Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 | /* * Copyright (c) 2000 Apple Computer, Inc. All rights reserved. * * @APPLE_LICENSE_HEADER_START@ * * The contents of this file constitute Original Code as defined in and * are subject to the Apple Public Source License Version 1.1 (the * "License"). You may not use this file except in compliance with the * License. Please obtain a copy of the License at * http://www.apple.com/publicsource and read it before using this file. * * This Original Code and all software distributed under the License are * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the * License for the specific language governing rights and limitations * under the License. * * @APPLE_LICENSE_HEADER_END@ */ /* * @OSF_COPYRIGHT@ */ /* * Mach Operating System * Copyright (c) 1991,1990,1989,1988,1987 Carnegie Mellon University * All Rights Reserved. * * Permission to use, copy, modify and distribute this software and its * documentation is hereby granted, provided that both the copyright * notice and this permission notice appear in all copies of the * software, derivative works or modified versions, and any portions * thereof, and that both notices appear in supporting documentation. * * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND FOR * ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE. * * Carnegie Mellon requests users of this software to return to * * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU * School of Computer Science * Carnegie Mellon University * Pittsburgh PA 15213-3890 * * any improvements or extensions that they make and grant Carnegie Mellon * the rights to redistribute these changes. */ /* */ /* * File: vm/vm_map.h * Author: Avadis Tevanian, Jr., Michael Wayne Young * Date: 1985 * * Virtual memory map module definitions. * * Contributors: * avie, dlb, mwyoung */ #ifndef _VM_VM_MAP_H_ #define _VM_VM_MAP_H_ #include <mach/mach_types.h> #include <mach/kern_return.h> #include <mach/boolean.h> #include <mach/vm_types.h> #include <mach/vm_prot.h> #include <mach/vm_inherit.h> #include <mach/vm_behavior.h> #include <vm/pmap.h> typedef struct vm_map_entry *vm_map_entry_t; extern void kernel_vm_map_reference(vm_map_t map); #ifndef MACH_KERNEL_PRIVATE struct vm_map_entry {}; extern void vm_map_reference(vm_map_t map); extern vm_map_t current_map(void); #else /* MACH_KERNEL_PRIVATE */ #include <cpus.h> #include <task_swapper.h> #include <mach_assert.h> #include <vm/vm_object.h> #include <vm/vm_page.h> #include <kern/lock.h> #include <kern/zalloc.h> #include <kern/macro_help.h> #define shared_region_mapping_lock_init(object) \ mutex_init(&(object)->Lock, ETAP_VM_OBJ) #define shared_region_mapping_lock(object) mutex_lock(&(object)->Lock) #define shared_region_mapping_unlock(object) mutex_unlock(&(object)->Lock) #include <kern/thread_act.h> #define current_map_fast() (current_act_fast()->map) #define current_map() (current_map_fast()) /* * Types defined: * * vm_map_t the high-level address map data structure. * vm_map_entry_t an entry in an address map. * vm_map_version_t a timestamp of a map, for use with vm_map_lookup * vm_map_copy_t represents memory copied from an address map, * used for inter-map copy operations */ /* * Type: vm_map_object_t [internal use only] * * Description: * The target of an address mapping, either a virtual * memory object or a sub map (of the kernel map). */ typedef union vm_map_object { struct vm_object *vm_object; /* object object */ struct vm_map *sub_map; /* belongs to another map */ } vm_map_object_t; #define named_entry_lock_init(object) mutex_init(&(object)->Lock, ETAP_VM_OBJ) #define named_entry_lock(object) mutex_lock(&(object)->Lock) #define named_entry_unlock(object) mutex_unlock(&(object)->Lock) /* * Type: vm_named_entry_t [internal use only] * * Description: * Description of a mapping to a memory cache object. * * Implementation: * While the handle to this object is used as a means to map * and pass around the right to map regions backed by pagers * of all sorts, the named_entry itself is only manipulated * by the kernel. Named entries hold information on the * right to map a region of a cached object. Namely, * the target cache object, the beginning and ending of the * region to be mapped, and the permissions, (read, write) * with which it can be mapped. * */ struct vm_named_entry { decl_mutex_data(, Lock) /* Synchronization */ vm_object_t object; /* object I point to */ vm_object_offset_t offset; /* offset into object */ union { memory_object_t pager; /* amo pager port */ vm_map_t map; /* map backing submap */ } backing; unsigned int size; /* size of region */ unsigned int protection; /* access permissions */ int ref_count; /* Number of references */ unsigned int /* boolean_t */ internal:1, /* is an internal object */ /* boolean_t */ is_sub_map:1; /* is object is a submap? */ }; /* * Type: vm_map_entry_t [internal use only] * * Description: * A single mapping within an address map. * * Implementation: * Address map entries consist of start and end addresses, * a VM object (or sub map) and offset into that object, * and user-exported inheritance and protection information. * Control information for virtual copy operations is also * stored in the address map entry. */ struct vm_map_links { struct vm_map_entry *prev; /* previous entry */ struct vm_map_entry *next; /* next entry */ vm_offset_t start; /* start address */ vm_offset_t end; /* end address */ }; struct vm_map_entry { struct vm_map_links links; /* links to other entries */ #define vme_prev links.prev #define vme_next links.next #define vme_start links.start #define vme_end links.end union vm_map_object object; /* object I point to */ vm_object_offset_t offset; /* offset into object */ unsigned int /* boolean_t */ is_shared:1, /* region is shared */ /* boolean_t */ is_sub_map:1, /* Is "object" a submap? */ /* boolean_t */ in_transition:1, /* Entry being changed */ /* boolean_t */ needs_wakeup:1, /* Waiters on in_transition */ /* vm_behavior_t */ behavior:2, /* user paging behavior hint */ /* behavior is not defined for submap type */ /* boolean_t */ needs_copy:1, /* object need to be copied? */ /* Only in task maps: */ /* vm_prot_t */ protection:3, /* protection code */ /* vm_prot_t */ max_protection:3,/* maximum protection */ /* vm_inherit_t */ inheritance:2, /* inheritance */ /* nested pmap */ use_pmap:1, /* nested pmaps */ /* user alias */ alias:8; unsigned short wired_count; /* can be paged if = 0 */ unsigned short user_wired_count; /* for vm_wire */ }; /* * wired_counts are unsigned short. This value is used to safeguard * against any mishaps due to runaway user programs. */ #define MAX_WIRE_COUNT 65535 /* * Type: struct vm_map_header * * Description: * Header for a vm_map and a vm_map_copy. */ struct vm_map_header { struct vm_map_links links; /* first, last, min, max */ int nentries; /* Number of entries */ boolean_t entries_pageable; /* are map entries pageable? */ }; /* * Type: vm_map_t [exported; contents invisible] * * Description: * An address map -- a directory relating valid * regions of a task's address space to the corresponding * virtual memory objects. * * Implementation: * Maps are doubly-linked lists of map entries, sorted * by address. One hint is used to start * searches again from the last successful search, * insertion, or removal. Another hint is used to * quickly find free space. */ struct vm_map { lock_t lock; /* uni- and smp-lock */ struct vm_map_header hdr; /* Map entry header */ #define min_offset hdr.links.start /* start of range */ #define max_offset hdr.links.end /* end of range */ pmap_t pmap; /* Physical map */ vm_size_t size; /* virtual size */ int ref_count; /* Reference count */ #if TASK_SWAPPER int res_count; /* Residence count (swap) */ int sw_state; /* Swap state */ #endif /* TASK_SWAPPER */ decl_mutex_data(, s_lock) /* Lock ref, res, hint fields */ vm_map_entry_t hint; /* hint for quick lookups */ vm_map_entry_t first_free; /* First free space hint */ boolean_t wait_for_space; /* Should callers wait for space? */ boolean_t wiring_required;/* All memory wired? */ boolean_t no_zero_fill; /* No zero fill absent pages */ unsigned int timestamp; /* Version number */ } ; #define vm_map_to_entry(map) ((struct vm_map_entry *) &(map)->hdr.links) #define vm_map_first_entry(map) ((map)->hdr.links.next) #define vm_map_last_entry(map) ((map)->hdr.links.prev) #if TASK_SWAPPER /* * VM map swap states. There are no transition states. */ #define MAP_SW_IN 1 /* map is swapped in; residence count > 0 */ #define MAP_SW_OUT 2 /* map is out (res_count == 0 */ #endif /* TASK_SWAPPER */ /* * Type: vm_map_version_t [exported; contents invisible] * * Description: * Map versions may be used to quickly validate a previous * lookup operation. * * Usage note: * Because they are bulky objects, map versions are usually * passed by reference. * * Implementation: * Just a timestamp for the main map. */ typedef struct vm_map_version { unsigned int main_timestamp; } vm_map_version_t; /* * Type: vm_map_copy_t [exported; contents invisible] * * Description: * A map copy object represents a region of virtual memory * that has been copied from an address map but is still * in transit. * * A map copy object may only be used by a single thread * at a time. * * Implementation: * There are three formats for map copy objects. * The first is very similar to the main * address map in structure, and as a result, some * of the internal maintenance functions/macros can * be used with either address maps or map copy objects. * * The map copy object contains a header links * entry onto which the other entries that represent * the region are chained. * * The second format is a single vm object. This is used * primarily in the pageout path. The third format is a * list of vm pages. An optional continuation provides * a hook to be called to obtain more of the memory, * or perform other operations. The continuation takes 3 * arguments, a saved arg buffer, a pointer to a new vm_map_copy * (returned) and an abort flag (abort if TRUE). */ #define VM_MAP_COPY_PAGE_LIST_MAX 20 #define VM_MAP_COPY_PAGE_LIST_MAX_SIZE (VM_MAP_COPY_PAGE_LIST_MAX * PAGE_SIZE) /* * Options for vm_map_copyin_page_list. */ #define VM_MAP_COPYIN_OPT_VM_PROT 0x7 #define VM_MAP_COPYIN_OPT_SRC_DESTROY 0x8 #define VM_MAP_COPYIN_OPT_STEAL_PAGES 0x10 #define VM_MAP_COPYIN_OPT_PMAP_ENTER 0x20 #define VM_MAP_COPYIN_OPT_NO_ZERO_FILL 0x40 /* * Continuation structures for vm_map_copyin_page_list. */ typedef struct { vm_map_t map; vm_offset_t src_addr; vm_size_t src_len; vm_offset_t destroy_addr; vm_size_t destroy_len; int options; } vm_map_copyin_args_data_t, *vm_map_copyin_args_t; #define VM_MAP_COPYIN_ARGS_NULL ((vm_map_copyin_args_t) 0) /* vm_map_copy_cont_t is a type definition/prototype * for the cont function pointer in vm_map_copy structure. */ typedef kern_return_t (*vm_map_copy_cont_t)( vm_map_copyin_args_t, vm_map_copy_t *); #define VM_MAP_COPY_CONT_NULL ((vm_map_copy_cont_t) 0) struct vm_map_copy { int type; #define VM_MAP_COPY_ENTRY_LIST 1 #define VM_MAP_COPY_OBJECT 2 #define VM_MAP_COPY_KERNEL_BUFFER 3 vm_object_offset_t offset; vm_size_t size; union { struct vm_map_header hdr; /* ENTRY_LIST */ struct { /* OBJECT */ vm_object_t object; vm_size_t index; /* record progress as pages * are moved from object to * page list; must be zero * when first invoking * vm_map_object_to_page_list */ } c_o; struct { /* KERNEL_BUFFER */ vm_offset_t kdata; vm_size_t kalloc_size; /* size of this copy_t */ } c_k; } c_u; }; #define cpy_hdr c_u.hdr #define cpy_object c_u.c_o.object #define cpy_index c_u.c_o.index #define cpy_kdata c_u.c_k.kdata #define cpy_kalloc_size c_u.c_k.kalloc_size /* * Useful macros for entry list copy objects */ #define vm_map_copy_to_entry(copy) \ ((struct vm_map_entry *) &(copy)->cpy_hdr.links) #define vm_map_copy_first_entry(copy) \ ((copy)->cpy_hdr.links.next) #define vm_map_copy_last_entry(copy) \ ((copy)->cpy_hdr.links.prev) /* * Macros: vm_map_lock, etc. [internal use only] * Description: * Perform locking on the data portion of a map. * When multiple maps are to be locked, order by map address. * (See vm_map.c::vm_remap()) */ #define vm_map_lock_init(map) \ MACRO_BEGIN \ lock_init(&(map)->lock, TRUE, ETAP_VM_MAP, ETAP_VM_MAP_I); \ (map)->timestamp = 0; \ MACRO_END #define vm_map_lock(map) \ MACRO_BEGIN \ lock_write(&(map)->lock); \ (map)->timestamp++; \ MACRO_END #define vm_map_unlock(map) lock_write_done(&(map)->lock) #define vm_map_lock_read(map) lock_read(&(map)->lock) #define vm_map_unlock_read(map) lock_read_done(&(map)->lock) #define vm_map_lock_write_to_read(map) \ lock_write_to_read(&(map)->lock) #define vm_map_lock_read_to_write(map) \ (lock_read_to_write(&(map)->lock) || (((map)->timestamp++), 0)) extern zone_t vm_map_copy_zone; /* zone for vm_map_copy structures */ /* * Exported procedures that operate on vm_map_t. */ /* Initialize the module */ extern void vm_map_init(void); /* Allocate a range in the specified virtual address map and * return the entry allocated for that range. */ extern kern_return_t vm_map_find_space( vm_map_t map, vm_offset_t *address, /* OUT */ vm_size_t size, vm_offset_t mask, vm_map_entry_t *o_entry); /* OUT */ /* Lookup map entry containing or the specified address in the given map */ extern boolean_t vm_map_lookup_entry( vm_map_t map, vm_offset_t address, vm_map_entry_t *entry); /* OUT */ /* Find the VM object, offset, and protection for a given virtual address * in the specified map, assuming a page fault of the type specified. */ extern kern_return_t vm_map_lookup_locked( vm_map_t *var_map, /* IN/OUT */ vm_offset_t vaddr, vm_prot_t fault_type, vm_map_version_t *out_version, /* OUT */ vm_object_t *object, /* OUT */ vm_object_offset_t *offset, /* OUT */ vm_prot_t *out_prot, /* OUT */ boolean_t *wired, /* OUT */ int *behavior, /* OUT */ vm_object_offset_t *lo_offset, /* OUT */ vm_object_offset_t *hi_offset, /* OUT */ vm_map_t *pmap_map); /* OUT */ /* Verifies that the map has not changed since the given version. */ extern boolean_t vm_map_verify( vm_map_t map, vm_map_version_t *version); /* REF */ /* Split a vm_map_entry into 2 entries */ extern void _vm_map_clip_start( struct vm_map_header *map_header, vm_map_entry_t entry, vm_offset_t start); extern vm_map_entry_t vm_map_entry_insert( vm_map_t map, vm_map_entry_t insp_entry, vm_offset_t start, vm_offset_t end, vm_object_t object, vm_object_offset_t offset, boolean_t needs_copy, boolean_t is_shared, boolean_t in_transition, vm_prot_t cur_protection, vm_prot_t max_protection, vm_behavior_t behavior, vm_inherit_t inheritance, unsigned wired_count); extern kern_return_t vm_remap_extract( vm_map_t map, vm_offset_t addr, vm_size_t size, boolean_t copy, struct vm_map_header *map_header, vm_prot_t *cur_protection, vm_prot_t *max_protection, vm_inherit_t inheritance, boolean_t pageable); extern kern_return_t vm_remap_range_allocate( vm_map_t map, vm_offset_t *address, vm_size_t size, vm_offset_t mask, boolean_t anywhere, vm_map_entry_t *map_entry); extern kern_return_t vm_remap_extract( vm_map_t map, vm_offset_t addr, vm_size_t size, boolean_t copy, struct vm_map_header *map_header, vm_prot_t *cur_protection, vm_prot_t *max_protection, vm_inherit_t inheritance, boolean_t pageable); extern kern_return_t vm_remap_range_allocate( vm_map_t map, vm_offset_t *address, vm_size_t size, vm_offset_t mask, boolean_t anywhere, vm_map_entry_t *map_entry); /* * Functions implemented as macros */ #define vm_map_min(map) ((map)->min_offset) /* Lowest valid address in * a map */ #define vm_map_max(map) ((map)->max_offset) /* Highest valid address */ #define vm_map_pmap(map) ((map)->pmap) /* Physical map associated * with this address map */ #define vm_map_verify_done(map, version) vm_map_unlock_read(map) /* Operation that required * a verified lookup is * now complete */ /* * Macros/functions for map residence counts and swapin/out of vm maps */ #if TASK_SWAPPER #if MACH_ASSERT /* Gain a reference to an existing map */ extern void vm_map_reference( vm_map_t map); /* Lose a residence count */ extern void vm_map_res_deallocate( vm_map_t map); /* Gain a residence count on a map */ extern void vm_map_res_reference( vm_map_t map); /* Gain reference & residence counts to possibly swapped-out map */ extern void vm_map_reference_swap( vm_map_t map); #else /* MACH_ASSERT */ #define vm_map_reference(map) \ MACRO_BEGIN \ vm_map_t Map = (map); \ if (Map) { \ mutex_lock(&Map->s_lock); \ Map->res_count++; \ Map->ref_count++; \ mutex_unlock(&Map->s_lock); \ } \ MACRO_END #define vm_map_res_reference(map) \ MACRO_BEGIN \ vm_map_t Lmap = (map); \ if (Lmap->res_count == 0) { \ mutex_unlock(&Lmap->s_lock); \ vm_map_lock(Lmap); \ vm_map_swapin(Lmap); \ mutex_lock(&Lmap->s_lock); \ ++Lmap->res_count; \ vm_map_unlock(Lmap); \ } else \ ++Lmap->res_count; \ MACRO_END #define vm_map_res_deallocate(map) \ MACRO_BEGIN \ vm_map_t Map = (map); \ if (--Map->res_count == 0) { \ mutex_unlock(&Map->s_lock); \ vm_map_lock(Map); \ vm_map_swapout(Map); \ vm_map_unlock(Map); \ mutex_lock(&Map->s_lock); \ } \ MACRO_END #define vm_map_reference_swap(map) \ MACRO_BEGIN \ vm_map_t Map = (map); \ mutex_lock(&Map->s_lock); \ ++Map->ref_count; \ vm_map_res_reference(Map); \ mutex_unlock(&Map->s_lock); \ MACRO_END #endif /* MACH_ASSERT */ extern void vm_map_swapin( vm_map_t map); extern void vm_map_swapout( vm_map_t map); #else /* TASK_SWAPPER */ #define vm_map_reference(map) \ MACRO_BEGIN \ vm_map_t Map = (map); \ if (Map) { \ mutex_lock(&Map->s_lock); \ Map->ref_count++; \ mutex_unlock(&Map->s_lock); \ } \ MACRO_END #define vm_map_reference_swap(map) vm_map_reference(map) #define vm_map_res_reference(map) #define vm_map_res_deallocate(map) #endif /* TASK_SWAPPER */ /* * Submap object. Must be used to create memory to be put * in a submap by vm_map_submap. */ extern vm_object_t vm_submap_object; /* * Wait and wakeup macros for in_transition map entries. */ #define vm_map_entry_wait(map, interruptible) \ MACRO_BEGIN \ assert_wait((event_t)&(map)->hdr, interruptible); \ vm_map_unlock(map); \ thread_block((void (*)(void))0); \ MACRO_END #define vm_map_entry_wakeup(map) thread_wakeup((event_t)(&(map)->hdr)) #define vm_map_ref_fast(map) \ MACRO_BEGIN \ mutex_lock(&map->s_lock); \ map->ref_count++; \ vm_map_res_reference(map); \ mutex_unlock(&map->s_lock); \ MACRO_END #define vm_map_dealloc_fast(map) \ MACRO_BEGIN \ register int c; \ \ mutex_lock(&map->s_lock); \ c = --map->ref_count; \ if (c > 0) \ vm_map_res_deallocate(map); \ mutex_unlock(&map->s_lock); \ if (c == 0) \ vm_map_destroy(map); \ MACRO_END /* simplify map entries */ extern void vm_map_simplify( vm_map_t map, vm_offset_t start); /* Steal all the pages from a vm_map_copy page_list */ extern void vm_map_copy_steal_pages( vm_map_copy_t copy); /* Discard a copy without using it */ extern void vm_map_copy_discard( vm_map_copy_t copy); /* Move the information in a map copy object to a new map copy object */ extern vm_map_copy_t vm_map_copy_copy( vm_map_copy_t copy); /* Overwrite existing memory with a copy */ extern kern_return_t vm_map_copy_overwrite( vm_map_t dst_map, vm_offset_t dst_addr, vm_map_copy_t copy, int interruptible); /* Create a copy object from an object. */ extern kern_return_t vm_map_copyin_object( vm_object_t object, vm_object_offset_t offset, vm_object_size_t size, vm_map_copy_t *copy_result); /* OUT */ extern vm_map_t vm_map_switch( vm_map_t map); extern int vm_map_copy_cont_is_valid( vm_map_copy_t copy); #endif /* !MACH_KERNEL_PRIVATE */ /* Get rid of a map */ extern void vm_map_destroy( vm_map_t map); /* Lose a reference */ extern void vm_map_deallocate( vm_map_t map); /* Create an empty map */ extern vm_map_t vm_map_create( pmap_t pmap, vm_offset_t min, vm_offset_t max, boolean_t pageable); /* Enter a mapping */ extern kern_return_t vm_map_enter( vm_map_t map, vm_offset_t *address, vm_size_t size, vm_offset_t mask, int flags, vm_object_t object, vm_object_offset_t offset, boolean_t needs_copy, vm_prot_t cur_protection, vm_prot_t max_protection, vm_inherit_t inheritance); extern kern_return_t vm_map_write_user( vm_map_t map, vm_offset_t src_addr, vm_offset_t dst_addr, vm_size_t size); extern kern_return_t vm_map_read_user( vm_map_t map, vm_offset_t src_addr, vm_offset_t dst_addr, vm_size_t size); /* Create a new task map using an existing task map as a template. */ extern vm_map_t vm_map_fork( vm_map_t old_map); /* Change protection */ extern kern_return_t vm_map_protect( vm_map_t map, vm_offset_t start, vm_offset_t end, vm_prot_t new_prot, boolean_t set_max); /* Change inheritance */ extern kern_return_t vm_map_inherit( vm_map_t map, vm_offset_t start, vm_offset_t end, vm_inherit_t new_inheritance); /* wire down a region */ extern kern_return_t vm_map_wire( vm_map_t map, vm_offset_t start, vm_offset_t end, vm_prot_t access_type, boolean_t user_wire); /* unwire a region */ extern kern_return_t vm_map_unwire( vm_map_t map, vm_offset_t start, vm_offset_t end, boolean_t user_wire); /* Deallocate a region */ extern kern_return_t vm_map_remove( vm_map_t map, vm_offset_t start, vm_offset_t end, boolean_t flags); /* Place a copy into a map */ extern kern_return_t vm_map_copyout( vm_map_t dst_map, vm_offset_t *dst_addr, /* OUT */ vm_map_copy_t copy); /* Add or remove machine-dependent attributes from map regions */ extern kern_return_t vm_map_machine_attribute( vm_map_t map, vm_offset_t address, vm_size_t size, vm_machine_attribute_t attribute, vm_machine_attribute_val_t* value); /* IN/OUT */ /* Set paging behavior */ extern kern_return_t vm_map_behavior_set( vm_map_t map, vm_offset_t start, vm_offset_t end, vm_behavior_t new_behavior); extern kern_return_t vm_map_copyin_common( vm_map_t src_map, vm_offset_t src_addr, vm_size_t len, boolean_t src_destroy, boolean_t src_volatile, vm_map_copy_t *copy_result, /* OUT */ boolean_t use_maxprot); extern kern_return_t vm_map_submap( vm_map_t map, vm_offset_t start, vm_offset_t end, vm_map_t submap, vm_offset_t offset, boolean_t use_pmap); extern kern_return_t vm_region_clone( ipc_port_t src_region, ipc_port_t dst_region); extern kern_return_t vm_map_region_replace( vm_map_t target_map, ipc_port_t old_region, ipc_port_t new_region, vm_offset_t start, vm_offset_t end); /* * Macros to invoke vm_map_copyin_common. vm_map_copyin is the * usual form; it handles a copyin based on the current protection * (current protection == VM_PROT_NONE) is a failure. * vm_map_copyin_maxprot handles a copyin based on maximum possible * access. The difference is that a region with no current access * BUT possible maximum access is rejected by vm_map_copyin(), but * returned by vm_map_copyin_maxprot. */ #define vm_map_copyin(src_map, src_addr, len, src_destroy, copy_result) \ vm_map_copyin_common(src_map, src_addr, len, src_destroy, \ FALSE, copy_result, FALSE) #define vm_map_copyin_maxprot(src_map, \ src_addr, len, src_destroy, copy_result) \ vm_map_copyin_common(src_map, src_addr, len, src_destroy, \ FALSE, copy_result, TRUE) #define VM_MAP_ENTRY_NULL ((vm_map_entry_t) 0) /* * Flags for vm_map_remove() and vm_map_delete() */ #define VM_MAP_NO_FLAGS 0x0 #define VM_MAP_REMOVE_KUNWIRE 0x1 #define VM_MAP_REMOVE_INTERRUPTIBLE 0x2 #define VM_MAP_REMOVE_WAIT_FOR_KWIRE 0x4 #ifdef MACH_KERNEL_PRIVATE /* address space shared region descriptor */ struct shared_region_mapping { decl_mutex_data(, Lock) /* Synchronization */ int ref_count; ipc_port_t text_region; vm_size_t text_size; ipc_port_t data_region; vm_size_t data_size; vm_offset_t region_mappings; vm_offset_t client_base; vm_offset_t alternate_base; vm_offset_t alternate_next; int flags; int depth; struct shared_region_object_chain *object_chain; struct shared_region_mapping *self; struct shared_region_mapping *next; }; typedef struct shared_region_mapping *shared_region_mapping_t; struct shared_region_object_chain { shared_region_mapping_t object_chain_region; int depth; struct shared_region_object_chain *next; }; typedef struct shared_region_object_chain *shared_region_object_chain_t; #else /* !MACH_KERNEL_PRIVATE */ typedef void *shared_region_mapping_t; #endif /* MACH_KERNEL_PRIVATE */ /* address space shared region descriptor */ extern kern_return_t shared_region_mapping_info( shared_region_mapping_t shared_region, ipc_port_t *text_region, vm_size_t *text_size, ipc_port_t *data_region, vm_size_t *data_size, vm_offset_t *region_mappings, vm_offset_t *client_base, vm_offset_t *alternate_base, vm_offset_t *alternate_next, int *flags, shared_region_mapping_t *next); extern kern_return_t shared_region_mapping_create( ipc_port_t text_region, vm_size_t text_size, ipc_port_t data_region, vm_size_t data_size, vm_offset_t region_mappings, vm_offset_t client_base, shared_region_mapping_t *shared_region, vm_offset_t alt_base, vm_offset_t alt_next); extern kern_return_t shared_region_mapping_ref( shared_region_mapping_t shared_region); extern kern_return_t shared_region_mapping_dealloc( shared_region_mapping_t shared_region); extern kern_return_t shared_region_object_chain_attach( shared_region_mapping_t target_region, shared_region_mapping_t object_chain); /* extern kern_return_t vm_get_shared_region( task_t task, shared_region_mapping_t *shared_region); extern kern_return_t vm_set_shared_region( task_t task, shared_region_mapping_t shared_region); */ #endif /* _VM_VM_MAP_H_ */ |