Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 | /* * Copyright (c) 2016 Apple Inc. All rights reserved. * * @APPLE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in * compliance with the License. Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this * file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. * * @APPLE_LICENSE_HEADER_END@ */ #ifndef __VM_H #define __VM_H #include <malloc/_ptrcheck.h> __ptrcheck_abi_assume_single() #if MALLOC_TARGET_EXCLAVES typedef struct _liblibc_plat_mem_map_t plat_map_t; #define mvm_plat_map(x) (&(x)) #else #define mvm_plat_map(x) NULL #endif // MALLOC_TARGET_EXCLAVES extern uint64_t malloc_entropy[2]; #if !MALLOC_TARGET_EXCLAVES static inline bool mvm_aslr_enabled(void) { extern struct mach_header __dso_handle; return _dyld_get_image_slide(&__dso_handle); } #endif // !MALLOC_TARGET_EXCLAVES MALLOC_NOEXPORT void mvm_aslr_init(void); MALLOC_NOEXPORT void * __alloc_size(2) __sized_by_or_null(size) mvm_allocate_plat(uintptr_t addr, size_t size, uint8_t align, int flags, int debug_flags, int label, plat_map_t *map_out); MALLOC_NOEXPORT void * __alloc_size(1) __sized_by_or_null(size) mvm_allocate_pages(size_t size, uint8_t align, uint32_t debug_flags, int vm_page_label); MALLOC_NOEXPORT void * __alloc_size(1) __sized_by_or_null(size) mvm_allocate_pages_plat(size_t size, uint8_t align, uint32_t debug_flags, int vm_page_label, plat_map_t *map_out); MALLOC_NOEXPORT void mvm_deallocate_plat(void * __sized_by(size) addr, size_t size, int debug_flags, plat_map_t *map); MALLOC_NOEXPORT void mvm_deallocate_pages(void * __sized_by(size) addr, size_t size, unsigned debug_flags); MALLOC_NOEXPORT void mvm_deallocate_pages_plat(void * __sized_by(size) addr, size_t size, unsigned debug_flags, plat_map_t *map); MALLOC_NOEXPORT int mvm_madvise(void * __sized_by(size) addr, size_t size, int advice, unsigned debug_flags); MALLOC_NOEXPORT int mvm_madvise_plat(void * __sized_by(size) addr, size_t size, int advice, unsigned debug_flags, plat_map_t *map); MALLOC_NOEXPORT int mvm_madvise_free(void *szone, void *r, uintptr_t pgLo, uintptr_t pgHi, uintptr_t *last, boolean_t scribble); MALLOC_NOEXPORT int mvm_madvise_free_plat(void *szone, void *r, uintptr_t pgLo, uintptr_t pgHi, uintptr_t *last, boolean_t scribble, plat_map_t *map); MALLOC_NOEXPORT void mvm_protect(void * __sized_by(size) address, size_t size, unsigned protection, unsigned debug_flags); MALLOC_NOEXPORT void mvm_protect_plat(void * __sized_by(size) address, size_t size, unsigned protection, unsigned debug_flags, plat_map_t *map); #if CONFIG_DEFERRED_RECLAIM MALLOC_NOEXPORT kern_return_t mvm_deferred_reclaim_init(void); MALLOC_NOEXPORT bool mvm_reclaim_mark_used(uint64_t id, mach_vm_address_t ptr, uint32_t size, unsigned int debug_flags); MALLOC_NOEXPORT uint64_t mvm_reclaim_mark_free(vm_address_t ptr, uint32_t size, unsigned int debug_flags); MALLOC_NOEXPORT bool mvm_reclaim_is_available(uint64_t id); #endif // CONFIG_DEFERRED_RECLAIM #endif // __VM_H |