| /xnu-8020.140.41/osfmk/mach/arm/ |
| H A D | vm_param.h | 175 #define VM_MIN_KERNEL_ADDRESS ((vm_address_t) 0x80000000) macro 214 #define VM_MIN_KERNEL_ADDRESS ((vm_address_t) (0ULL - TiB(2))) macro 217 #define VM_MAX_KERNEL_ADDRESS ((vm_address_t) (VM_MIN_KERNEL_ADDRESS + GiB(64) + GiB(512) - 1)) 237 #define VM_MIN_KERNEL_ADDRESS ((vm_address_t) 0xffffffe000000000ULL) macro 247 #define VM_MIN_KERNEL_ADDRESS ((vm_address_t) (0ULL - TiB(2))) macro 254 #define VM_MIN_KERNEL_AND_KEXT_ADDRESS VM_MIN_KERNEL_ADDRESS 287 ((VM_KERNEL_STRIP_UPTR(_va) >= VM_MIN_KERNEL_ADDRESS) && \
|
| /xnu-8020.140.41/osfmk/mach/i386/ |
| H A D | vm_param.h | 227 #define VM_MIN_KERNEL_ADDRESS ((vm_offset_t) 0xFFFFFF8000000000UL) macro 229 #define VM_MIN_KERNEL_AND_KEXT_ADDRESS (VM_MIN_KERNEL_ADDRESS - 0x80000000ULL)
|
| /xnu-8020.140.41/san/memory/ |
| H A D | kasan-arm64.c | 90 _Static_assert((VM_MIN_KERNEL_ADDRESS >> KASAN_SCALE) + KASAN_OFFSET_ARM64 >= KASAN_SHADOW_MIN, "KA… 317 kasan_map_shadow_internal(VM_MIN_KERNEL_ADDRESS, in kasan_arch_init() 318 VM_MAX_KERNEL_ADDRESS - VM_MIN_KERNEL_ADDRESS + 1, KASAN_ARM64_PREALLOCATE_TRANSLATION); in kasan_arch_init()
|
| /xnu-8020.140.41/osfmk/i386/AT386/ |
| H A D | model_dep.c | 232 bool dump_kernel_stack = (fp >= VM_MIN_KERNEL_ADDRESS); in print_one_backtrace() 238 if (dump_kernel_stack && ((fp < VM_MIN_KERNEL_ADDRESS) || (fp > VM_MAX_KERNEL_ADDRESS))) { in print_one_backtrace() 241 if ((!dump_kernel_stack) && (fp >= VM_MIN_KERNEL_ADDRESS)) { in print_one_backtrace() 454 if (vm_addr < VM_MIN_KERNEL_ADDRESS) { in efi_init() 455 vm_addr |= VM_MIN_KERNEL_ADDRESS; in efi_init() 530 if (vm_addr < VM_MIN_KERNEL_ADDRESS) { in hibernate_newruntime_map() 531 vm_addr |= VM_MIN_KERNEL_ADDRESS; in hibernate_newruntime_map() 550 if (vm_addr < VM_MIN_KERNEL_ADDRESS) { in hibernate_newruntime_map() 551 vm_addr |= VM_MIN_KERNEL_ADDRESS; in hibernate_newruntime_map() 1806 if (addr >= VM_MIN_KERNEL_ADDRESS) { in efi_efiboot_virtual_to_physical()
|
| /xnu-8020.140.41/san/coverage/ |
| H A D | kcov_ksancov.c | 161 uint32_t pc = (uint32_t)(VM_KERNEL_UNSLIDE(caller) - VM_MIN_KERNEL_ADDRESS - 1); in kcov_ksancov_trace_guard() 181 uint32_t pc = (uint32_t)(VM_KERNEL_UNSLIDE(caller) - VM_MIN_KERNEL_ADDRESS - 1); in kcov_ksancov_trace_pc() 459 trace->kt_offset = VM_MIN_KERNEL_ADDRESS; in ksancov_trace_alloc() 771 ksancov_edgemap->ke_offset = VM_MIN_KERNEL_ADDRESS; in ksancov_init_dev()
|
| H A D | kcov_ksancov_data.h | 40 #define KSANCOV_PC_OFFSET VM_MIN_KERNEL_ADDRESS
|
| /xnu-8020.140.41/osfmk/arm/ |
| H A D | model_dep.c | 200 bool dump_kernel_stack = (fp >= VM_MIN_KERNEL_ADDRESS); in print_one_backtrace() 206 if (dump_kernel_stack && ((fp < VM_MIN_KERNEL_ADDRESS) || (fp > VM_MAX_KERNEL_ADDRESS))) { in print_one_backtrace() 209 if ((!dump_kernel_stack) && (fp >= VM_MIN_KERNEL_ADDRESS)) { in print_one_backtrace()
|
| H A D | cpu_data_internal.h | 77 static_assert((CPUWINDOWS_BASE >= VM_MIN_KERNEL_ADDRESS) && ((CPUWINDOWS_TOP - 1) <= VM_MAX_KERNEL_…
|
| H A D | genassym.c | 195 DECLARE("KERNELBASE", VM_MIN_KERNEL_ADDRESS); in main()
|
| H A D | machine_routines.c | 817 if (vaddr < VM_MIN_KERNEL_ADDRESS) { in ml_static_protect() 896 if (vaddr < VM_MIN_KERNEL_ADDRESS) { in ml_static_mfree()
|
| H A D | loose_ends.c | 655 if (__improbable(kernel_addr < VM_MIN_KERNEL_ADDRESS || in copy_validate()
|
| /xnu-8020.140.41/osfmk/i386/ |
| H A D | pal_routines.c | 175 if (func < VM_MIN_KERNEL_ADDRESS) { in pal_efi_call_in_64bit_mode()
|
| H A D | machine_routines.c | 134 return (vm_offset_t)(((unsigned long) paddr) | VM_MIN_KERNEL_ADDRESS); in ml_static_ptovirt() 204 assert(vaddr >= VM_MIN_KERNEL_ADDRESS); in ml_static_mfree()
|
| H A D | genassym.c | 273 DECLARE("KERNELBASE", VM_MIN_KERNEL_ADDRESS); in main()
|
| /xnu-8020.140.41/osfmk/kern/ |
| H A D | kext_alloc.c | 128 kext_alloc_base = VM_MIN_KERNEL_ADDRESS; in kext_alloc_init()
|
| /xnu-8020.140.41/osfmk/kdp/ml/arm/ |
| H A D | kdp_machdep.c | 426 stacklimit_bottom = VM_MIN_KERNEL_ADDRESS; in machine_trace_thread() 583 stacklimit_bottom = VM_MIN_KERNEL_ADDRESS; in machine_trace_thread64()
|
| /xnu-8020.140.41/osfmk/arm64/ |
| H A D | copyio.c | 155 bool in_kva = (VM_KERNEL_STRIP_UPTR(kernel_addr) >= VM_MIN_KERNEL_ADDRESS) && in copy_validate()
|
| H A D | arm_vm_init.c | 1829 if (mem_size >= ((VM_MAX_KERNEL_ADDRESS - VM_MIN_KERNEL_ADDRESS) / 2)) { in arm_vm_init() 1835 physmap_base = VM_MIN_KERNEL_ADDRESS - (physmap_l1_entries << ARM_TT_L1_SHIFT); in arm_vm_init() 2253 va_l1 = VM_MIN_KERNEL_ADDRESS & ~ARM_TT_L1_OFFMASK; in arm_vm_init() 2254 …init_ptpages(cpu_tte, VM_MIN_KERNEL_ADDRESS & ~ARM_TT_L1_OFFMASK, VM_MAX_KERNEL_ADDRESS, FALSE, AR… in arm_vm_init()
|
| H A D | genassym.c | 248 DECLARE("VM_MIN_KERNEL_ADDRESS", VM_MIN_KERNEL_ADDRESS); in main()
|
| H A D | machine_routines.c | 1538 if (vaddr < VM_MIN_KERNEL_ADDRESS) { in ml_static_protect() 1539 panic("ml_static_protect(): %p < %p", (void *) vaddr, (void *) VM_MIN_KERNEL_ADDRESS); in ml_static_protect() 1679 if (vaddr < VM_MIN_KERNEL_ADDRESS) { in ml_static_mfree()
|
| /xnu-8020.140.41/bsd/sys/ |
| H A D | dtrace_glue.h | 478 #define KERNELBASE VM_MIN_KERNEL_ADDRESS
|
| /xnu-8020.140.41/tools/lldbmacros/core/ |
| H A D | kernelcore.py | 717 return self.VM_MIN_KERNEL_ADDRESS - 0x80000000 719 return self.VM_MIN_KERNEL_ADDRESS
|
| /xnu-8020.140.41/bsd/dev/arm/ |
| H A D | dtrace_isa.c | 610 func(0x0, VM_MIN_KERNEL_ADDRESS); in dtrace_toxic_ranges()
|
| /xnu-8020.140.41/bsd/dev/arm64/ |
| H A D | dtrace_isa.c | 700 func(0x0, VM_MIN_KERNEL_ADDRESS); in dtrace_toxic_ranges()
|
| /xnu-8020.140.41/osfmk/x86_64/ |
| H A D | pmap.c | 521 virtual_avail = (vm_offset_t)(VM_MIN_KERNEL_ADDRESS) + (vm_offset_t)first_avail; in pmap_bootstrap() 923 assert((uintptr_t)VM_MIN_KERNEL_ADDRESS + avail_start <= (uintptr_t)vm_page_array_beginning_addr); in pmap_init() 924 pmap_pv_fixup((uintptr_t)VM_MIN_KERNEL_ADDRESS, (uintptr_t)VM_MIN_KERNEL_ADDRESS + avail_start); in pmap_init()
|