| /xnu-8019.80.24/osfmk/i386/ |
| H A D | pal_routines.c | 162 return KERN_INVALID_ADDRESS; in pal_efi_call_in_64bit_mode() 180 return KERN_INVALID_ADDRESS; in pal_efi_call_in_64bit_mode() 210 return KERN_INVALID_ADDRESS; in pal_efi_call_in_32bit_mode()
|
| H A D | pmap.h | 881 (KERN_INVALID_ADDRESS) 883 (KERN_INVALID_ADDRESS)
|
| H A D | bsd_i386.c | 323 regs->eax = KERN_INVALID_ADDRESS; in machdep_syscall()
|
| /xnu-8019.80.24/tests/vm/ |
| H A D | vm_allocation.c | 1811 …size ? KERN_INVALID_ADDRESS : (get_allocator() != wrapper_mach_vm_allocate) ? KERN_INVALID_ARGUMEN… in test_allocate_at_zero() 2190 assert_read_return(address, size, &read_address, &read_size, KERN_INVALID_ADDRESS); in test_read_partially_deallocated_range() 2191 logv("Returned expected error: %s.", mach_error_string(KERN_INVALID_ADDRESS)); in test_read_partially_deallocated_range() 2207 …kern_return_t kr_expected = (size < vm_page_size * 2) ? KERN_INVALID_ADDRESS : KERN_PROTECTION_FAI… in test_read_partially_unreadable_range() 2294 T_QUIET; T_ASSERT_EQ(kr, KERN_INVALID_ADDRESS, in test_read_wrapped_around_ranges() 2299 mach_error_string(KERN_INVALID_ADDRESS)); in test_read_wrapped_around_ranges() 2301 logv("Returned expected error on each range: %s.", mach_error_string(KERN_INVALID_ADDRESS)); in test_read_wrapped_around_ranges() 2447 assert_write_return(address, data, buffer_size, KERN_INVALID_ADDRESS); in test_write_on_partially_deallocated_range() 2448 logv("Returned expected error: %s.", mach_error_string(KERN_INVALID_ADDRESS)); in test_write_on_partially_deallocated_range() 2464 …kern_return_t kr_expected = (buffer_size < vm_page_size * 2) ? KERN_INVALID_ADDRESS : KERN_PROTECT… in test_write_on_partially_unwritable_range() [all …]
|
| /xnu-8019.80.24/osfmk/mach/ |
| H A D | kern_return.h | 74 #define KERN_INVALID_ADDRESS 1 macro
|
| /xnu-8019.80.24/osfmk/vm/ |
| H A D | vm_kern.c | 1507 kr = KERN_INVALID_ADDRESS; in copyinmap() 1513 kr = KERN_INVALID_ADDRESS; in copyinmap() 1542 kr = KERN_INVALID_ADDRESS; in copyoutmap() 1548 kr = KERN_INVALID_ADDRESS; in copyoutmap() 1576 kr = KERN_INVALID_ADDRESS; in copyoutmap_atomic32() 1582 kr = KERN_INVALID_ADDRESS; in copyoutmap_atomic32() 1604 kr = KERN_INVALID_ADDRESS; in copyoutmap_atomic64() 1610 kr = KERN_INVALID_ADDRESS; in copyoutmap_atomic64()
|
| H A D | vm_map.c | 891 kr = KERN_INVALID_ADDRESS; in vm_map_apple_protected() 897 kr = KERN_INVALID_ADDRESS; in vm_map_apple_protected() 2713 RETURN(KERN_INVALID_ADDRESS); in vm_map_enter() 3583 RETURN(KERN_INVALID_ADDRESS); in vm_map_enter_fourk() 5891 return KERN_INVALID_ADDRESS; in vm_map_protect() 5959 return KERN_INVALID_ADDRESS; in vm_map_protect() 5969 return KERN_INVALID_ADDRESS; in vm_map_protect() 5996 return KERN_INVALID_ADDRESS; in vm_map_protect() 6088 return KERN_INVALID_ADDRESS; in vm_map_protect() 6509 rc = KERN_INVALID_ADDRESS; in vm_map_wire_nested() [all …]
|
| H A D | vm32_user.c | 401 return KERN_INVALID_ADDRESS; in vm32_region() 479 return KERN_INVALID_ADDRESS; in vm32_region_recurse()
|
| H A D | vm_tests.c | 630 …assertf(kr == KERN_INVALID_ADDRESS, "fault_addr = 0x%lx kr = 0x%x expected 0x%x", fault_addr, kr, … in vm_test_4k()
|
| H A D | vm_shared_region.c | 994 kr = KERN_INVALID_ADDRESS; in vm_shared_region_start_address() 2318 kr = KERN_INVALID_ADDRESS; in vm_shared_region_slide_mapping() 2335 kr = KERN_INVALID_ADDRESS; in vm_shared_region_slide_mapping()
|
| H A D | vm_user.c | 1967 return KERN_INVALID_ADDRESS; in vm_region_64() 2002 return KERN_INVALID_ADDRESS; in vm_region() 2082 return KERN_INVALID_ADDRESS; in vm_region_recurse_64() 2138 return KERN_INVALID_ADDRESS; in vm_region_recurse()
|
| H A D | vm_fourk_pager.c | 1302 return KERN_INVALID_ADDRESS; in fourk_pager_populate()
|
| /xnu-8019.80.24/bsd/kern/ |
| H A D | posix_sem.c | 394 kern_return_t kret = KERN_INVALID_ADDRESS; /* default fail */ in sem_open() 846 case KERN_INVALID_ADDRESS: in sem_wait_nocancel() 908 case KERN_INVALID_ADDRESS: in sem_trywait() 968 case KERN_INVALID_ADDRESS: in sem_post() 1045 case KERN_INVALID_ADDRESS: in psem_delete()
|
| H A D | kern_mman.c | 829 case KERN_INVALID_ADDRESS: in mmap() 933 case KERN_INVALID_ADDRESS: /* hole in region being sync'ed */ in msync_nocancel() 1066 case KERN_INVALID_ADDRESS: in mprotect() 1085 case KERN_INVALID_ADDRESS: in mprotect() 1209 case KERN_INVALID_ADDRESS: in madvise()
|
| H A D | posix_shm.c | 661 case KERN_INVALID_ADDRESS: in pshm_truncate() 936 case KERN_INVALID_ADDRESS: in pshm_mmap()
|
| H A D | kern_core.c | 447 kret = KERN_INVALID_ADDRESS; in coredump()
|
| H A D | sysv_shm.c | 562 case KERN_INVALID_ADDRESS: in shmat() 884 case KERN_INVALID_ADDRESS: in shmget_allocate_segment()
|
| /xnu-8019.80.24/bsd/uxkern/ |
| H A D | ux_exception.c | 69 if (code == KERN_INVALID_ADDRESS) { in ux_exception()
|
| /xnu-8019.80.24/osfmk/tests/ |
| H A D | ptrauth_data_tests.c | 59 kr = KERN_INVALID_ADDRESS; \
|
| /xnu-8019.80.24/osfmk/kern/ |
| H A D | kext_alloc.c | 224 rval = KERN_INVALID_ADDRESS; in kext_alloc()
|
| /xnu-8019.80.24/osfmk/arm/ |
| H A D | genassym.c | 198 DECLARE("KERN_INVALID_ADDRESS", KERN_INVALID_ADDRESS); in main()
|
| H A D | pmap.h | 569 #define pmap_attribute(pmap, addr, size, attr, value) (KERN_INVALID_ADDRESS)
|
| /xnu-8019.80.24/tests/ |
| H A D | vm_test_mach_map.c | 1012 if (kr == KERN_INVALID_ADDRESS) { 1333 if (kr == KERN_INVALID_ADDRESS) { 1651 if (kr == KERN_INVALID_ADDRESS) { 1676 kr == KERN_INVALID_ADDRESS) {
|
| /xnu-8019.80.24/bsd/dev/arm64/ |
| H A D | fasttrap_isa.c | 289 t->uu_exception = KERN_INVALID_ADDRESS; /* SIGSEGV */
|
| /xnu-8019.80.24/bsd/dev/i386/ |
| H A D | unix_signal.c | 573 case KERN_INVALID_ADDRESS: in sendsig()
|