| /xnu-12377.81.4/tests/vm/ |
| H A D | vm_ranges.c | 140 T_LOG("checking if %llx <= %llx <= %llx", range.min_address, addr, in assert_in_range() 142 T_EXPECT_GE(addr, range.min_address, "allocation above min address"); in assert_in_range() 260 assert_allocate_in_range(HEAP, range.min_address - ALLOCATION_SIZE, RANGE_HEAP_FLAGS); 296 if (def.max_address + 3 * ALLOCATION_SIZE <= heap.min_address) { in ensure_rogue_fixed_fails() 297 addr = heap.min_address - 2 * ALLOCATION_SIZE; in ensure_rogue_fixed_fails() 373 mach_vm_address_t target = range.min_address - ALLOCATION_SIZE; 481 T_EXPECT_GT(heap.min_address, def.max_address, in ensure_range() 485 T_EXPECT_EQ(heap.min_address, in ensure_range() 486 heap.min_address & (unsigned long)~0x1FFFFF, in ensure_range() 496 T_QUIET; T_EXPECT_EQ(def.min_address, parent_default.min_address, in ensure_child_range() [all …]
|
| /xnu-12377.81.4/osfmk/kern/ |
| H A D | thread_test_context.c | 158 if (thread_get_test_option(ttc_testing_ttc_struct).min_address != 0) { in thread_test_context_tests() 163 if (thread_get_test_option(ttc_testing_ttc_struct).min_address != 0) { in thread_test_context_tests() 181 if (thread_get_test_option(ttc_testing_ttc_struct).min_address != 33) { in thread_test_context_tests() 185 if (thread_get_test_option(ttc_testing_ttc_struct).min_address != 55) { in thread_test_context_tests()
|
| H A D | zalloc.c | 1455 r.min_address = addr; in zone_kmem_suballoc() 1560 return (struct zone_bits_allocator_header *)zone_info.zi_bits_range.min_address; in zba_base_header() 1599 bits_addr = zone_info.zi_bits_range.min_address + n * bits_size; in zba_populate() 1600 xtra_addr = zone_info.zi_xtra_range.min_address + n * xtra_size; in zba_populate() 1928 vm_offset_t base = zone_info.zi_xtra_range.min_address; in zba_extra_ref_ptr() 6611 (void *)zone_info.zi_ro_range.min_address, in zalloc_ro_mut_validation_panic() 7757 (void *)zone_info.zi_map_range.min_address, in panic_display_zone_info() 7774 (void *)zone_info.zi_meta_range.min_address, in panic_display_zone_info() 7776 (void *)zone_info.zi_bits_range.min_address, in panic_display_zone_info() 7778 (void *)zone_info.zi_xtra_range.min_address, in panic_display_zone_info() [all …]
|
| H A D | task.c | 6150 vm_info->min_address = map->min_offset; in task_info()
|
| /xnu-12377.81.4/bsd/dev/ |
| H A D | unix_startup.c | 157 &bufferhdr_range.min_address, in bsd_startupearly() 165 &(vm_offset_t){ bufferhdr_range.min_address }, in bsd_startupearly() 170 buf_headers = (struct buf *)bufferhdr_range.min_address; in bsd_startupearly() 214 &mb_range.min_address, in bsd_bufferinit() 220 mbutl = (unsigned char *)mb_range.min_address; in bsd_bufferinit()
|
| /xnu-12377.81.4/tests/ |
| H A D | task_info.c | 80 vm_info.min_address = CANARY; 96 T_EXPECT_EQ(vm_info.min_address, CANARY, 99 vm_info.min_address, CANARY); 112 vm_info.min_address = CANARY; 128 T_EXPECT_EQ(vm_info.min_address, CANARY, 131 vm_info.min_address, CANARY); 144 vm_info.min_address = CANARY; 160 T_EXPECT_NE(vm_info.min_address, CANARY, 163 vm_info.min_address, CANARY); 176 vm_info.min_address = CANARY; [all …]
|
| /xnu-12377.81.4/osfmk/ipc/ |
| H A D | ipc_init.c | 153 ipc_kernel_map = kmem_suballoc(kernel_map, &ipc_kernel_range.min_address, in ipc_init() 158 ipc_kernel_copy_map = kmem_suballoc(kernel_map, &ipc_kernel_copy_range.min_address, in ipc_init()
|
| /xnu-12377.81.4/osfmk/vm/ |
| H A D | vm_kern.c | 267 ({ (rmin) = (r)->min_address; (rmax) = (r)->max_address; }) 282 mach_vm_offset_t min_address, in __mach_vm_range_invalid() argument 286 min_address, max_address); in __mach_vm_range_invalid() 343 r2_min = r2->min_address; in mach_vm_range_intersects() 366 r2.min_address = addr; in mach_vm_range_intersects() 2820 *range_start = kmem_ranges[range_id].min_address; in kmem_addr_to_meta() 2868 vm_map_offset_t addr = kmem_meta_range[i].min_address; in kmem_metadata_init() 2879 kmem_meta_range[i].min_address = addr; in kmem_metadata_init() 2882 meta = (struct kmem_page_meta *) kmem_meta_range[i].min_address; in kmem_metadata_init() 2980 (void *)slot->min_address, (void *)slot->max_address, in kmem_invalid_slot_for_addr() [all …]
|
| H A D | vm_user.c | 2066 if (r1->range.min_address != r2->range.min_address) { in mach_vm_range_recipe_v1_cmp() 2067 return r1->range.min_address < r2->range.min_address ? -1 : 1; in mach_vm_range_recipe_v1_cmp() 2146 .min_address = map->default_range.max_address, in mach_vm_range_create_v1() 2147 .max_address = map->data_range.min_address, in mach_vm_range_create_v1() 2150 .min_address = map->data_range.max_address, in mach_vm_range_create_v1() 2191 if (!mach_vm_range_contains(&void1, r->min_address, s) && in mach_vm_range_create_v1() 2192 !mach_vm_range_contains(&void2, r->min_address, s)) { in mach_vm_range_create_v1() 2198 recipe[i].range.min_address) { in mach_vm_range_create_v1() 2224 } else if (r2->vmur_max_address <= r1->min_address) { in mach_vm_range_create_v1() 2248 .vmur_min_address = recipe[i].range.min_address, in mach_vm_range_create_v1() [all …]
|
| H A D | vm_map.c | 2313 if (map->large_file_range.min_address != map->large_file_range.max_address) { in vm_map_get_range() 2346 effective_range.min_address = MAX(map->min_offset, VM_MAP_PAGE_SIZE(map)); in vm_map_get_range() 2429 if (hint <= effective_range.min_address) { in vm_map_locate_space_anywhere() 2432 limit = effective_range.min_address; in vm_map_locate_space_anywhere() 2462 if (hint < effective_range.min_address) { in vm_map_locate_space_anywhere() 2463 hint = effective_range.min_address; in vm_map_locate_space_anywhere() 2479 effective_range.max_address - effective_range.min_address) { in vm_map_locate_space_anywhere() 2591 effective_min_offset = r.min_address; in vm_map_locate_space_fixed() 2599 effective_min_offset = r->min_address; in vm_map_locate_space_fixed() 2881 addr_space_size = effective_range.max_address - effective_range.min_address; in vm_map_random_address_for_size() [all …]
|
| H A D | vm_compressor.c | 1126 compressor_map = kmem_suballoc(kernel_map, &compressor_range.min_address, in vm_compressor_init()
|
| /xnu-12377.81.4/osfmk/i386/ |
| H A D | io_map.c | 78 io_submap = kmem_suballoc(kernel_map, &io_range.min_address, IO_MAP_SIZE, in io_map_init()
|
| /xnu-12377.81.4/osfmk/arm/ |
| H A D | io_map.c | 74 io_submap = kmem_suballoc(kernel_map, &io_range.min_address, IO_MAP_SIZE, in io_map_init()
|
| /xnu-12377.81.4/osfmk/mach/ |
| H A D | vm_types.h | 164 mach_vm_offset_t min_address; member
|
| H A D | task_info.h | 391 mach_vm_address_t min_address; member
|
| H A D | vm_param.h | 711 vm_offset_t min_address,
|
| /xnu-12377.81.4/iokit/Kernel/ |
| H A D | IOLib.cpp | 194 &gIOKitPageableFixedRange.min_address, in IOLibInit() 201 gIOKitPageableMap.address = gIOKitPageableFixedRange.min_address; in IOLibInit()
|
| /xnu-12377.81.4/bsd/kern/ |
| H A D | bsd_init.c | 677 &bsd_pageable_range.min_address, in bsd_init()
|
| /xnu-12377.81.4/osfmk/tests/ |
| H A D | vm_parameter_validation.h | 4546 .min_address = map->default_range.max_address, in test_mach_vm_range_create() 4547 .max_address = map->data_range.min_address, in test_mach_vm_range_create() 4550 .min_address = map->data_range.max_address, in test_mach_vm_range_create() 4562 if (void1.min_address + (PAGE_SIZE * 6) < void1.max_address) { in test_mach_vm_range_create() 4565 } else if (void2.min_address + (PAGE_SIZE * 6) < void2.max_address) { in test_mach_vm_range_create() 4570 mach_vm_address_t addr_base = range_to_test.min_address + PAGE_SIZE * 2; in test_mach_vm_range_create()
|
| /xnu-12377.81.4/tools/lldbmacros/ |
| H A D | memory.py | 1588 start_vaddr = range.min_address 1696 min_address = vmrange.min_address 1698 range_size = max_address - min_address 1699 out_string += format_string.format(range_name, min_address, max_address, range_size)
|
| H A D | kasan.py | 419 if addr < unsigned(ranges[i].min_address):
|
| /xnu-12377.81.4/osfmk/x86_64/ |
| H A D | pmap.c | 836 pmap_struct_map = kmem_suballoc(kernel_map, &pmap_struct_range.min_address, in pmap_init()
|
| /xnu-12377.81.4/bsd/vm/ |
| H A D | vm_unix.c | 3627 r1.max_address, r2.min_address,
|