Lines Matching refs:ALLOCATION_SIZE

18 #define ALLOCATION_SIZE (PAGE_SIZE)  macro
104 int ret = mach_vm_allocate(mach_task_self(), &dst, ALLOCATION_SIZE, vm_flags); in assert_allocate()
129 void *ret = mmap(addr, ALLOCATION_SIZE, VM_PROT_READ | VM_PROT_WRITE, in assert_mmap()
142 ALLOCATION_SIZE, vm_flags), "vm_allocate"); in assert_allocate_eq()
224 assert_allocate_in_range(HEAP, range.min_address - ALLOCATION_SIZE, RANGE_HEAP_FLAGS);
241 vm_deallocate(mach_task_self(), addr, ALLOCATION_SIZE); in ensure_fixed_mappings_succeed_cross()
245 vm_deallocate(mach_task_self(), addr, ALLOCATION_SIZE); in ensure_fixed_mappings_succeed_cross()
249 vm_deallocate(mach_task_self(), addr, ALLOCATION_SIZE); in ensure_fixed_mappings_succeed_cross()
260 if (def.max_address + 3 * ALLOCATION_SIZE <= heap.min_address) { in ensure_rogue_fixed_fails()
261 addr = heap.min_address - 2 * ALLOCATION_SIZE; in ensure_rogue_fixed_fails()
268 addr = heap.max_address + ALLOCATION_SIZE; in ensure_rogue_fixed_fails()
272 ALLOCATION_SIZE, VM_FLAGS_FIXED); in ensure_rogue_fixed_fails()
311 T_ASSERT_POSIX_SUCCESS(ftruncate(fd, (off_t)ALLOCATION_SIZE), NULL);
337 mach_vm_address_t target = range.min_address - ALLOCATION_SIZE;
348 mach_vm_address_t target = range.max_address - 100 * ALLOCATION_SIZE;
379 int ret = mach_vm_map(mach_task_self(), &addr, ALLOCATION_SIZE,
395 mach_vm_address_t target = addr + ALLOCATION_SIZE;
397 int ret = mach_vm_remap(mach_task_self(), &target, ALLOCATION_SIZE,
412 mach_vm_address_t target = addr + ALLOCATION_SIZE;
414 int ret = mach_vm_remap(mach_task_self(), &target, ALLOCATION_SIZE,
432 int ret = mach_vm_remap(mach_task_self(), &target, ALLOCATION_SIZE,
591 kr = mach_vm_allocate(mach_task_self(), &addr, ALLOCATION_SIZE,