Lines Matching refs:assertf
714 assertf(!new->use_pmap, "old %p new %p\n", old, new); in vm_map_entry_copy()
1217 assertf(kr == KERN_SUCCESS, in vm_map_apple_protected()
1219 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
2097 assertf(!pgz_owned(address), in vm_map_lookup_entry()
2899 assertf(vmk_flags.__vmkf_unused2 == 0, "vmk_flags unused2=0x%llx\n", vmk_flags.__vmkf_unused2); in vm_map_enter()
3133 assertf(VM_MAP_PAGE_ALIGNED(*address, FOURK_PAGE_MASK), "0x%llx", (uint64_t)*address); in vm_map_enter()
3134 assertf(VM_MAP_PAGE_ALIGNED(size, FOURK_PAGE_MASK), "0x%llx", (uint64_t)size); in vm_map_enter()
3136 assertf(page_aligned(*address), "0x%llx", (uint64_t)*address); in vm_map_enter()
3137 assertf(page_aligned(size), "0x%llx", (uint64_t)size); in vm_map_enter()
4097 assertf(vmk_flags.__vmkf_unused2 == 0, "vmk_flags unused2=0x%llx\n", vmk_flags.__vmkf_unused2); in vm_map_enter_mem_object()
4859 assertf(FALSE, "kernel_prefault && !UPL_VALID_PAGE"); in vm_map_enter_mem_object()
7002 …assertf(vm_object_round_page(VME_OFFSET(entry) + size) - vm_object_trunc_page(VME_OFFSET(entry)) =… in vm_map_wire_nested()
7010 assertf(os_ref_get_count_raw(&object->ref_count) == 1, in vm_map_wire_nested()
7013 assertf(!entry->needs_copy, in vm_map_wire_nested()
8786 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_delete()
10412 assertf(copy->size == copy_size, in vm_map_copy_overwrite()
12530 assertf(new_entry->use_pmap, "src_map %p new_entry %p\n", src_map, new_entry); in vm_map_copyin_internal()
13453 assertf(pre_nested_start <= pre_nested_end, in vm_map_fork_unnest()
13456 assertf(start <= end, in vm_map_fork_unnest()
13488 assertf(kr == KERN_SUCCESS, in vm_map_fork_unnest()
14178 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_lookup_and_lock_object()
14246 assertf(VM_MAP_PAGE_ALIGNED(local_vaddr, fault_page_mask), in vm_map_lookup_and_lock_object()
14369 assertf(VM_MAP_PAGE_ALIGNED(VME_OFFSET(submap_entry), VM_MAP_PAGE_MASK(map)), in vm_map_lookup_and_lock_object()
17453 …assertf(!os_add_overflow(end - start, offset, &end_offset), "size 0x%llx, offset 0x%llx caused ove… in vm_map_entry_insert()
18066 assertf(!src_entry->use_pmap, in vm_map_remap_extract()
18080 assertf(src_entry->use_pmap, in vm_map_remap_extract()
18373 assertf(new_entry->use_pmap, "map %p new_entry %p\n", map, new_entry); in vm_map_remap_extract()
18426 assertf(new_entry->use_pmap, "map %p new_entry %p\n", map, new_entry); in vm_map_remap_extract()
18986 assertf(target_size == src_copy_map_size - *trimmed_start_p - trimmed_end, in vm_map_copy_adjust_to_target()
20190 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20214 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20243 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20273 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20293 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20298 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20310 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20588 assertf(VM_MAP_PAGE_SHIFT(sub_map) >= VM_MAP_PAGE_SHIFT(map), in vm_map_page_range_info_internal()
23705 assertf(kmr.kmr_return == KERN_SUCCESS, in vm_map_corpse_footprint_collect_done()