Lines Matching refs:assertf
715 assertf(!new->use_pmap, "old %p new %p\n", old, new); in vm_map_entry_copy()
1218 assertf(kr == KERN_SUCCESS, in vm_map_apple_protected()
1220 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
2099 assertf(!pgz_owned(address), in vm_map_lookup_entry()
2895 assertf(vmk_flags.__vmkf_unused2 == 0, "vmk_flags unused2=0x%llx\n", vmk_flags.__vmkf_unused2); in vm_map_enter()
3129 assertf(VM_MAP_PAGE_ALIGNED(*address, FOURK_PAGE_MASK), "0x%llx", (uint64_t)*address); in vm_map_enter()
3130 assertf(VM_MAP_PAGE_ALIGNED(size, FOURK_PAGE_MASK), "0x%llx", (uint64_t)size); in vm_map_enter()
3132 assertf(page_aligned(*address), "0x%llx", (uint64_t)*address); in vm_map_enter()
3133 assertf(page_aligned(size), "0x%llx", (uint64_t)size); in vm_map_enter()
4093 assertf(vmk_flags.__vmkf_unused2 == 0, "vmk_flags unused2=0x%llx\n", vmk_flags.__vmkf_unused2); in vm_map_enter_mem_object()
4855 assertf(FALSE, "kernel_prefault && !UPL_VALID_PAGE"); in vm_map_enter_mem_object()
7003 …assertf(vm_object_round_page(VME_OFFSET(entry) + size) - vm_object_trunc_page(VME_OFFSET(entry)) =… in vm_map_wire_nested()
7011 assertf(os_ref_get_count_raw(&object->ref_count) == 1, in vm_map_wire_nested()
7014 assertf(!entry->needs_copy, in vm_map_wire_nested()
8791 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_delete()
10418 assertf(copy->size == copy_size, in vm_map_copy_overwrite()
12537 assertf(new_entry->use_pmap, "src_map %p new_entry %p\n", src_map, new_entry); in vm_map_copyin_internal()
13460 assertf(pre_nested_start <= pre_nested_end, in vm_map_fork_unnest()
13463 assertf(start <= end, in vm_map_fork_unnest()
13495 assertf(kr == KERN_SUCCESS, in vm_map_fork_unnest()
14185 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_lookup_and_lock_object()
14253 assertf(VM_MAP_PAGE_ALIGNED(local_vaddr, fault_page_mask), in vm_map_lookup_and_lock_object()
14376 assertf(VM_MAP_PAGE_ALIGNED(VME_OFFSET(submap_entry), VM_MAP_PAGE_MASK(map)), in vm_map_lookup_and_lock_object()
17464 …assertf(!os_add_overflow(end - start, offset, &end_offset), "size 0x%llx, offset 0x%llx caused ove… in vm_map_entry_insert()
18077 assertf(!src_entry->use_pmap, in vm_map_remap_extract()
18091 assertf(src_entry->use_pmap, in vm_map_remap_extract()
18384 assertf(new_entry->use_pmap, "map %p new_entry %p\n", map, new_entry); in vm_map_remap_extract()
18437 assertf(new_entry->use_pmap, "map %p new_entry %p\n", map, new_entry); in vm_map_remap_extract()
18997 assertf(target_size == src_copy_map_size - *trimmed_start_p - trimmed_end, in vm_map_copy_adjust_to_target()
20201 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20225 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20254 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20284 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20304 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20309 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20321 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20597 assertf(VM_MAP_PAGE_SHIFT(sub_map) >= VM_MAP_PAGE_SHIFT(map), in vm_map_page_range_info_internal()
23715 assertf(kmr.kmr_return == KERN_SUCCESS, in vm_map_corpse_footprint_collect_done()