Lines Matching refs:assertf

683 		assertf(!new->use_pmap, "old %p new %p\n", old, new);  in vm_map_entry_copy()
1186 assertf(kr == KERN_SUCCESS, in vm_map_apple_protected()
1188 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
2024 assertf(!pgz_owned(address), in vm_map_lookup_entry()
2811 assertf(vmk_flags.__vmkf_unused == 0, "vmk_flags unused=0x%x\n", vmk_flags.__vmkf_unused); in vm_map_enter()
3045 assertf(VM_MAP_PAGE_ALIGNED(*address, FOURK_PAGE_MASK), "0x%llx", (uint64_t)*address); in vm_map_enter()
3046 assertf(VM_MAP_PAGE_ALIGNED(size, FOURK_PAGE_MASK), "0x%llx", (uint64_t)size); in vm_map_enter()
3048 assertf(page_aligned(*address), "0x%llx", (uint64_t)*address); in vm_map_enter()
3049 assertf(page_aligned(size), "0x%llx", (uint64_t)size); in vm_map_enter()
3996 assertf(vmk_flags.__vmkf_unused == 0, "vmk_flags unused=0x%x\n", vmk_flags.__vmkf_unused); in vm_map_enter_mem_object()
4750 assertf(FALSE, "kernel_prefault && !UPL_VALID_PAGE"); in vm_map_enter_mem_object()
6769assertf(vm_object_round_page(VME_OFFSET(entry) + size) - vm_object_trunc_page(VME_OFFSET(entry)) =… in vm_map_wire_nested()
6777 assertf(object->ref_count == 1, in vm_map_wire_nested()
6780 assertf(!entry->needs_copy, in vm_map_wire_nested()
8471 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_delete()
10018 assertf(copy->size == copy_size, in vm_map_copy_overwrite()
12109 assertf(new_entry->use_pmap, "src_map %p new_entry %p\n", src_map, new_entry); in vm_map_copyin_internal()
13032 assertf(pre_nested_start <= pre_nested_end, in vm_map_fork_unnest()
13035 assertf(start <= end, in vm_map_fork_unnest()
13067 assertf(kr == KERN_SUCCESS, in vm_map_fork_unnest()
13733 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_lookup_and_lock_object()
13801 assertf(VM_MAP_PAGE_ALIGNED(local_vaddr, fault_page_mask), in vm_map_lookup_and_lock_object()
13924 assertf(VM_MAP_PAGE_ALIGNED(VME_OFFSET(submap_entry), VM_MAP_PAGE_MASK(map)), in vm_map_lookup_and_lock_object()
17030assertf(!os_add_overflow(end - start, offset, &end_offset), "size 0x%llx, offset 0x%llx caused ove… in vm_map_entry_insert()
17649 assertf(!src_entry->use_pmap, in vm_map_remap_extract()
17663 assertf(src_entry->use_pmap, in vm_map_remap_extract()
17934 assertf(new_entry->use_pmap, "map %p new_entry %p\n", map, new_entry); in vm_map_remap_extract()
17987 assertf(new_entry->use_pmap, "map %p new_entry %p\n", map, new_entry); in vm_map_remap_extract()
18548 assertf(target_size == src_copy_map_size - *trimmed_start_p - trimmed_end, in vm_map_copy_adjust_to_target()
19650 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
19674 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
19703 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
19733 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
19753 assertf(!map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
19758 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
19770 assertf(map_entry->use_pmap, "offset 0x%llx map_entry %p", (uint64_t) curr_s_offset, map_entry); in vm_map_footprint_query_page_info()
20035 assertf(VM_MAP_PAGE_SHIFT(sub_map) >= VM_MAP_PAGE_SHIFT(map), in vm_map_page_range_info_internal()
23110 assertf(kr == KERN_SUCCESS, in vm_map_corpse_footprint_collect_done()
23121 assertf(kr == KERN_SUCCESS, in vm_map_corpse_footprint_collect_done()
23261 assertf(kr == KERN_SUCCESS, "kr=0x%x\n", kr); in vm_map_corpse_footprint_destroy()