| /xnu-12377.1.9/osfmk/vm/ |
| H A D | vm_tests.c | 563 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 564 …assertf(alloc1_addr == expected_addr, "alloc1_addr = 0x%lx expected 0x%lx", alloc1_addr, expected_… in vm_test_4k() 570 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 581 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 582 …assertf(alloc1_addr == expected_addr, "alloc1_addr = 0x%lx expected 0x%lx", alloc1_addr, expected_… in vm_test_4k() 593 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 594 …assertf(alloc2_addr == expected_addr, "alloc2_addr = 0x%lx expected 0x%lx", alloc2_addr, expected_… in vm_test_4k() 605 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 606 …assertf(alloc3_addr == expected_addr, "alloc3_addr = 0x%lx expected 0x%lx\n", alloc3_addr, expecte… in vm_test_4k() 617 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() [all …]
|
| H A D | vm_page_internal.h | 150 assertf(fault_phys_offset < PAGE_SIZE && in VMP_CS_FOR_OFFSET() 161 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_VALIDATED() 175 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_TAINTED() 189 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_NX() 204 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_VALIDATED() 226 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_TAINTED() 248 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_NX()
|
| H A D | vm_compressor_algorithms.c | 402 assertf(rval == PAGE_SIZE, "LZ4 decode: size != pgsize %d, header: 0x%x, 0x%x, 0x%x", in metadecompressor() 461 assertf(((new_codec == VM_COMPRESSOR_DEFAULT_CODEC) || (new_codec == CMODE_WK) || in vm_compressor_algorithm_init()
|
| H A D | vm_map.c | 739 assertf(!new->use_pmap, "old %p new %p\n", old, new); in vm_map_entry_copy() 1262 assertf(kr == KERN_SUCCESS, in vm_map_apple_protected() 1264 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected() 2941 assertf(vmk_flags.__vmkf_unused2 == 0, "vmk_flags unused2=0x%llx\n", vmk_flags.__vmkf_unused2); in vm_map_enter() 3189 assertf(VM_MAP_PAGE_ALIGNED(*address, VM_MAP_PAGE_MASK(map)), "0x%llx", (uint64_t)*address); in vm_map_enter() 3190 assertf(VM_MAP_PAGE_ALIGNED(size, VM_MAP_PAGE_MASK(map)), "0x%llx", (uint64_t)size); in vm_map_enter() 4139 assertf(vmk_flags.__vmkf_unused2 == 0, "vmk_flags unused2=0x%llx\n", vmk_flags.__vmkf_unused2); in vm_map_enter_mem_object() 4923 assertf(FALSE, "kernel_prefault && !UPL_VALID_PAGE"); in vm_map_enter_mem_object() 4934 assertf(p, "offset 0x%llx: no VM page", page_offset); in vm_map_enter_mem_object() 4950 assertf(VM_PAGE_GET_PHYS_PAGE(p) == UPL_PHYS_PAGE(page_list, i), in vm_map_enter_mem_object() [all …]
|
| H A D | vm_map_store.c | 146 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_store_entry_link() 215 assertf(!entry->vme_permanent, in _vm_map_store_entry_unlink()
|
| H A D | vm_fault.c | 2674 assertf(fault_phys_offset == 0, in vm_fault_cs_need_validation() 3661 assertf((!(fault_phys_offset & FOURK_PAGE_MASK) && in vm_fault_attempt_pmap_enter() 3665 assertf(fault_phys_offset == 0, in vm_fault_attempt_pmap_enter() 4094 assertf(VM_PAGE_OBJECT(m) != VM_OBJECT_NULL, "m=%p", m); in vm_fault_enter() 5202 assertf((!(fault_phys_offset & FOURK_PAGE_MASK) && in vm_fault_internal() 5206 assertf(fault_phys_offset == 0, in vm_fault_internal() 5220 assertf(VM_PAGE_OBJECT(m) == m_object, "m=%p m_object=%p object=%p", m, m_object, object); in vm_fault_internal() 5874 assertf(!((fault_type & VM_PROT_WRITE) && object->vo_copy), in vm_fault_internal() 5918 assertf(VM_PAGE_OBJECT(m) == object, "m=%p object=%p", m, object); in vm_fault_internal() 6019 assertf(!((prot & VM_PROT_WRITE) && object->vo_copy), in vm_fault_internal() [all …]
|
| H A D | vm_compressor.c | 895 …assertf((tmp_slot_ptr.s_cseg == c_segments_limit), "vm_compressor_init: overflowed s_cseg field in… in vm_compressor_set_size() 935 …assertf((tmp_slot_ptr.s_cseg == c_segments_limit), "vm_compressor_init: freezer reserve overflowed… in vm_compressor_set_size() 2394 …assertf(c_seg_dst->c_has_donated_pages == c_seg_src->c_has_donated_pages, "Mismatched donation sta… in c_seg_major_compact() 2421 …assertf(c_seg_dst->c_populated_offset >= c_seg_dst->c_nextoffset, "Unexpected segment offsets: %u,… in c_seg_major_compact() 4911 assertf(((c_size <= max_csize_adj) && (c_size >= -1)), in c_compress_page() 5470 …assertf(c_segment_pages_compressed_incore >= 0, "-ve incore count %p 0x%x", c_seg, c_segment_pages… in c_decompress_page() 5473 …assertf(c_segment_pages_compressed_incore_late_swapout >= 0, "-ve lateswapout count %p 0x%x", c_se… in c_decompress_page() 6350 assertf(vaddr < maxvaddr, "0x%llx 0x%llx", vaddr, maxvaddr); in vm_uncompressed_return_space_to_swap() 6352 assertf((uncompressed_file0_space_bitmap[chunkidx] & ((uint64_t)1 << chunkoffset)), in vm_uncompressed_return_space_to_swap() 6355 assertf(!(uncompressed_file0_space_bitmap[chunkidx] & ((uint64_t)1 << chunkoffset)), in vm_uncompressed_return_space_to_swap() [all …]
|
| H A D | vm_kern.c | 2156 assertf(!VM_PAGE_PAGEABLE(mem), in kmem_realloc_guard() 2162 assertf(VM_PAGE_WIRED(mem), in kmem_realloc_guard() 2167 assertf(mem->vmp_wire_count >= 1, in kmem_realloc_guard() 2269 assertf(!VM_PAGE_PAGEABLE(mem), in kmem_realloc_guard() 2275 assertf(VM_PAGE_WIRED(mem), in kmem_realloc_guard() 2280 assertf(mem->vmp_wire_count >= 2, in kmem_realloc_guard() 4776 assertf(can_write_at(offs, page), \ 4780 assertf(!can_write_at(offs, page), \ 4840 assertf(addr != 0ull, "kma(%p, 10p, 0, KO | GF | GL)", map); in kmem_alloc_basic_test() 4845 assertf(e, "unable to find address %p in map %p", (void *)addr, map); in kmem_alloc_basic_test() [all …]
|
| H A D | vm_object.c | 3037 assertf(!(offset & (pmap_page_size - 1)) && !(size & (pmap_page_size - 1)), 3825 assertf(page_aligned(copy_size), 3877 assertf(page_aligned(copy_size), 4216 assertf(page_aligned(result->vo_shadow_offset), 4787 assertf(page_aligned(object->vo_shadow_offset), 4790 assertf(page_aligned(backing_object->vo_shadow_offset), 4864 assertf(page_aligned(object->vo_shadow_offset), 4867 assertf(page_aligned(backing_object->vo_shadow_offset), 4981 assertf(page_aligned(hint_offset), "hint_offset 0x%llx", hint_offset); 5562 assertf(page_aligned(newsize), [all …]
|
| /xnu-12377.1.9/iokit/Kernel/ |
| H A D | IOPerfControl.cpp | 96 …assertf(workTableLength <= kWorkTableMaxSize, "%zu exceeds max allowed capacity of %zu", workTable… in init() 211 …assertf(false, "Unexpected device type for IOPerfControlClient::accountResources: %llu", static_ca… in accountResources() 276 …assertf(token != kIOPerfControlClientWorkUntracked, "Attempt to deallocate token kIOPerfControlCli… in deallocateToken() 277 …assertf(token <= workTableLength, "Attempt to deallocate token %llu which is greater than the tabl… in deallocateToken() 309 assertf(entry->thread_group, "Invalid work token: %llu", token); in getEntryForToken() 494 assertf(!entry->started, "Work for token %llu was already started", token); in workBegin() 588 …assertf(thread_group == nullptr, "IOPerfControlWorkContext ID %llu being released without calling … in free() 589 …assertf(coal == nullptr, "IOPerfControlWorkContext ID %llu being released without calling workEnd!… in free() 653 …assertf(!work_context->started, "IOPerfControlWorkContext ID %llu was already started", work_conte… in workSubmitWithContext() 654 …assertf(work_context->thread_group == nullptr, "IOPerfControlWorkContext ID %llu has already taken… in workSubmitWithContext() [all …]
|
| /xnu-12377.1.9/osfmk/kern/ |
| H A D | lock_rw.c | 450 assertf(entry->rwlde_mode_count == 1, in change_held_rwlock_slow() 459 assertf(entry->rwlde_mode_count == -1, in change_held_rwlock_slow() 544 assertf(entry->rwlde_mode_count != INT8_MAX, in add_held_rwlock_slow() 914 assertf(lock->lck_rw_owner != self->ctid, 1117 assertf((get_preemption_level() == 0 && ml_get_interrupts_enabled()) || in lck_rw_lock_check_preemption() 1168 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", in lck_rw_lock_exclusive_check_contended() 1210 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", 1310 assertf(lck->lck_rw_owner != self->ctid, 1478 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", 1733 assertf(lock->lck_rw_priv_excl != 0, "lock %p thread %p", lock, current_thread()); in lck_rw_lock_shared_to_exclusive() [all …]
|
| H A D | assert.h | 200 #define assertf(ex, fmt, args...) ({ \ macro 259 #define assertf(ex, fmt, args...) ((void)0) macro
|
| H A D | smr.h | 125 assertf(held_cond, "smr_serialized_load: lock not held"); \ 169 assertf(held_cond, "smr_serialized_store: lock not held"); \ 201 assertf(held_cond, "smr_serialized_store_relaxed: lock not held"); \ 231 assertf(held_cond, "smr_serialized_store: lock not held"); \
|
| H A D | exclaves_boot.c | 440 assertf(false, "unknown boot status %u", boot_status); in exclaves_get_boot_status_string()
|
| /xnu-12377.1.9/osfmk/arm64/ |
| H A D | lowmem_vectors.c | 113 assertf((vm_first_phys != 0) && (vm_last_phys != 0), in patch_low_glo_static_region() 139 assertf((physmap_base != 0) && (physmap_end != 0), in patch_low_glo_static_region()
|
| H A D | bsd_arm64.c | 299 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_syscall()
|
| H A D | hibernate_arm64.c | 280 assertf(SPTMArgs->hib_metadata->iboot_loaded_ranges != NULL, in hibernate_page_list_set_volatile()
|
| /xnu-12377.1.9/osfmk/i386/ |
| H A D | bsd_i386.c | 373 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in machdep_syscall() 447 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in machdep_syscall64() 604 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_call_munger() 721 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_call_munger64()
|
| /xnu-12377.1.9/bsd/dev/i386/ |
| H A D | systemcalls.c | 244 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall() 453 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall64() 587 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall_return()
|
| /xnu-12377.1.9/security/ |
| H A D | mac_label.c | 206 assertf(label->l_owner != (struct label **)-1, in mac_label_set()
|
| /xnu-12377.1.9/bsd/dev/arm/ |
| H A D | systemcalls.c | 204 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall() 259 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall_return()
|
| /xnu-12377.1.9/osfmk/arm/pmap/ |
| H A D | pmap_internal.h | 186 #define ASSERT_NOT_HIBERNATING() (assertf(!hib_entry_pmap_lockdown, \
|
| /xnu-12377.1.9/osfmk/arm64/sptm/pmap/ |
| H A D | pmap_data.h | 295 assertf(os_atomic_load(&pv_head_table[index], relaxed) & PVH_LOCK_FLAGS, in pvh_assert_locked() 353 assertf(wres == THREAD_WAITING, "%s: unexpected wait result %d", __func__, wres); in pvh_lock() 948 assertf(pvh_pve_list(locked_pvh->pvh) == pvep, "%s: pvh %p != pvep %p", in pve_remove()
|
| /xnu-12377.1.9/tests/sched/sched_test_harness/shadow_headers/ |
| H A D | sched_prim.c | 418 assertf(false, "unimplemented"); in thread_setrun()
|
| /xnu-12377.1.9/iokit/Tests/ |
| H A D | TestIOMemoryDescriptor.cpp | 220 assertf((0x53535300 | dir) == data, "mismatch 0x%x", data); in IODMACommandForceDoubleBufferTest() 234 assertf((0x11223300 | dir) == data, "mismatch 0x%x", data); in IODMACommandForceDoubleBufferTest() 316 …assertf(segments[0].fIOVMAddr != segPhys, "phys !local 0x%qx, 0x%qx, %p", segments[0].fIOVMAddr, s… in IODMACommandLocalMappedNonContig()
|