| /xnu-8792.41.9/osfmk/vm/ |
| H A D | vm_tests.c | 535 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 536 …assertf(alloc1_addr == expected_addr, "alloc1_addr = 0x%lx expected 0x%lx", alloc1_addr, expected_… in vm_test_4k() 542 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 553 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 554 …assertf(alloc1_addr == expected_addr, "alloc1_addr = 0x%lx expected 0x%lx", alloc1_addr, expected_… in vm_test_4k() 565 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 566 …assertf(alloc2_addr == expected_addr, "alloc2_addr = 0x%lx expected 0x%lx", alloc2_addr, expected_… in vm_test_4k() 577 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() 578 …assertf(alloc3_addr == expected_addr, "alloc3_addr = 0x%lx expected 0x%lx\n", alloc3_addr, expecte… in vm_test_4k() 589 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k() [all …]
|
| H A D | vm_page.h | 293 assertf(fault_phys_offset < PAGE_SIZE && in VMP_CS_FOR_OFFSET() 304 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_VALIDATED() 318 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_TAINTED() 332 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_NX() 347 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_VALIDATED() 369 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_TAINTED() 391 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_NX()
|
| H A D | vm_compressor_algorithms.c | 402 assertf(rval == PAGE_SIZE, "LZ4 decode: size != pgsize %d, header: 0x%x, 0x%x, 0x%x", in metadecompressor() 461 assertf(((new_codec == VM_COMPRESSOR_DEFAULT_CODEC) || (new_codec == CMODE_WK) || in vm_compressor_algorithm_init()
|
| H A D | vm_map_store.c | 160 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_store_entry_link() 195 …assertf(!entry->vme_permanent, "mapHdr %p entry %p [ 0x%llx end 0x%llx ] prot 0x%x/0x%x submap %d\… in _vm_map_store_entry_unlink()
|
| H A D | vm_fault.c | 2552 assertf(fault_phys_offset == 0, in vm_fault_cs_need_validation() 3376 assertf((!(fault_phys_offset & FOURK_PAGE_MASK) && in vm_fault_attempt_pmap_enter() 3380 assertf(fault_phys_offset == 0, in vm_fault_attempt_pmap_enter() 3737 assertf(VM_PAGE_OBJECT(m) != VM_OBJECT_NULL, "m=%p", m); in vm_fault_enter() 4719 assertf((!(fault_phys_offset & FOURK_PAGE_MASK) && in vm_fault_internal() 4723 assertf(fault_phys_offset == 0, in vm_fault_internal() 4727 assertf(VM_PAGE_OBJECT(m) == m_object, "m=%p m_object=%p object=%p", m, m_object, object); in vm_fault_internal() 5359 assertf(VM_PAGE_OBJECT(m) == object, "m=%p object=%p", m, object); in vm_fault_internal() 5624 assertf(object->internal, "object %p", object); in vm_fault_internal() 5856 assertf(VM_PAGE_OBJECT(m) == m_object, "m=%p m_object=%p", m, m_object); in vm_fault_internal() [all …]
|
| H A D | vm_object.c | 2686 assertf(!(offset & (pmap_page_size - 1)) && !(size & (pmap_page_size - 1)), in vm_object_pmap_protect_options() 3360 assertf(page_aligned(src_offset + size), in vm_object_copy_call() 3587 assertf(page_aligned(copy_size), in vm_object_copy_delayed() 3628 assertf(page_aligned(copy_size), in vm_object_copy_delayed() 3956 assertf(page_aligned(result->vo_shadow_offset), in vm_object_shadow() 4478 assertf(page_aligned(object->vo_shadow_offset), in vm_object_do_collapse() 4481 assertf(page_aligned(backing_object->vo_shadow_offset), in vm_object_do_collapse() 4555 assertf(page_aligned(object->vo_shadow_offset), in vm_object_do_bypass() 4558 assertf(page_aligned(backing_object->vo_shadow_offset), in vm_object_do_bypass() 4670 assertf(page_aligned(hint_offset), "hint_offset 0x%llx", hint_offset); in vm_object_collapse() [all …]
|
| H A D | vm_map.c | 498 assertf(!new->use_pmap, "old %p new %p\n", old, new); in vm_map_entry_copy() 1018 assertf(kr == KERN_SUCCESS, in vm_map_apple_protected() 1020 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected() 1838 assertf(!pgz_owned(address), in vm_map_lookup_entry() 2437 assertf(vmk_flags.__vmkf_unused == 0, "vmk_flags unused=0x%x\n", vmk_flags.__vmkf_unused); in vm_map_enter() 2596 assertf(VM_MAP_PAGE_ALIGNED(*address, FOURK_PAGE_MASK), "0x%llx", (uint64_t)*address); in vm_map_enter() 2597 assertf(VM_MAP_PAGE_ALIGNED(size, FOURK_PAGE_MASK), "0x%llx", (uint64_t)size); in vm_map_enter() 2599 assertf(page_aligned(*address), "0x%llx", (uint64_t)*address); in vm_map_enter() 2600 assertf(page_aligned(size), "0x%llx", (uint64_t)size); in vm_map_enter() 4011 assertf(vmk_flags.__vmkf_unused == 0, "vmk_flags unused=0x%x\n", vmk_flags.__vmkf_unused); in vm_map_enter_mem_object_helper() [all …]
|
| H A D | vm_kern.c | 2638 assertf(can_write_at(offs, page), \ 2642 assertf(!can_write_at(offs, page), \ 2668 assertf(addr != 0ull, "kma(%p, 10p, 0, KO | GF | GL)", map); in kmem_alloc_basic_test() 2673 assertf(e, "unable to find address %p in map %p", (void *)addr, map); in kmem_alloc_basic_test() 2694 assertf(addr != 0ull, "kma(%p, 10p, 0, KO | PG)", map); in kmem_alloc_basic_test()
|
| /xnu-8792.41.9/osfmk/kern/ |
| H A D | smr.h | 182 assertf(held_cond, "smr_serialized_load: lock not held"); \ 226 assertf(held_cond, "smr_serialized_store: lock not held"); \ 258 assertf(held_cond, "smr_serialized_store_relaxed: lock not held"); \ 288 assertf(held_cond, "smr_serialized_store: lock not held"); \
|
| H A D | assert.h | 98 #define assertf(ex, fmt, args...) \ macro 149 #define assertf(ex, fmt, args...) \ 156 #define assertf(ex, fmt, args...) ((void)0)
|
| H A D | lock_rw.c | 425 assertf(entry->rwlde_mode_count == 1, in change_held_rwlock() 434 assertf(entry->rwlde_mode_count == -1, in change_held_rwlock() 516 assertf(entry->rwlde_mode_count != INT8_MAX, in add_held_rwlock() 895 assertf(lock->lck_rw_owner != self->ctid, 1135 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", in lck_rw_lock_exclusive_check_contended() 1176 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", 1275 assertf(lck->lck_rw_owner != self->ctid, 1442 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", 1707 assertf(lock->lck_rw_priv_excl != 0, "lock %p thread %p", lock, current_thread()); in lck_rw_lock_shared_to_exclusive() 1742 assertf(lock->lck_rw_owner == 0, "state=0x%x, owner=%p", in lck_rw_lock_shared_to_exclusive() [all …]
|
| H A D | cpu_quiesce.c | 358 assertf(!cpu_quiescent_counter_needs_commit(orig_state), in cpu_quiescent_counter_checkin()
|
| H A D | recount.c | 189 assertf((new & 1) == 1, "invalid seqlock in exclusive lock"); in _seqlock_excl_lock_relaxed() 202 assertf((new & 1) == 0, "invalid seqlock in exclusive unlock"); in _seqlock_excl_unlock_relaxed() 1006 assertf(to == RCT_TOPO_CPU_KIND, in recount_convert_topo_index()
|
| H A D | remote_time.c | 280 assertf(recalculate_count <= MAX_RECALCULATE_COUNT, "bt_caliberation_thread: recalculate \ in bt_calibration_thread()
|
| H A D | smr.c | 249 assertf(smr_boot.used <= SMR_EARLY_COUNT, in __smr_init()
|
| /xnu-8792.41.9/iokit/Kernel/ |
| H A D | IOPerfControl.cpp | 93 …assertf(workTableLength <= kWorkTableMaxSize, "%zu exceeds max allowed capacity of %zu", workTable… in init() 225 …assertf(token != kIOPerfControlClientWorkUntracked, "Attempt to deallocate token kIOPerfControlCli… in deallocateToken() 226 …assertf(token <= workTableLength, "Attempt to deallocate token %llu which is greater than the tabl… in deallocateToken() 254 assertf(entry->thread_group, "Invalid work token: %llu", token); in getEntryForToken() 439 assertf(!entry->started, "Work for token %llu was already started", token); in workBegin() 522 …assertf(thread_group == nullptr, "IOPerfControlWorkContext ID %llu being released without calling … in free() 586 …assertf(!work_context->started, "IOPerfControlWorkContext ID %llu was already started", work_conte… in workSubmitWithContext() 587 …assertf(work_context->thread_group == nullptr, "IOPerfControlWorkContext ID %llu has already taken… in workSubmitWithContext() 655 …assertf(!work_context->started, "IOPerfControlWorkContext %llu was already started", work_context-… in workBeginWithContext()
|
| H A D | IOTimerEventSource.cpp | 287 assertf(func, "IOTimerEventSource options 0x%x", options); in setTimeoutFunc()
|
| /xnu-8792.41.9/osfmk/arm64/ |
| H A D | lowmem_vectors.c | 109 assertf((vm_first_phys != 0) && (vm_last_phys != 0), in patch_low_glo_static_region() 135 assertf((physmap_base != 0) && (physmap_end != 0), in patch_low_glo_static_region()
|
| H A D | bsd_arm64.c | 299 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_syscall()
|
| /xnu-8792.41.9/osfmk/i386/ |
| H A D | bsd_i386.c | 373 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in machdep_syscall() 447 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in machdep_syscall64() 604 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_call_munger() 721 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_call_munger64()
|
| H A D | fpu.c | 84 assertf(!(((uintptr_t)(p)) & ((a) - 1)), \ 532 assertf((size & (sizeof(uint64_t) - 1)) == 0, "FP save area component not a multiple of 8 bytes"); in fpu_allzeroes() 708 …assertf(fpu_YMM_capable ? (xgetbv(XCR0) == xstate_xmask[cdp->cpu_xstate]) : TRUE, "XCR0 mismatch: … in fpu_switch_context()
|
| /xnu-8792.41.9/bsd/dev/i386/ |
| H A D | systemcalls.c | 248 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall() 461 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall64() 595 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall_return()
|
| /xnu-8792.41.9/bsd/dev/arm/ |
| H A D | systemcalls.c | 204 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall() 259 …assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall_return()
|
| /xnu-8792.41.9/osfmk/arm/pmap/ |
| H A D | pmap_internal.h | 179 #define ASSERT_NOT_HIBERNATING() (assertf(!hib_entry_pmap_lockdown, \
|
| /xnu-8792.41.9/iokit/Tests/ |
| H A D | TestIOMemoryDescriptor.cpp | 218 assertf((0x53535300 | dir) == data, "mismatch 0x%x", data); in IODMACommandForceDoubleBufferTest() 232 assertf((0x11223300 | dir) == data, "mismatch 0x%x", data); in IODMACommandForceDoubleBufferTest() 313 …assertf(segments[0].fIOVMAddr != segPhys, "phys !local 0x%qx, 0x%qx, %p", segments[0].fIOVMAddr, s… in IODMACommandLocalMappedNonContig()
|