Home
last modified time | relevance | path

Searched refs:assertf (Results 1 – 25 of 59) sorted by relevance

123

/xnu-8020.101.4/osfmk/vm/
H A Dvm_tests.c539 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k()
540assertf(alloc1_addr == expected_addr, "alloc1_addr = 0x%lx expected 0x%lx", alloc1_addr, expected_… in vm_test_4k()
546 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k()
557 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k()
558assertf(alloc1_addr == expected_addr, "alloc1_addr = 0x%lx expected 0x%lx", alloc1_addr, expected_… in vm_test_4k()
569 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k()
570assertf(alloc2_addr == expected_addr, "alloc2_addr = 0x%lx expected 0x%lx", alloc2_addr, expected_… in vm_test_4k()
581 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k()
582assertf(alloc3_addr == expected_addr, "alloc3_addr = 0x%lx expected 0x%lx\n", alloc3_addr, expecte… in vm_test_4k()
593 assertf(kr == KERN_SUCCESS, "kr = 0x%x", kr); in vm_test_4k()
[all …]
H A Dvm_page.h290 assertf(fault_phys_offset < PAGE_SIZE && in VMP_CS_FOR_OFFSET()
301 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_VALIDATED()
315 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_TAINTED()
329 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_NX()
344 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_VALIDATED()
366 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_TAINTED()
388 assertf(fault_page_size <= PAGE_SIZE, in VMP_CS_SET_NX()
H A Dvm_compressor_algorithms.c402 assertf(rval == PAGE_SIZE, "LZ4 decode: size != pgsize %d, header: 0x%x, 0x%x, 0x%x", in metadecompressor()
461 assertf(((new_codec == VM_COMPRESSOR_DEFAULT_CODEC) || (new_codec == CMODE_WK) || in vm_compressor_algorithm_init()
H A Dvm_map_store.c166 assertf(VM_MAP_PAGE_SHIFT(VME_SUBMAP(entry)) >= VM_MAP_PAGE_SHIFT(map), in vm_map_store_entry_link()
H A Dvm_fault.c2583 assertf(fault_phys_offset == 0, in vm_fault_cs_need_validation()
3409 assertf((!(fault_phys_offset & FOURK_PAGE_MASK) && in vm_fault_attempt_pmap_enter()
3413 assertf(fault_phys_offset == 0, in vm_fault_attempt_pmap_enter()
3770 assertf(VM_PAGE_OBJECT(m) != VM_OBJECT_NULL, "m=%p", m); in vm_fault_enter()
4789 assertf((!(fault_phys_offset & FOURK_PAGE_MASK) && in vm_fault_internal()
4793 assertf(fault_phys_offset == 0, in vm_fault_internal()
4797 assertf(VM_PAGE_OBJECT(m) == m_object, "m=%p m_object=%p object=%p", m, m_object, object); in vm_fault_internal()
5408 assertf(VM_PAGE_OBJECT(m) == object, "m=%p object=%p", m, object); in vm_fault_internal()
5678 assertf(object->internal, "object %p", object); in vm_fault_internal()
5910 assertf(VM_PAGE_OBJECT(m) == m_object, "m=%p m_object=%p", m, m_object); in vm_fault_internal()
[all …]
H A Dvm_object.c2690 assertf(!(offset & (pmap_page_size - 1)) && !(size & (pmap_page_size - 1)), in vm_object_pmap_protect_options()
3363 assertf(page_aligned(src_offset + size), in vm_object_copy_call()
3590 assertf(page_aligned(copy_size), in vm_object_copy_delayed()
3631 assertf(page_aligned(copy_size), in vm_object_copy_delayed()
3967 assertf(page_aligned(result->vo_shadow_offset), in vm_object_shadow()
4489 assertf(page_aligned(object->vo_shadow_offset), in vm_object_do_collapse()
4492 assertf(page_aligned(backing_object->vo_shadow_offset), in vm_object_do_collapse()
4566 assertf(page_aligned(object->vo_shadow_offset), in vm_object_do_bypass()
4569 assertf(page_aligned(backing_object->vo_shadow_offset), in vm_object_do_bypass()
4682 assertf(page_aligned(hint_offset), "hint_offset 0x%llx", hint_offset); in vm_object_collapse()
[all …]
H A Dvm_map.c463 assertf(!new->use_pmap, "old %p new %p\n", old, new); in vm_map_entry_copy()
983 assertf(kr == KERN_SUCCESS, in vm_map_apple_protected()
985 assertf(map_addr == tmp_entry.vme_start, in vm_map_apple_protected()
1665 assertf(!pgz_owned(address), in vm_map_lookup_entry()
2210 assertf(vmk_flags.__vmkf_unused == 0, "vmk_flags unused=0x%x\n", vmk_flags.__vmkf_unused); in vm_map_enter()
2387 assertf(VM_MAP_PAGE_ALIGNED(*address, FOURK_PAGE_MASK), "0x%llx", (uint64_t)*address); in vm_map_enter()
2388 assertf(VM_MAP_PAGE_ALIGNED(size, FOURK_PAGE_MASK), "0x%llx", (uint64_t)size); in vm_map_enter()
2390 assertf(page_aligned(*address), "0x%llx", (uint64_t)*address); in vm_map_enter()
2391 assertf(page_aligned(size), "0x%llx", (uint64_t)size); in vm_map_enter()
4089 assertf(vmk_flags.__vmkf_unused == 0, "vmk_flags unused=0x%x\n", vmk_flags.__vmkf_unused); in vm_map_enter_mem_object_helper()
[all …]
/xnu-8020.101.4/osfmk/kern/
H A Dsmr.h172 assertf(held_cond, "smr_serialized_load: lock not held"); \
216 assertf(held_cond, "smr_serialized_store: lock not held"); \
246 assertf(held_cond, "smr_serialized_store: lock not held"); \
H A Dassert.h98 #define assertf(ex, fmt, args...) \ macro
149 #define assertf(ex, fmt, args...) \
156 #define assertf(ex, fmt, args...) ((void)0)
H A Dhazard.h159 assertf(held_cond, "hazard_ptr_serialized_load: lock not held"); \
205 assertf(held_cond, "hazard_ptr_serialized_store: lock not held"); \
235 assertf(held_cond, "hazard_ptr_serialized_store: lock not held"); \
H A Dlock_rw.c419 assertf(entry->rwlde_mode_count == 1, in change_held_rwlock()
428 assertf(entry->rwlde_mode_count == -1, in change_held_rwlock()
510 assertf(entry->rwlde_mode_count != INT8_MAX, in add_held_rwlock()
819 assertf(owner != current_thread(), "Lock already held state=0x%x, owner=%p", in lck_rw_lock_exclusive_gen()
1038 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_exclusive_check_contended()
1072 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_exclusive_internal_inline()
1137 assertf(owner != current_thread(), "Lock already held state=0x%x, owner=%p", in lck_rw_lock_shared_gen()
1281 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_shared_internal_inline()
1511 assertf(lock->lck_rw_priv_excl != 0, "lock %p thread %p", lock, current_thread()); in lck_rw_lock_shared_to_exclusive()
1547 assertf(owner == THREAD_NULL, "state=0x%x, owner=%p", ordered_load_rw(lock), owner); in lck_rw_lock_shared_to_exclusive()
[all …]
H A Dlock_ticket.c466 assertf(tlock->lck_owner != (uintptr_t) cthread, in lck_ticket_contended()
717 assertf(tmp.cticket != tmp.nticket, in hw_lck_ticket_unlock()
728 assertf(tlock->lck_owner == (uintptr_t)current_thread(), in lck_ticket_unlock()
743 assertf(owner == self, "lck_ticket_assert_owned: owner %p, current: %p", in lck_ticket_assert_owned()
H A Dhazard.c142 assertf((hga->hga_mask & mask) == 0, "slot %d in use", in __hazard_guard_get()
157 assertf(slot < HAZARD_GUARD_SLOTS, "invalid guard %p", guard); in __hazard_guard_put()
158 assertf((hga->hga_mask & mask) == mask, "slot %d free", in __hazard_guard_put()
H A Dcpu_quiesce.c358 assertf(!cpu_quiescent_counter_needs_commit(orig_state), in cpu_quiescent_counter_checkin()
H A Dsmr.c83 assertf(smr_boot.used <= SMR_EARLY_COUNT, in __smr_init()
/xnu-8020.101.4/iokit/Kernel/
H A DIOPerfControl.cpp93assertf(workTableLength <= kWorkTableMaxSize, "%zu exceeds max allowed capacity of %zu", workTable… in init()
225assertf(token != kIOPerfControlClientWorkUntracked, "Attempt to deallocate token kIOPerfControlCli… in deallocateToken()
226assertf(token <= workTableLength, "Attempt to deallocate token %llu which is greater than the tabl… in deallocateToken()
254 assertf(entry->thread_group, "Invalid work token: %llu", token); in getEntryForToken()
439 assertf(!entry->started, "Work for token %llu was already started", token); in workBegin()
522assertf(thread_group == nullptr, "IOPerfControlWorkContext ID %llu being released without calling … in free()
586assertf(!work_context->started, "IOPerfControlWorkContext ID %llu was already started", work_conte… in workSubmitWithContext()
587assertf(work_context->thread_group == nullptr, "IOPerfControlWorkContext ID %llu has already taken… in workSubmitWithContext()
655assertf(!work_context->started, "IOPerfControlWorkContext %llu was already started", work_context-… in workBeginWithContext()
/xnu-8020.101.4/osfmk/arm64/
H A Dlowmem_vectors.c109 assertf((vm_first_phys != 0) && (vm_last_phys != 0), in patch_low_glo_static_region()
135 assertf((physmap_base != 0) && (physmap_end != 0), in patch_low_glo_static_region()
H A Dbsd_arm64.c300assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_syscall()
/xnu-8020.101.4/osfmk/i386/
H A Dbsd_i386.c374assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in machdep_syscall()
448assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in machdep_syscall64()
605assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_call_munger()
722assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in mach_call_munger64()
/xnu-8020.101.4/osfmk/arm/
H A Dlowmem_vectors.c93 assertf((vm_first_phys != 0) && (vm_last_phys != 0), in patch_low_glo_static_region()
H A Dmachine_routines.c757assertf(((vm_address_t)(vaddr) - gVirtBase) < gPhysSize, "%s: illegal vaddr: %p", __func__, (void*… in ml_static_vtop()
775 assertf((pa - gPhysBase) < gPhysSize, "%s: illegal PA: 0x%lx", __func__, (unsigned long)pa); in phystokv_range()
/xnu-8020.101.4/bsd/dev/i386/
H A Dsystemcalls.c248assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall()
461assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall64()
595assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall_return()
/xnu-8020.101.4/osfmk/arm/pmap/
H A Dpmap_internal.h181 #define ASSERT_NOT_HIBERNATING() (assertf(!hib_entry_pmap_lockdown, \
/xnu-8020.101.4/bsd/dev/arm/
H A Dsystemcalls.c210assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall()
268assertf(prior == NULL, "thread_set_allocation_name(\"%s\") not cleared", kern_allocation_get_name(… in unix_syscall_return()
/xnu-8020.101.4/iokit/Tests/
H A DTestIOMemoryDescriptor.cpp218 assertf((0x53535300 | dir) == data, "mismatch 0x%x", data); in IODMACommandForceDoubleBufferTest()
232 assertf((0x11223300 | dir) == data, "mismatch 0x%x", data); in IODMACommandForceDoubleBufferTest()
313assertf(segments[0].fIOVMAddr != segPhys, "phys !local 0x%qx, 0x%qx, %p", segments[0].fIOVMAddr, s… in IODMACommandLocalMappedNonContig()

123