| /xnu-12377.61.12/bsd/skywalk/mem/ |
| H A D | skmem_cache.c | 548 struct skmem_cache *__single skm; in skmem_cache_create() local 631 skm = (void *)((char *)buf + diff); in skmem_cache_create() 632 skm->skm_start = buf; in skmem_cache_create() 640 skm = zalloc_flags_buf(skm_zone, Z_WAITOK | Z_ZERO); in skmem_cache_create() 642 skm->skm_cpu_cache_count = ncpu; in skmem_cache_create() 644 VERIFY(IS_P2ALIGNED(skm, CHANNEL_CACHE_ALIGN_MAX)); in skmem_cache_create() 653 skm->skm_mode |= SKM_MODE_NOMAGAZINES; in skmem_cache_create() 666 skm->skm_mode |= SKM_MODE_DYNAMIC; in skmem_cache_create() 671 skm->skm_mode |= SKM_MODE_NOREDIRECT; in skmem_cache_create() 691 skm->skm_mode |= SKM_MODE_BATCH; in skmem_cache_create() [all …]
|
| H A D | skmem_slab.c | 64 skmem_slab_create(struct skmem_cache *skm, uint32_t skmflag) in skmem_slab_create() argument 66 struct skmem_region *skr = skm->skm_region; in skmem_slab_create() 68 size_t slabsize = skm->skm_slabsize; in skmem_slab_create() 93 sl->sl_cache = skm; in skmem_slab_create() 99 ASSERT(skm->skm_objsize == objsize); in skmem_slab_create() 137 SK_KVA(skm), SK_KVA(sl)); in skmem_slab_create() 144 skmem_slab_destroy(skm, sl); in skmem_slab_create() 151 os_atomic_inc(&skm->skm_sl_alloc_fail, relaxed); in skmem_slab_create() 160 skmem_slab_destroy(struct skmem_cache *skm, struct skmem_slab *sl) in skmem_slab_destroy() argument 169 SK_KVA(skm), SK_KVA(sl)); in skmem_slab_destroy() [all …]
|
| H A D | skmem.c | 609 struct skmem_cache *skm; in skmem_dump() local 633 if ((skm = skr->skr_cache[i]) == NULL) { in skmem_dump() 642 skm, skm->skm_mode, skm->skm_sl_bufinuse, in skmem_dump() 643 skm->skm_sl_bufmax, skm->skm_sl_alloc_fail, in skmem_dump() 644 skm->skm_sl_create, skm->skm_sl_destroy, in skmem_dump() 645 skm->skm_sl_alloc, skm->skm_sl_free); in skmem_dump()
|
| H A D | skmem_cache_var.h | 436 skmem_cache_alloc(struct skmem_cache *skm, uint32_t skmflag) in skmem_cache_alloc() argument 440 (void) skmem_cache_batch_alloc(skm, &buf, skm->skm_objsize, 1, skmflag); in skmem_cache_alloc() 443 return __unsafe_forge_bidi_indexable(void *, buf, buf ? skm->skm_objsize : 0); in skmem_cache_alloc()
|
| H A D | skmem_test.c | 1737 struct skmem_cache *skm; in skmem_tests() local 1752 skm = skmem_cache_create(name, bufsize, align, skmem_test_ctor, in skmem_tests() 1757 objary[i] = skmem_cache_alloc(skm, SKMEM_SLEEP); in skmem_tests() 1763 skmem_cache_free(skm, objary[i]); in skmem_tests() 1766 skmem_cache_destroy(skm); in skmem_tests()
|
| H A D | skmem_region.c | 1045 skmem_region_slab_config(struct skmem_region *skr, struct skmem_cache *skm, in skmem_region_slab_config() argument 1059 skr->skr_cache[i] = skm; in skmem_region_slab_config() 1064 for (i = 0; i < SKR_MAX_CACHES && skr->skr_cache[i] != skm; in skmem_region_slab_config() 1069 ASSERT(skr->skr_cache[i] == skm); in skmem_region_slab_config()
|
| /xnu-12377.61.12/tools/lldbmacros/ |
| H A D | skywalk.py | 172 for skm in IterateTAILQ_HEAD(skmhead, "skm_link") : 174 …skm, skm.skm_sl_bufinuse, skm.skm_sl_bufmax, skm.skm_sl_rescale, skm.skm_sl_create, skm.skm_sl_des… 186 skm = kern.GetValueFromAddress(cmd_args[0], 'skmem_cache *') 188 for slab in IterateTAILQ_HEAD(skm.skm_sl_partial_list, "sl_link") : 200 for slab in IterateTAILQ_HEAD(skm.skm_sl_empty_list, "sl_link") : 213 for i in range(0, skm.skm_hash_mask + 1) : 217 print(format_string.format(addressof(skm.skm_hash_table[i]), i)) 221 for bc in IterateListEntry(skm.skm_hash_table[i].bcb_head, 'bc_link',
|
| /xnu-12377.61.12/bsd/skywalk/packet/ |
| H A D | pbufpool.c | 2486 struct skmem_cache *skm = large ? PP_BUF_CACHE_LARGE(pp): in pp_alloc_buffer_common() local 2489 ASSERT(skm != NULL); in pp_alloc_buffer_common() 2491 baddr = (mach_vm_address_t)skmem_cache_alloc(skm, skmflag); in pp_alloc_buffer_common() 2503 skmem_cache_free(skm, in pp_alloc_buffer_common() 2515 skmem_cache_get_obj_info(skm, in pp_alloc_buffer_common()
|