| /xnu-12377.61.12/bsd/skywalk/nexus/flowswitch/flow/ |
| H A D | flow_classifier.c | 224 if (__probable(IS_P2ALIGNED(&iph->ip_src, 8))) { in flow_pkt_classify() 227 } else if (IS_P2ALIGNED(&iph->ip_src, 4)) { in flow_pkt_classify() 271 if (__probable(IS_P2ALIGNED(&ip6->ip6_src, 8))) { in flow_pkt_classify() 279 } else if (IS_P2ALIGNED(&ip6->ip6_src, 4)) { in flow_pkt_classify() 367 if (__probable(IS_P2ALIGNED(&tcph->th_sport, 4))) { in flow_pkt_classify()
|
| H A D | flow_stats.c | 92 ASSERT(IS_P2ALIGNED(fs, 16)); in flow_stats_alloc()
|
| H A D | flow_agg.c | 344 ASSERT(IS_P2ALIGNED(tcp, sizeof(uint16_t))); in mbuf_csum() 1055 ASSERT(IS_P2ALIGNED(siph, sizeof(uint16_t))); in can_agg_slowpath() 1056 ASSERT(IS_P2ALIGNED(iph, sizeof(uint16_t))); in can_agg_slowpath() 1107 ASSERT(IS_P2ALIGNED(sip6, sizeof(uint16_t))); in can_agg_slowpath() 1196 if (IS_P2ALIGNED(stcp + 1, sizeof(uint32_t))) { in can_agg_slowpath() 1201 if (IS_P2ALIGNED(tcp + 1, sizeof(uint32_t))) { in can_agg_slowpath() 1350 ASSERT(IS_P2ALIGNED(siph, sizeof(uint16_t))); in flow_agg_merge_hdr() 1366 ASSERT(IS_P2ALIGNED(sip6, sizeof(uint16_t))); in flow_agg_merge_hdr() 1394 ASSERT(IS_P2ALIGNED(stcp, sizeof(uint16_t))); in flow_agg_merge_hdr() 1395 ASSERT(IS_P2ALIGNED(tcp, sizeof(uint16_t))); in flow_agg_merge_hdr()
|
| H A D | flow_owner.c | 71 ASSERT(IS_P2ALIGNED(fob, cache_sz)); in flow_owner_buckets_alloc() 92 ASSERT(IS_P2ALIGNED(fob, skmem_cpu_cache_line_size())); in flow_owner_bucket_init()
|
| H A D | flow_route.c | 160 ASSERT(IS_P2ALIGNED(frb, cache_sz)); in flow_route_buckets_alloc() 181 ASSERT(IS_P2ALIGNED(frb, skmem_cpu_cache_line_size())); in flow_route_bucket_init() 257 ASSERT(IS_P2ALIGNED(frib, cache_sz)); in flow_route_id_buckets_alloc() 278 ASSERT(IS_P2ALIGNED(frib, skmem_cpu_cache_line_size())); in flow_route_id_bucket_init()
|
| /xnu-12377.61.12/bsd/kern/ |
| H A D | mcache.c | 387 VERIFY(IS_P2ALIGNED(ccp, CPU_CACHE_LINE_SIZE)); in mcache_create_common() 988 VERIFY(IS_P2ALIGNED(base, cp->mc_align)); in mcache_slab_alloc() 1021 VERIFY(IS_P2ALIGNED(base, cp->mc_align)); in mcache_slab_free() 1059 VERIFY(IS_P2ALIGNED(base, cp->mc_align)); in mcache_slab_audit() 1588 VERIFY(IS_P2ALIGNED(buf_arg, sizeof(u_int64_t))); in mcache_set_pattern() 1589 VERIFY(IS_P2ALIGNED(size, sizeof(u_int64_t))); in mcache_set_pattern() 1602 VERIFY(IS_P2ALIGNED(buf_arg, sizeof(u_int64_t))); in mcache_verify_pattern() 1603 VERIFY(IS_P2ALIGNED(size, sizeof(u_int64_t))); in mcache_verify_pattern() 1620 VERIFY(IS_P2ALIGNED(buf_arg, sizeof(u_int64_t))); in mcache_verify_set_pattern() 1621 VERIFY(IS_P2ALIGNED(size, sizeof(u_int64_t))); in mcache_verify_set_pattern()
|
| H A D | uipc_mbuf2.c | 476 VERIFY(IS_P2ALIGNED(hdr + 1, sizeof(u_int64_t))); in m_tag_alloc_mbuf() 480 VERIFY(IS_P2ALIGNED(t, sizeof(u_int64_t))); in m_tag_alloc_mbuf() 494 VERIFY(IS_P2ALIGNED(t, sizeof(u_int64_t))); in m_tag_alloc_mbuf() 551 VERIFY(IS_P2ALIGNED(tag, sizeof(uint64_t))); in m_tag_kalloc() 557 VERIFY(IS_P2ALIGNED(tag->m_tag_data, sizeof(uint64_t))); in m_tag_kalloc() 646 VERIFY(IS_P2ALIGNED(hdr + 1, sizeof(u_int64_t))); in m_tag_create_mbuf() 653 VERIFY(IS_P2ALIGNED(t, sizeof(u_int64_t))); in m_tag_create_mbuf() 687 VERIFY(IS_P2ALIGNED(hdr + 1, sizeof(u_int64_t))); in m_tag_free_mbuf()
|
| H A D | uipc_mbuf_mcache.c | 1413 VERIFY(IS_P2ALIGNED(buf, MCLBYTES)); in slab_free() 1423 VERIFY(IS_P2ALIGNED(buf, MBIGCLBYTES)); in slab_free() 1436 VERIFY(IS_P2ALIGNED(buf, PAGE_SIZE)); in slab_free() 2503 VERIFY(IS_P2ALIGNED(page, PAGE_SIZE)); 4947 VERIFY(IS_P2ALIGNED(msc, sizeof(u_int64_t))); 5003 VERIFY(IS_P2ALIGNED(mobj, MIN(m_maxsize(class), PAGE_SIZE)));
|
| H A D | uipc_mbuf.c | 3030 (!align || IS_P2ALIGNED((m->m_data - len), sizeof(u_int32_t)))) { in m_prepend_2() 4148 ((IS_P2ALIGNED((m)->m_data, PAGE_SIZE) \ 4150 (!IS_P2ALIGNED((m)->m_data, PAGE_SIZE) && \ 4170 if (IS_P2ALIGNED(data, PAGE_SIZE) && len0 > PAGE_SIZE) { in m_expand() 4172 } else if (!IS_P2ALIGNED(data, PAGE_SIZE) && in m_expand()
|
| /xnu-12377.61.12/bsd/sys/ |
| H A D | mcache.h | 69 #ifndef IS_P2ALIGNED 70 #define IS_P2ALIGNED(v, a) \ macro
|
| H A D | dtrace_glue.h | 455 #define IS_P2ALIGNED(v, a) ((((uintptr_t)(v)) & ((uintptr_t)(a) - 1)) == 0) macro
|
| H A D | mbuf.h | 174 VERIFY(IS_P2ALIGNED((tag), sizeof(u_int64_t))); \ 1562 if (!IS_P2ALIGNED((_m)->m_data, sizeof (u_int32_t))) { \
|
| /xnu-12377.61.12/bsd/skywalk/core/ |
| H A D | skywalk.c | 1264 if (!IS_P2ALIGNED(s1, SK_DUMP_BUF_ALIGN)) { in skywalk_memcmp_mask_self_tests() 1268 ASSERT(IS_P2ALIGNED(s1, SK_DUMP_BUF_ALIGN)); in skywalk_memcmp_mask_self_tests() 1270 ASSERT(IS_P2ALIGNED(s2, SK_DUMP_BUF_ALIGN)); in skywalk_memcmp_mask_self_tests() 1272 ASSERT(IS_P2ALIGNED(s3, SK_DUMP_BUF_ALIGN)); in skywalk_memcmp_mask_self_tests() 1426 if (!IS_P2ALIGNED(s1, SK_DUMP_BUF_ALIGN)) { in skywalk_self_tests() 1430 ASSERT(IS_P2ALIGNED(s1, SK_DUMP_BUF_ALIGN)); in skywalk_self_tests() 1432 ASSERT(IS_P2ALIGNED(s2, SK_DUMP_BUF_ALIGN)); in skywalk_self_tests() 1434 ASSERT(IS_P2ALIGNED(s3, SK_DUMP_BUF_ALIGN)); in skywalk_self_tests()
|
| H A D | skywalk_common.h | 102 #ifndef IS_P2ALIGNED 103 #define IS_P2ALIGNED(v, a) \ macro
|
| /xnu-12377.61.12/bsd/skywalk/packet/ |
| H A D | packet_copy.c | 50 if (__probable(IS_P2ALIGNED(src, 8) && IS_P2ALIGNED(dst, 8))) { in _pkt_copy() 61 if (IS_P2ALIGNED(len, 64)) { in _pkt_copy() 64 } else if (IS_P2ALIGNED(len, 32)) { in _pkt_copy() 67 } else if (IS_P2ALIGNED(len, 8)) { in _pkt_copy() 70 } else if (IS_P2ALIGNED(len, 4)) { in _pkt_copy() 173 if (IS_P2ALIGNED(dbaddr + stuff, sizeof(csum))) { in pkt_copy_from_pkt() 747 if (IS_P2ALIGNED(baddr + stuff, sizeof(csum))) { in pkt_copy_multi_buflet_from_pkt() 910 if (IS_P2ALIGNED(baddr + stuff, sizeof(csum))) { in pkt_copy_from_mbuf() 1311 if (IS_P2ALIGNED(baddr + stuff, sizeof(csum))) { in pkt_copy_multi_buflet_from_mbuf() 1621 if (IS_P2ALIGNED(dp + stuff, sizeof(csum))) { in pkt_copy_to_mbuf() [all …]
|
| H A D | packet_common.h | 449 if (__probable(IS_P2ALIGNED(token, 8))) { in __packet_opt_set_token() 977 if (__probable(IS_P2ALIGNED(flow_uuid, sizeof(uint64_t)))) { in __packet_set_flow_uuid() 981 } else if (__probable(IS_P2ALIGNED(flow_uuid, sizeof(uint32_t)))) { in __packet_set_flow_uuid() 1002 if (__probable(IS_P2ALIGNED(flow_uuid, sizeof(uint64_t)))) { in __packet_get_flow_uuid() 1006 } else if (__probable(IS_P2ALIGNED(flow_uuid, sizeof(uint32_t)))) { in __packet_get_flow_uuid()
|
| /xnu-12377.61.12/bsd/net/ |
| H A D | dlil_ctl.c | 211 VERIFY(dlifp1 == NULL || (IS_P2ALIGNED(dlifp1, sizeof(u_int64_t)) && in dlil_if_acquire() 212 IS_P2ALIGNED(&ifp1->if_data, sizeof(u_int64_t)))); in dlil_if_acquire()
|
| H A D | rtsock.c | 2020 VERIFY(IS_P2ALIGNED(cp, sizeof(u_int32_t))); in sysctl_iflist() 2074 VERIFY(IS_P2ALIGNED(cp, in sysctl_iflist() 2112 VERIFY(IS_P2ALIGNED(cp, sizeof(u_int32_t))); in sysctl_iflist() 2208 VERIFY(IS_P2ALIGNED(cp, sizeof(u_int32_t))); in sysctl_iflist2() 2262 VERIFY(IS_P2ALIGNED(cp, in sysctl_iflist2() 2322 VERIFY(IS_P2ALIGNED(cp, in sysctl_iflist2() 2358 VERIFY(IS_P2ALIGNED(cp, sizeof(u_int32_t))); in sysctl_iflist2()
|
| H A D | dlil_subr.c | 363 VERIFY(IS_P2ALIGNED(ifp->if_tcp_stat, sizeof(u_int64_t)) && in dlil_alloc_local_stats() 364 IS_P2ALIGNED(ifp->if_udp_stat, sizeof(u_int64_t))); in dlil_alloc_local_stats()
|
| H A D | content_filter.c | 2556 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_ctl_q_in_enqueued, in cfil_init() 2558 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_ctl_q_out_enqueued, in cfil_init() 2560 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_ctl_q_in_peeked, in cfil_init() 2562 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_ctl_q_out_peeked, in cfil_init() 2565 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_pending_q_in_enqueued, in cfil_init() 2567 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_pending_q_out_enqueued, in cfil_init() 2570 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_inject_q_in_enqueued, in cfil_init() 2572 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_inject_q_out_enqueued, in cfil_init() 2574 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_inject_q_in_passed, in cfil_init() 2576 VERIFY(IS_P2ALIGNED(&cfil_stats.cfs_inject_q_out_passed, in cfil_init()
|
| /xnu-12377.61.12/bsd/skywalk/mem/ |
| H A D | skmem_slab.c | 506 VERIFY(IS_P2ALIGNED(buf, skm->skm_bufalign)); in skmem_slab_alloc_pseudo_locked() 720 VERIFY(IS_P2ALIGNED(obj, skm->skm_bufalign)); in skmem_slab_free_pseudo_locked()
|
| H A D | skmem_cache.c | 644 VERIFY(IS_P2ALIGNED(skm, CHANNEL_CACHE_ALIGN_MAX)); in skmem_cache_create() 800 VERIFY(IS_P2ALIGNED(ccp, CHANNEL_CACHE_ALIGN_MAX)); in skmem_cache_create()
|
| /xnu-12377.61.12/bsd/netinet6/ |
| H A D | ipsec.h | 313 #define IPSEC_IS_P2ALIGNED(p) IS_P2ALIGNED(p, sizeof (u_int32_t))
|
| /xnu-12377.61.12/bsd/netinet/ |
| H A D | cpu_in_cksum_gen.c | 118 IS_P2ALIGNED(data, sizeof(uint32_t)) && in os_cpu_in_cksum()
|
| /xnu-12377.61.12/libsyscall/wrappers/skywalk/ |
| H A D | os_channel.c | 1204 if (__probable(IS_P2ALIGNED(id, sizeof(uint64_t)))) { in _flowadv_id_equal() 1208 } else if (__probable(IS_P2ALIGNED(id, sizeof(uint32_t)))) { in _flowadv_id_equal()
|