| /xnu-12377.41.6/bsd/skywalk/packet/ ! |
| H A D | os_packet_private.h | 641 struct __user_quantum pkt_qum; member 648 #define pkt_flow_id pkt_qum.qum_flow_id 649 #define pkt_flow_id_64 pkt_qum.qum_flow_id_val64 650 #define pkt_qum_qflags pkt_qum.qum_qflags 651 #define pkt_length pkt_qum.qum_len 652 #define pkt_qum_buf pkt_qum.qum_buf[0] 653 #define pkt_svc_class pkt_qum.qum_svc_class 665 #define pkt_flow_token pkt_qum.qum_flow_id_val32[0]
|
| H A D | packet_var.h | 353 struct __kern_quantum pkt_qum; member 354 #define pkt_user pkt_qum.qum_user 551 KQUM_CTOR(&(_pk)->pkt_qum, _midx, \ 552 (((_pu) == NULL) ? NULL : &(_pu)->pkt_qum), _pp, _qflags); \ 560 KQUM_INIT(&(_pk)->pkt_qum, _flags); \
|
| H A D | packet_common.h | 1074 PP_HAS_BUFFER_ON_DEMAND(PKT_ADDR(ph)->pkt_qum.qum_pp)); in __packet_get_buflet_count() 1100 VERIFY(PP_HAS_BUFFER_ON_DEMAND(PKT_ADDR(ph)->pkt_qum.qum_pp)); in __packet_add_buflet() 1174 PP_HAS_BUFFER_ON_DEMAND(PKT_ADDR(ph)->pkt_qum.qum_pp)); in __packet_get_next_buflet() 1439 VERIFY(pkt->pkt_qum.qum_qflags & QUM_F_INTERNALIZED); in __packet_initialize_with_mbufchain() 1444 pkt->pkt_qum.qum_qflags &= ~(QUM_F_DROPPED | QUM_F_FINALIZED); in __packet_initialize_with_mbufchain() 1452 pkt->pkt_qum.qum_qflags |= QUM_F_FINALIZED; in __packet_initialize_with_mbufchain() 1481 ASSERT(pkt->pkt_qum.qum_qflags & QUM_F_INTERNALIZED); in __packet_finalize_with_mbuf() 1486 pkt->pkt_qum.qum_qflags &= ~(QUM_F_DROPPED | QUM_F_FINALIZED); in __packet_finalize_with_mbuf() 1530 pkt->pkt_qum.qum_qflags |= QUM_F_FINALIZED; in __packet_finalize_with_mbuf() 1533 pkt->pkt_qum.qum_qflags |= QUM_F_DROPPED; in __packet_finalize_with_mbuf()
|
| H A D | packet_kern.c | 571 if ((p1->pkt_qum.qum_qflags & QUM_F_DROPPED) != 0 || in kern_packet_clone_internal() 632 _QUM_COPY(&(p1)->pkt_qum, &(p2)->pkt_qum); in kern_packet_clone_internal() 640 p2->pkt_qum.qum_qflags &= ~(QUM_F_FINALIZED | QUM_F_FLOW_CLASSIFIED); in kern_packet_clone_internal()
|
| H A D | packet_copy.c | 112 VERIFY((doff + len) <= PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp)); in pkt_copy_from_pkt() 678 VERIFY((doff + len) <= (PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp) * in pkt_copy_multi_buflet_from_pkt() 740 PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp)); in pkt_copy_multi_buflet_from_pkt() 833 VERIFY((poff + len) <= PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)); in pkt_copy_from_mbuf() 1233 VERIFY((poff + len) <= (PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp) * in pkt_copy_multi_buflet_from_mbuf() 1500 VERIFY((poff + len) <= PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)); in pkt_copy_to_mbuf() 2076 ASSERT(len <= PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)); in pkt_add_trailers() 2080 (len + extra) > PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)) { in pkt_add_trailers()
|
| H A D | pbufpool.c | 763 ASSERT(kpkt->pkt_qum.qum_pp == pp); in pp_metadata_destruct_common() 767 ASSERT(kpkt->pkt_qum.qum_ksd == NULL); in pp_metadata_destruct_common() 1645 kpkt->pkt_qum.qum_user); in pp_remove_upp_bft_chain_locked() 2250 ASSERT(kpkt->pkt_qum.qum_ksd == NULL); in pp_metadata_fini() 2277 pp = __DECONST(struct kern_pbufpool *, pkt_chain->pkt_qum.qum_pp); in pp_free_packet_chain() 2360 kpkt->pkt_qum.qum_pid, NULL, -1, NULL, 0, 0); in pp_drop_pktq() 2442 pkt->pkt_qum.qum_pp), SK_PTR_ADDR(pkt)); in pp_free_packet_single() 2467 pkt->pkt_qum.qum_pid, NULL, -1, NULL, 0, 0); in pp_drop_packet_single()
|
| H A D | pbufpool_kern.c | 396 VERIFY(pp == pkt_chain->pkt_qum.qum_pp); in kern_pbufpool_free_chain()
|
| /xnu-12377.41.6/bsd/skywalk/nexus/flowswitch/ ! |
| H A D | fsw_dp.c | 331 fsw->fsw_ifp, kpkt->pkt_qum.qum_pid, NULL, -1, NULL, \ 346 fsw->fsw_ifp, (pkt)->pkt_qum.qum_pid, NULL, -1, NULL, 0, 0); \ 359 fsw->fsw_ifp, (pkt)->pkt_qum.qum_pid, NULL, -1, NULL, 0, 0); \ 374 NULL, pkt->pkt_qum.qum_pid, NULL, -1, NULL, \ 538 PP_BUF_SIZE_DEF(SK_PTR_ADDR_KPKT(dph)->pkt_qum.qum_pp)) - in validate_pkt_len() 593 PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp)); in copy_packet_from_dev() 599 _QUM_COPY(&(spkt)->pkt_qum, &(dpkt)->pkt_qum); in copy_packet_from_dev() 601 ASSERT(!(dpkt->pkt_qum.qum_qflags & QUM_F_KERNEL_ONLY) || in copy_packet_from_dev() 602 PP_KERNEL_ONLY(dpkt->pkt_qum.qum_pp)); in copy_packet_from_dev() 1354 pp = __DECONST(struct kern_pbufpool *, KPKTQ_FIRST(pktq)->pkt_qum.qum_pp); in convert_native_pktq_to_mbufs() [all …]
|
| H A D | fsw_ip_frag.c | 1051 pp_free_packet(__DECONST(struct kern_pbufpool *, pkt->pkt_qum.qum_pp), in ipf_free_pkt()
|
| /xnu-12377.41.6/bsd/skywalk/channel/ ! |
| H A D | channel_ring.c | 1717 static_assert(offsetof(struct __kern_packet, pkt_qum) == 0); in kr_internalize_metadata() 1718 static_assert(offsetof(struct __user_packet, pkt_qum) == 0); in kr_internalize_metadata() 1773 …static_assert(offsetof(struct __kern_packet, pkt_length) == offsetof(struct __kern_packet, pkt_qum… in kr_internalize_metadata() 1896 kbuf = &kpkt->pkt_qum.qum_buf[0]; in kr_externalize_metadata_internal() 1899 &kpkt->pkt_qum.qum_user->qum_buf[0]); in kr_externalize_metadata_internal() 1909 static_assert(offsetof(struct __kern_packet, pkt_qum) == 0); in kr_externalize_metadata_internal() 1923 &kpkt->pkt_qum.qum_user->qum_buf[0]); in kr_externalize_metadata_internal()
|
| H A D | channel.c | 205 static_assert(offsetof(struct __user_packet, pkt_qum) == 0); in channel_init() 206 static_assert(offsetof(struct __kern_packet, pkt_qum) == 0); in channel_init()
|
| /xnu-12377.41.6/bsd/skywalk/nexus/flowswitch/flow/ ! |
| H A D | flow_agg.c | 1498 _QUM_COPY(&(spkt)->pkt_qum, &(dpkt)->pkt_qum); in pkt_copy_metadata() 1700 ASSERT(pkt->pkt_qum.qum_pp != dpp); in flow_rx_agg_channel()
|
| /xnu-12377.41.6/bsd/net/pktsched/ ! |
| H A D | pktsched.c | 516 flags, ifp, kpkt->pkt_qum.qum_pid, NULL, -1, NULL, 0, 0); in pktsched_drop_pkt()
|
| /xnu-12377.41.6/bsd/skywalk/nexus/netif/ ! |
| H A D | nx_netif_vp.c | 112 if (pkt->pkt_qum.qum_pp == ring->ckr_pp) { in netif_copy_or_attach_pkt()
|
| /xnu-12377.41.6/bsd/net/ ! |
| H A D | dlil.c | 3387 kpkt->pkt_qum.qum_pp), SK_PTR_ADDR(kpkt)); in ifnet_enqueue_pkt() 3395 kpkt->pkt_qum.qum_pp), SK_PTR_ADDR(kpkt)); in ifnet_enqueue_pkt() 3400 kpkt->pkt_qum.qum_pp), SK_PTR_ADDR(kpkt)); in ifnet_enqueue_pkt()
|