| /xnu-10063.141.1/bsd/skywalk/nexus/netif/ |
| H A D | nx_netif_util.c | 176 if (__improbable((off + mlen) > PP_BUF_SIZE_DEF(pp))) { in nx_netif_mbuf_to_filter_pkt() 177 VERIFY(off < PP_BUF_SIZE_DEF(pp)); in nx_netif_mbuf_to_filter_pkt() 178 mlen = PP_BUF_SIZE_DEF(pp) - off; in nx_netif_mbuf_to_filter_pkt() 184 uint32_t, PP_BUF_SIZE_DEF(pp)); in nx_netif_mbuf_to_filter_pkt() 369 if (__improbable((off + plen) > PP_BUF_SIZE_DEF(pp))) { in nx_netif_pkt_to_filter_pkt() 370 VERIFY(off < PP_BUF_SIZE_DEF(pp)); in nx_netif_pkt_to_filter_pkt() 371 plen = PP_BUF_SIZE_DEF(pp) - off; in nx_netif_pkt_to_filter_pkt() 377 int, plen, uint32_t, PP_BUF_SIZE_DEF(pp)); in nx_netif_pkt_to_filter_pkt() 633 if (__improbable((off + len) > PP_BUF_SIZE_DEF(pp))) { in nx_netif_pkt_to_pkt() 638 uint32_t, PP_BUF_SIZE_DEF(pp)); in nx_netif_pkt_to_pkt()
|
| H A D | nx_netif_host.c | 597 VERIFY((poff + len) <= (PP_BUF_SIZE_DEF(pp) * pp->pp_max_frags)); in nx_netif_mbuf_to_kpkt()
|
| H A D | nx_netif_gso.c | 286 VERIFY((tx_headroom + state->hlen + mss) <= PP_BUF_SIZE_DEF(pp)); in netif_gso_tcp_segment_mbuf()
|
| H A D | nx_netif_vp.c | 887 buf_sz = PP_BUF_SIZE_DEF(nx->nx_tx_pp); in netif_vp_region_params_setup()
|
| H A D | nx_netif_compat.c | 1442 } else if (__probable(mlen <= (int)PP_BUF_SIZE_DEF(pp))) { in nx_netif_compat_na_rxsync()
|
| H A D | nx_netif.c | 2770 VERIFY(PP_BUF_SIZE_DEF(nif->nif_nx->nx_tx_pp) >= in nx_netif_verify_tso_config()
|
| /xnu-10063.141.1/bsd/skywalk/packet/ |
| H A D | packet_var.h | 84 PP_BUF_SIZE_DEF(_pp), 0, 0, (_kbuf)->buf_nbft_addr, \ 129 PP_BUF_SIZE_DEF((_pp)); \ 146 PP_BUF_SIZE_LARGE((_pp)) : PP_BUF_SIZE_DEF((_pp))); \ 223 PP_BUF_SIZE_DEF((_pp)), _midx); \
|
| H A D | packet_copy.c | 113 VERIFY((doff + len) <= PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp)); in pkt_copy_from_pkt() 652 VERIFY((doff + len) <= (PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp) * in pkt_copy_multi_buflet_from_pkt() 714 PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp)); in pkt_copy_multi_buflet_from_pkt() 803 VERIFY((poff + len) <= PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)); in pkt_copy_from_mbuf() 1184 VERIFY((poff + len) <= (PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp) * in pkt_copy_multi_buflet_from_mbuf() 1427 VERIFY((poff + len) <= PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)); in pkt_copy_to_mbuf() 1645 VERIFY((poff + len) <= (PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp) * in pkt_copy_multi_buflet_to_mbuf() 1977 ASSERT(len <= PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)); in pkt_add_trailers() 1981 (len + extra) > PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp)) { in pkt_add_trailers()
|
| H A D | pbufpool_var.h | 143 #define PP_BUF_SIZE_DEF(_pp) ((_pp)->pp_buf_size[PBUFPOOL_BUF_IDX_DEF]) macro
|
| H A D | pbufpool_kern.c | 264 pp_info->kpm_bufsize = PP_BUF_SIZE_DEF(pp); in kern_pbufpool_get_memory_info()
|
| H A D | pbufpool.c | 1117 PP_BUF_SIZE_DEF(pp) = def_buf_size; in pp_create() 2134 SK_ROUNDUP(size, PP_BUF_SIZE_DEF(pp)) / PP_BUF_SIZE_DEF(pp); in pp_alloc_packet_by_size()
|
| /xnu-10063.141.1/bsd/net/ |
| H A D | if_ipsec.c | 959 if (length > PP_BUF_SIZE_DEF(rx_pp)) { in ipsec_kpipe_sync_rx_mbuf() 964 pcb->ipsec_ifp->if_xname, length, PP_BUF_SIZE_DEF(rx_pp)); in ipsec_kpipe_sync_rx_mbuf() 1792 if (length > PP_BUF_SIZE_DEF(rx_pp) || in ipsec_netif_sync_rx_mbuf() 1796 u_int32_t fragment_mtu = PP_BUF_SIZE_DEF(rx_pp); in ipsec_netif_sync_rx_mbuf() 1798 pcb->ipsec_input_frag_size < PP_BUF_SIZE_DEF(rx_pp)) { in ipsec_netif_sync_rx_mbuf() 1851 pcb->ipsec_ifp->if_xname, length, PP_BUF_SIZE_DEF(rx_pp)); in ipsec_netif_sync_rx_mbuf() 2086 if (length > PP_BUF_SIZE_DEF(rx_pp)) { in ipsec_netif_sync_rx_mbuf() 2093 pcb->ipsec_ifp->if_xname, length, PP_BUF_SIZE_DEF(rx_pp)); in ipsec_netif_sync_rx_mbuf() 2400 if (length > PP_BUF_SIZE_DEF(rx_pp) || in ipsec_netif_sync_rx_packet() 2404 u_int32_t fragment_mtu = PP_BUF_SIZE_DEF(rx_pp); in ipsec_netif_sync_rx_packet() [all …]
|
| H A D | if_utun.c | 690 if (length > PP_BUF_SIZE_DEF(rx_pp)) { in utun_netif_sync_rx() 697 pcb->utun_ifp->if_xname, length, PP_BUF_SIZE_DEF(rx_pp)); in utun_netif_sync_rx() 3418 length > PP_BUF_SIZE_DEF(rx_pp) || in utun_kpipe_sync_rx()
|
| /xnu-10063.141.1/bsd/skywalk/nexus/flowswitch/flow/ |
| H A D | flow_agg.c | 1611 if (fe->fe_rx_largest_size <= PP_BUF_SIZE_DEF(dpp) || in flow_rx_agg_channel() 1613 agg_bufsize = PP_BUF_SIZE_DEF(dpp); in flow_rx_agg_channel() 1618 bh_cnt = estimate_buf_cnt(fe, PP_BUF_SIZE_DEF(dpp), in flow_rx_agg_channel() 1628 agg_bufsize = PP_BUF_SIZE_DEF(dpp); in flow_rx_agg_channel() 1742 tmp = estimate_buf_cnt(fe, PP_BUF_SIZE_DEF(dpp), in flow_rx_agg_channel()
|
| /xnu-10063.141.1/bsd/skywalk/nexus/flowswitch/ |
| H A D | fsw_dp.c | 468 PP_BUF_SIZE_DEF(SK_PTR_ADDR_KPKT(dph)->pkt_qum.qum_pp)) - in validate_pkt_len() 520 PP_BUF_SIZE_DEF(dpkt->pkt_qum.qum_pp)); in copy_packet_from_dev() 1250 max_pkt_len = PP_BUF_SIZE_DEF(pp) * pp->pp_max_frags; in convert_native_pktq_to_mbufs() 1749 cnt = howmany(fe->fe_rx_pktq_bytes, PP_BUF_SIZE_DEF(dpp)); in dp_flow_rx_process() 1800 n_bufs = howmany(pkt->pkt_length, PP_BUF_SIZE_DEF(dpp)); in dp_flow_rx_process() 1807 PP_BUF_SIZE_DEF(dpp)); in dp_flow_rx_process() 2496 (uint32_t)pp->pp_max_frags * PP_BUF_SIZE_DEF(pp)); in dp_copy_to_dev_log() 2562 if (dev_pkt_len > pp->pp_max_frags * PP_BUF_SIZE_DEF(pp)) { in dp_copy_to_dev() 3552 bufsz = PP_BUF_SIZE_DEF(pkt->pkt_qum.qum_pp); in do_gso()
|
| /xnu-10063.141.1/bsd/skywalk/channel/ |
| H A D | channel_ring.c | 81 PP_BUF_SIZE_LARGE(_pp) : PP_BUF_SIZE_DEF(_pp)) 1234 slot_size = PP_BUF_SIZE_DEF(kring->ckr_pp); in kr_txsync_finalize() 1801 bdlim = PP_BUF_SIZE_DEF(kqum->qum_pp); in kr_internalize_metadata()
|
| /xnu-10063.141.1/bsd/skywalk/nexus/monitor/ |
| H A D | nx_monitor.c | 1198 PP_BUF_SIZE_DEF(mkring->ckr_pp); in nx_mon_parent_sync()
|
| /xnu-10063.141.1/bsd/skywalk/nexus/ |
| H A D | nexus_adapter.c | 1496 PP_BUF_SIZE_DEF(pp)) * in na_kr_create()
|