| /xnu-12377.1.9/bsd/skywalk/nexus/netif/ |
| H A D | nx_netif_filter.c | 83 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_inject() local 84 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_inject() 90 lck_mtx_lock(&nif->nif_filter_lock); in nx_netif_filter_inject() 91 if ((nif->nif_filter_flags & NETIF_FILTER_FLAG_ENABLED) == 0) { in nx_netif_filter_inject() 110 f = STAILQ_FIRST(&nif->nif_filter_list); in nx_netif_filter_inject() 129 lck_mtx_unlock(&nif->nif_filter_lock); in nx_netif_filter_inject() 166 lck_mtx_lock(&nif->nif_filter_lock); in nx_netif_filter_inject() 172 lck_mtx_unlock(&nif->nif_filter_lock); in nx_netif_filter_inject() 177 nx_netif_filter_add(struct nx_netif *nif, nexus_port_t port, void *cb_arg, in nx_netif_filter_add() argument 182 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_add() [all …]
|
| H A D | nx_netif_flow.c | 352 nx_netif_flow_deliver(struct nx_netif *nif, struct netif_flow *f, in nx_netif_flow_deliver() argument 355 #pragma unused(nif) in nx_netif_flow_deliver() 360 nx_netif_snoop(struct nx_netif *nif, struct __kern_packet *pkt, in nx_netif_snoop() argument 364 if (!NETIF_IS_LOW_LATENCY(nif)) { in nx_netif_snoop() 368 pktap_input_packet(nif->nif_ifp, AF_INET6, DLT_EN10MB, in nx_netif_snoop() 372 pktap_output_packet(nif->nif_ifp, AF_INET6, DLT_EN10MB, in nx_netif_snoop() 384 nx_netif_validate_macaddr(struct nx_netif *nif, struct __kern_packet *pkt, in nx_netif_validate_macaddr() argument 387 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_validate_macaddr() 388 struct ifnet *ifp = nif->nif_ifp; in nx_netif_validate_macaddr() 401 DTRACE_SKYWALK2(bad__pkt__sz, struct nx_netif *, nif, in nx_netif_validate_macaddr() [all …]
|
| H A D | nx_netif_netagent.c | 47 get_mac_addr(struct nx_netif *nif, struct ether_addr *addr) in get_mac_addr() argument 49 struct ifnet *ifp = nif->nif_ifp; in get_mac_addr() 209 nx_netif_netagent_fill_port_info(struct nx_netif *nif, struct nx_flow_req *nfr, in nx_netif_netagent_fill_port_info() argument 212 #pragma unused(nif) in nx_netif_netagent_fill_port_info() 215 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_netagent_fill_port_info() 265 nx_netif_netagent_flow_bind(struct nx_netif *nif, struct nx_flow_req *nfr) in nx_netif_netagent_flow_bind() argument 271 struct kern_nexus *nx = nif->nif_nx; in nx_netif_netagent_flow_bind() 299 err = nx_netif_netagent_fill_port_info(nif, nfr, &npi); in nx_netif_netagent_flow_bind() 333 nx_netif_netagent_flow_unbind(struct nx_netif *nif, struct nx_flow_req *nfr) in nx_netif_netagent_flow_unbind() argument 336 struct kern_nexus *nx = nif->nif_nx; in nx_netif_netagent_flow_unbind() [all …]
|
| H A D | nx_netif_llink.c | 301 nx_netif_generate_internal_llink_id(struct nx_netif *nif) in nx_netif_generate_internal_llink_id() argument 304 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_generate_internal_llink_id() 309 STAILQ_FOREACH(llink, &nif->nif_llink_list, nll_link) { in nx_netif_generate_internal_llink_id() 324 nx_netif_llink_initialize(struct netif_llink *llink, struct nx_netif *nif, in nx_netif_llink_initialize() argument 328 struct ifnet *ifp = nif->nif_ifp; in nx_netif_llink_initialize() 330 LCK_RW_ASSERT(&nif->nif_llink_lock, LCK_RW_ASSERT_EXCLUSIVE); in nx_netif_llink_initialize() 332 llink->nll_nif = nif; in nx_netif_llink_initialize() 337 llink->nll_link_id_internal = nx_netif_generate_internal_llink_id(nif); in nx_netif_llink_initialize() 409 nx_netif_llink_create_locked(struct nx_netif *nif, in nx_netif_llink_create_locked() argument 413 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_llink_create_locked() [all …]
|
| H A D | nx_netif_vp.c | 143 struct nx_netif *nif = nifna->nifna_netif; in netif_deliver_pkt() local 144 struct netif_stats *nifs = &nif->nif_stats; in netif_deliver_pkt() 177 if (NETIF_IS_LOW_LATENCY(nif)) { in netif_deliver_pkt() 294 netif_change_pending(struct nx_netif *nif) in netif_change_pending() argument 297 while ((nif->nif_flags & NETIF_FLAG_CHANGE_PENDING) != 0) { in netif_change_pending() 298 DTRACE_SKYWALK1(change__pending__wait, struct nx_netif *, nif); in netif_change_pending() 299 (void) msleep(&nif->nif_flags, &sk_lock, (PZERO - 1), in netif_change_pending() 301 DTRACE_SKYWALK1(change__pending__wake, struct nx_netif *, nif); in netif_change_pending() 303 nif->nif_flags |= NETIF_FLAG_CHANGE_PENDING; in netif_change_pending() 307 netif_change_done(struct nx_netif *nif) in netif_change_done() argument [all …]
|
| H A D | nx_netif.c | 635 struct nx_netif *nif; in nx_netif_get_llink_info() local 641 nif = NX_NETIF_PRIVATE(nx); in nx_netif_get_llink_info() 642 if (!NETIF_LLINK_ENABLED(nif)) { in nx_netif_get_llink_info() 646 lck_rw_lock_shared(&nif->nif_llink_lock); in nx_netif_get_llink_info() 647 llink_cnt = nif->nif_llink_cnt; in nx_netif_get_llink_info() 680 STAILQ_FOREACH(llink, &nif->nif_llink_list, nll_link) { in nx_netif_get_llink_info() 717 lck_rw_unlock_shared(&nif->nif_llink_lock); in nx_netif_get_llink_info() 909 struct nx_netif *nif = NX_NETIF_PRIVATE(nx); in __netif_mib_get_stats() local 910 struct ifnet *ifp = nif->nif_ifp; in __netif_mib_get_stats() 919 sns->sns_nifs = nif->nif_stats; in __netif_mib_get_stats() [all …]
|
| H A D | nx_netif_filter_compat.c | 37 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_mbuf_enqueue() local 38 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_mbuf_enqueue() 40 if (nif->nif_filter_cnt == 0) { in nx_netif_filter_tx_mbuf_enqueue() 45 DTRACE_SKYWALK2(mbuf__default__drop, struct nx_netif *, nif, in nx_netif_filter_tx_mbuf_enqueue() 81 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_mbuf_dequeue() local 105 m = get_next_mbuf(nif->nif_tx_processed_mbq, &curr, end); in nx_netif_filter_tx_processed_mbuf_dequeue() 123 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_mbuf_enqueue() local 124 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_processed_mbuf_enqueue() 138 q = &nif->nif_tx_processed_mbq[tc]; in nx_netif_filter_tx_processed_mbuf_enqueue() 163 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_tx_processed_mbuf_get_len() local [all …]
|
| H A D | nx_netif_filter_native.c | 37 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_pkt_enqueue() local 38 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_pkt_enqueue() 40 if (nif->nif_filter_cnt == 0) { in nx_netif_filter_tx_pkt_enqueue() 44 DTRACE_SKYWALK2(pkt__default__drop, struct nx_netif *, nif, in nx_netif_filter_tx_pkt_enqueue() 80 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_pkt_dequeue() local 104 p = get_next_pkt(nif->nif_tx_processed_pktq, &curr, end); in nx_netif_filter_tx_processed_pkt_dequeue() 123 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_pkt_enqueue() local 124 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_processed_pkt_enqueue() 138 q = &nif->nif_tx_processed_pktq[tc]; in nx_netif_filter_tx_processed_pkt_enqueue() 165 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_tx_processed_pkt_get_len() local [all …]
|
| H A D | nx_netif_host.c | 49 struct nx_netif *nif = ((struct nexus_netif_adapter *)na)->nifna_netif; in nx_netif_host_is_gso_needed() local 66 if (((nif->nif_hwassist & IFNET_TSO_IPV4) != 0) || in nx_netif_host_is_gso_needed() 67 ((nif->nif_hwassist & IFNET_TSO_IPV6) != 0)) { in nx_netif_host_is_gso_needed() 82 struct nx_netif *nif = ((struct nexus_netif_adapter *)na)->nifna_netif; in nx_netif_host_adjust_if_capabilities() local 89 nif->nif_hwassist = ifp->if_hwassist; in nx_netif_host_adjust_if_capabilities() 90 nif->nif_capabilities = ifp->if_capabilities; in nx_netif_host_adjust_if_capabilities() 91 nif->nif_capenable = ifp->if_capenable; in nx_netif_host_adjust_if_capabilities() 111 IFNET_MULTIPAGES | (nif->nif_hwassist & in nx_netif_host_adjust_if_capabilities() 114 (nif->nif_capabilities & in nx_netif_host_adjust_if_capabilities() 117 (nif->nif_capenable & in nx_netif_host_adjust_if_capabilities() [all …]
|
| H A D | nx_netif_filter_vp.c | 311 struct nx_netif *__single nif = nx->nx_arg; in netif_filter_na_mem_new() local 315 NETIF_WLOCK_ASSERT_HELD(nif); in netif_filter_na_mem_new() 316 ASSERT(nif->nif_ifp != NULL); in netif_filter_na_mem_new() 327 if (nif->nif_filter_pp == NULL) { in netif_filter_na_mem_new() 334 "%s_netif_filter_pp", if_name(nif->nif_ifp)); in netif_filter_na_mem_new() 344 nif->nif_filter_pp = pp; in netif_filter_na_mem_new() 347 &nif->nif_filter_pp, NULL, FALSE, FALSE, NULL, &err); in netif_filter_na_mem_new() 486 struct nx_netif *nif = NX_NETIF_PRIVATE(nx); in netif_filter_na_dtor() local 489 NETIF_WLOCK(nif); in netif_filter_na_dtor() 496 nif->nif_filter_vp_cnt--; in netif_filter_na_dtor() [all …]
|
| H A D | nx_netif_util.c | 132 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_mbuf_to_filter_pkt() local 133 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_mbuf_to_filter_pkt() 134 struct kern_pbufpool *pp = nif->nif_filter_pp; in nx_netif_mbuf_to_filter_pkt() 135 ifnet_t ifp = nif->nif_ifp; in nx_netif_mbuf_to_filter_pkt() 199 nif->nif_pkt_copy_from_mbuf(type, fph, off, m, 0, in nx_netif_mbuf_to_filter_pkt() 310 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_pkt_to_filter_pkt() local 311 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_pkt_to_filter_pkt() 312 struct kern_pbufpool *pp = nif->nif_filter_pp; in nx_netif_pkt_to_filter_pkt() 313 ifnet_t ifp = nif->nif_ifp; in nx_netif_pkt_to_filter_pkt() 395 nif->nif_pkt_copy_from_mbuf(type, fph, off, m, 0, in nx_netif_pkt_to_filter_pkt() [all …]
|
| H A D | nx_netif_compat.c | 395 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_compat_na_activate() local 508 nx_netif_mit_init(nif, na->na_ifp, in nx_netif_compat_na_activate() 521 nx_netif_mit_init(nif, na->na_ifp, in nx_netif_compat_na_activate() 1280 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_compat_na_rxsync() local 1457 nif->nif_pkt_copy_from_mbuf(NR_RX, ph, 0, m, 0, in nx_netif_compat_na_rxsync() 1562 struct nx_netif *nif = NX_NETIF_PRIVATE(nx); in nx_netif_compat_attach() local 1604 devnca->nca_up.nifna_netif = nif; in nx_netif_compat_attach() 1605 nx_netif_retain(nif); in nx_netif_compat_attach() 1672 nif->nif_pkt_copy_from_mbuf = in nx_netif_compat_attach() 1674 nif->nif_pkt_copy_to_mbuf = in nx_netif_compat_attach() [all …]
|
| H A D | nx_netif_gso.c | 151 struct nx_netif *nif = NA(ifp)->nifna_netif; in netif_gso_check_netif_active() local 152 struct netif_stats *nifs = &nif->nif_stats; in netif_gso_check_netif_active() 153 struct kern_nexus *nx = nif->nif_nx; in netif_gso_check_netif_active() 190 struct nx_netif *nif = NA(ifp)->nifna_netif; in netif_gso_send() local 191 struct netif_stats *nifs = &nif->nif_stats; in netif_gso_send()
|
| H A D | nx_netif_mit.c | 155 nx_netif_mit_init(struct nx_netif *nif, const struct ifnet *ifp, in nx_netif_mit_init() argument 159 #pragma unused(nif) in nx_netif_mit_init() 284 skoid_create(&mit->mit_skoid, SKOID_DNODE(nif->nif_skoid), oid_name, 0); in nx_netif_mit_init()
|
| /xnu-12377.1.9/bsd/skywalk/nexus/ |
| H A D | nexus_traffic_rule_eth.c | 496 struct nx_netif *nif; in eth_traffic_rule_notify() local 509 nif = NA(ifp)->nifna_netif; in eth_traffic_rule_notify() 510 if (!NX_LLINK_PROV(nif->nif_nx)) { in eth_traffic_rule_notify() 516 qset = nx_netif_find_qset(nif, ntre->ntre_ra.ras_qset_id); in eth_traffic_rule_notify() 517 err = nx_netif_notify_steering_info(nif, qset, in eth_traffic_rule_notify()
|
| H A D | nexus_traffic_rule_inet.c | 754 struct nx_netif *nif; in inet_traffic_rule_notify() local 767 nif = NA(ifp)->nifna_netif; in inet_traffic_rule_notify() 768 if (!NX_LLINK_PROV(nif->nif_nx)) { in inet_traffic_rule_notify() 774 qset = nx_netif_find_qset(nif, ntri->ntri_ra.ras_qset_id); in inet_traffic_rule_notify() 780 err = nx_netif_notify_steering_info(nif, qset, in inet_traffic_rule_notify()
|
| /xnu-12377.1.9/bsd/net/ |
| H A D | if_redirect.c | 287 redirect_enqueue_pkt(struct nx_netif *nif, struct __kern_packet *pkt, in redirect_enqueue_pkt() argument 290 ifnet_t ifp = nif->nif_ifp; in redirect_enqueue_pkt() 294 if (NX_LLINK_PROV(nif->nif_nx) && in redirect_enqueue_pkt() 304 qset = nx_netif_find_qset(nif, qset_id); in redirect_enqueue_pkt() 317 redirect_enqueue_mbuf(struct nx_netif *nif, struct mbuf *m, in redirect_enqueue_mbuf() argument 320 return ifnet_enqueue_mbuf(nif->nif_ifp, m, flush, drop); in redirect_enqueue_mbuf() 327 struct nx_netif *nif; in redirect_tx_submit() local 347 nif = dev_nifna->nifna_netif; in redirect_tx_submit() 348 nifs = &nif->nif_stats; in redirect_tx_submit() 370 err = redirect_enqueue_pkt(nif, pkt, FALSE, &drop); in redirect_tx_submit() [all …]
|
| /xnu-12377.1.9/bsd/skywalk/nexus/flowswitch/flow/ |
| H A D | flow_entry.c | 366 struct nx_netif *nif; in flow_qset_select_dynamic() local 423 nif = NX_NETIF_PRIVATE(fsw->fsw_dev_ch->ch_na->na_nx); in flow_qset_select_dynamic() 425 fe->fe_qset = nx_netif_find_qset(nif, qset_id); in flow_qset_select_dynamic() 439 struct nx_netif *nif; in flow_entry_alloc() local 536 nif = NX_NETIF_PRIVATE(dev_na->na_nx); in flow_entry_alloc() 537 if (NX_LLINK_PROV(nif->nif_nx) && in flow_entry_alloc() 543 fe->fe_qset = nx_netif_find_qset(nif, req->nfr_qset_id); in flow_entry_alloc()
|
| /xnu-12377.1.9/bsd/skywalk/nexus/flowswitch/ |
| H A D | fsw.c | 335 struct nx_netif *nif = NA(ifp)->nifna_netif; in fsw_tso_setup() local 348 if (((nif->nif_hwassist & IFNET_TSO_IPV4) != 0) || in fsw_tso_setup() 349 ((nif->nif_hwassist & IFNET_TSO_IPV6) != 0)) { in fsw_tso_setup()
|