Home
last modified time | relevance | path

Searched refs:nif (Results 1 – 19 of 19) sorted by relevance

/xnu-12377.1.9/bsd/skywalk/nexus/netif/
H A Dnx_netif_filter.c83 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_inject() local
84 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_inject()
90 lck_mtx_lock(&nif->nif_filter_lock); in nx_netif_filter_inject()
91 if ((nif->nif_filter_flags & NETIF_FILTER_FLAG_ENABLED) == 0) { in nx_netif_filter_inject()
110 f = STAILQ_FIRST(&nif->nif_filter_list); in nx_netif_filter_inject()
129 lck_mtx_unlock(&nif->nif_filter_lock); in nx_netif_filter_inject()
166 lck_mtx_lock(&nif->nif_filter_lock); in nx_netif_filter_inject()
172 lck_mtx_unlock(&nif->nif_filter_lock); in nx_netif_filter_inject()
177 nx_netif_filter_add(struct nx_netif *nif, nexus_port_t port, void *cb_arg, in nx_netif_filter_add() argument
182 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_add()
[all …]
H A Dnx_netif_flow.c352 nx_netif_flow_deliver(struct nx_netif *nif, struct netif_flow *f, in nx_netif_flow_deliver() argument
355 #pragma unused(nif) in nx_netif_flow_deliver()
360 nx_netif_snoop(struct nx_netif *nif, struct __kern_packet *pkt, in nx_netif_snoop() argument
364 if (!NETIF_IS_LOW_LATENCY(nif)) { in nx_netif_snoop()
368 pktap_input_packet(nif->nif_ifp, AF_INET6, DLT_EN10MB, in nx_netif_snoop()
372 pktap_output_packet(nif->nif_ifp, AF_INET6, DLT_EN10MB, in nx_netif_snoop()
384 nx_netif_validate_macaddr(struct nx_netif *nif, struct __kern_packet *pkt, in nx_netif_validate_macaddr() argument
387 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_validate_macaddr()
388 struct ifnet *ifp = nif->nif_ifp; in nx_netif_validate_macaddr()
401 DTRACE_SKYWALK2(bad__pkt__sz, struct nx_netif *, nif, in nx_netif_validate_macaddr()
[all …]
H A Dnx_netif_netagent.c47 get_mac_addr(struct nx_netif *nif, struct ether_addr *addr) in get_mac_addr() argument
49 struct ifnet *ifp = nif->nif_ifp; in get_mac_addr()
209 nx_netif_netagent_fill_port_info(struct nx_netif *nif, struct nx_flow_req *nfr, in nx_netif_netagent_fill_port_info() argument
212 #pragma unused(nif) in nx_netif_netagent_fill_port_info()
215 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_netagent_fill_port_info()
265 nx_netif_netagent_flow_bind(struct nx_netif *nif, struct nx_flow_req *nfr) in nx_netif_netagent_flow_bind() argument
271 struct kern_nexus *nx = nif->nif_nx; in nx_netif_netagent_flow_bind()
299 err = nx_netif_netagent_fill_port_info(nif, nfr, &npi); in nx_netif_netagent_flow_bind()
333 nx_netif_netagent_flow_unbind(struct nx_netif *nif, struct nx_flow_req *nfr) in nx_netif_netagent_flow_unbind() argument
336 struct kern_nexus *nx = nif->nif_nx; in nx_netif_netagent_flow_unbind()
[all …]
H A Dnx_netif_llink.c301 nx_netif_generate_internal_llink_id(struct nx_netif *nif) in nx_netif_generate_internal_llink_id() argument
304 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_generate_internal_llink_id()
309 STAILQ_FOREACH(llink, &nif->nif_llink_list, nll_link) { in nx_netif_generate_internal_llink_id()
324 nx_netif_llink_initialize(struct netif_llink *llink, struct nx_netif *nif, in nx_netif_llink_initialize() argument
328 struct ifnet *ifp = nif->nif_ifp; in nx_netif_llink_initialize()
330 LCK_RW_ASSERT(&nif->nif_llink_lock, LCK_RW_ASSERT_EXCLUSIVE); in nx_netif_llink_initialize()
332 llink->nll_nif = nif; in nx_netif_llink_initialize()
337 llink->nll_link_id_internal = nx_netif_generate_internal_llink_id(nif); in nx_netif_llink_initialize()
409 nx_netif_llink_create_locked(struct nx_netif *nif, in nx_netif_llink_create_locked() argument
413 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_llink_create_locked()
[all …]
H A Dnx_netif_vp.c143 struct nx_netif *nif = nifna->nifna_netif; in netif_deliver_pkt() local
144 struct netif_stats *nifs = &nif->nif_stats; in netif_deliver_pkt()
177 if (NETIF_IS_LOW_LATENCY(nif)) { in netif_deliver_pkt()
294 netif_change_pending(struct nx_netif *nif) in netif_change_pending() argument
297 while ((nif->nif_flags & NETIF_FLAG_CHANGE_PENDING) != 0) { in netif_change_pending()
298 DTRACE_SKYWALK1(change__pending__wait, struct nx_netif *, nif); in netif_change_pending()
299 (void) msleep(&nif->nif_flags, &sk_lock, (PZERO - 1), in netif_change_pending()
301 DTRACE_SKYWALK1(change__pending__wake, struct nx_netif *, nif); in netif_change_pending()
303 nif->nif_flags |= NETIF_FLAG_CHANGE_PENDING; in netif_change_pending()
307 netif_change_done(struct nx_netif *nif) in netif_change_done() argument
[all …]
H A Dnx_netif.c635 struct nx_netif *nif; in nx_netif_get_llink_info() local
641 nif = NX_NETIF_PRIVATE(nx); in nx_netif_get_llink_info()
642 if (!NETIF_LLINK_ENABLED(nif)) { in nx_netif_get_llink_info()
646 lck_rw_lock_shared(&nif->nif_llink_lock); in nx_netif_get_llink_info()
647 llink_cnt = nif->nif_llink_cnt; in nx_netif_get_llink_info()
680 STAILQ_FOREACH(llink, &nif->nif_llink_list, nll_link) { in nx_netif_get_llink_info()
717 lck_rw_unlock_shared(&nif->nif_llink_lock); in nx_netif_get_llink_info()
909 struct nx_netif *nif = NX_NETIF_PRIVATE(nx); in __netif_mib_get_stats() local
910 struct ifnet *ifp = nif->nif_ifp; in __netif_mib_get_stats()
919 sns->sns_nifs = nif->nif_stats; in __netif_mib_get_stats()
[all …]
H A Dnx_netif_filter_compat.c37 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_mbuf_enqueue() local
38 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_mbuf_enqueue()
40 if (nif->nif_filter_cnt == 0) { in nx_netif_filter_tx_mbuf_enqueue()
45 DTRACE_SKYWALK2(mbuf__default__drop, struct nx_netif *, nif, in nx_netif_filter_tx_mbuf_enqueue()
81 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_mbuf_dequeue() local
105 m = get_next_mbuf(nif->nif_tx_processed_mbq, &curr, end); in nx_netif_filter_tx_processed_mbuf_dequeue()
123 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_mbuf_enqueue() local
124 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_processed_mbuf_enqueue()
138 q = &nif->nif_tx_processed_mbq[tc]; in nx_netif_filter_tx_processed_mbuf_enqueue()
163 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_tx_processed_mbuf_get_len() local
[all …]
H A Dnx_netif_filter_native.c37 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_pkt_enqueue() local
38 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_pkt_enqueue()
40 if (nif->nif_filter_cnt == 0) { in nx_netif_filter_tx_pkt_enqueue()
44 DTRACE_SKYWALK2(pkt__default__drop, struct nx_netif *, nif, in nx_netif_filter_tx_pkt_enqueue()
80 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_pkt_dequeue() local
104 p = get_next_pkt(nif->nif_tx_processed_pktq, &curr, end); in nx_netif_filter_tx_processed_pkt_dequeue()
123 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_filter_tx_processed_pkt_enqueue() local
124 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_filter_tx_processed_pkt_enqueue()
138 q = &nif->nif_tx_processed_pktq[tc]; in nx_netif_filter_tx_processed_pkt_enqueue()
165 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_tx_processed_pkt_get_len() local
[all …]
H A Dnx_netif_host.c49 struct nx_netif *nif = ((struct nexus_netif_adapter *)na)->nifna_netif; in nx_netif_host_is_gso_needed() local
66 if (((nif->nif_hwassist & IFNET_TSO_IPV4) != 0) || in nx_netif_host_is_gso_needed()
67 ((nif->nif_hwassist & IFNET_TSO_IPV6) != 0)) { in nx_netif_host_is_gso_needed()
82 struct nx_netif *nif = ((struct nexus_netif_adapter *)na)->nifna_netif; in nx_netif_host_adjust_if_capabilities() local
89 nif->nif_hwassist = ifp->if_hwassist; in nx_netif_host_adjust_if_capabilities()
90 nif->nif_capabilities = ifp->if_capabilities; in nx_netif_host_adjust_if_capabilities()
91 nif->nif_capenable = ifp->if_capenable; in nx_netif_host_adjust_if_capabilities()
111 IFNET_MULTIPAGES | (nif->nif_hwassist & in nx_netif_host_adjust_if_capabilities()
114 (nif->nif_capabilities & in nx_netif_host_adjust_if_capabilities()
117 (nif->nif_capenable & in nx_netif_host_adjust_if_capabilities()
[all …]
H A Dnx_netif_filter_vp.c311 struct nx_netif *__single nif = nx->nx_arg; in netif_filter_na_mem_new() local
315 NETIF_WLOCK_ASSERT_HELD(nif); in netif_filter_na_mem_new()
316 ASSERT(nif->nif_ifp != NULL); in netif_filter_na_mem_new()
327 if (nif->nif_filter_pp == NULL) { in netif_filter_na_mem_new()
334 "%s_netif_filter_pp", if_name(nif->nif_ifp)); in netif_filter_na_mem_new()
344 nif->nif_filter_pp = pp; in netif_filter_na_mem_new()
347 &nif->nif_filter_pp, NULL, FALSE, FALSE, NULL, &err); in netif_filter_na_mem_new()
486 struct nx_netif *nif = NX_NETIF_PRIVATE(nx); in netif_filter_na_dtor() local
489 NETIF_WLOCK(nif); in netif_filter_na_dtor()
496 nif->nif_filter_vp_cnt--; in netif_filter_na_dtor()
[all …]
H A Dnx_netif_util.c132 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_mbuf_to_filter_pkt() local
133 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_mbuf_to_filter_pkt()
134 struct kern_pbufpool *pp = nif->nif_filter_pp; in nx_netif_mbuf_to_filter_pkt()
135 ifnet_t ifp = nif->nif_ifp; in nx_netif_mbuf_to_filter_pkt()
199 nif->nif_pkt_copy_from_mbuf(type, fph, off, m, 0, in nx_netif_mbuf_to_filter_pkt()
310 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_pkt_to_filter_pkt() local
311 struct netif_stats *nifs = &nif->nif_stats; in nx_netif_pkt_to_filter_pkt()
312 struct kern_pbufpool *pp = nif->nif_filter_pp; in nx_netif_pkt_to_filter_pkt()
313 ifnet_t ifp = nif->nif_ifp; in nx_netif_pkt_to_filter_pkt()
395 nif->nif_pkt_copy_from_mbuf(type, fph, off, m, 0, in nx_netif_pkt_to_filter_pkt()
[all …]
H A Dnx_netif_compat.c395 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_compat_na_activate() local
508 nx_netif_mit_init(nif, na->na_ifp, in nx_netif_compat_na_activate()
521 nx_netif_mit_init(nif, na->na_ifp, in nx_netif_compat_na_activate()
1280 struct nx_netif *nif = nifna->nifna_netif; in nx_netif_compat_na_rxsync() local
1457 nif->nif_pkt_copy_from_mbuf(NR_RX, ph, 0, m, 0, in nx_netif_compat_na_rxsync()
1562 struct nx_netif *nif = NX_NETIF_PRIVATE(nx); in nx_netif_compat_attach() local
1604 devnca->nca_up.nifna_netif = nif; in nx_netif_compat_attach()
1605 nx_netif_retain(nif); in nx_netif_compat_attach()
1672 nif->nif_pkt_copy_from_mbuf = in nx_netif_compat_attach()
1674 nif->nif_pkt_copy_to_mbuf = in nx_netif_compat_attach()
[all …]
H A Dnx_netif_gso.c151 struct nx_netif *nif = NA(ifp)->nifna_netif; in netif_gso_check_netif_active() local
152 struct netif_stats *nifs = &nif->nif_stats; in netif_gso_check_netif_active()
153 struct kern_nexus *nx = nif->nif_nx; in netif_gso_check_netif_active()
190 struct nx_netif *nif = NA(ifp)->nifna_netif; in netif_gso_send() local
191 struct netif_stats *nifs = &nif->nif_stats; in netif_gso_send()
H A Dnx_netif_mit.c155 nx_netif_mit_init(struct nx_netif *nif, const struct ifnet *ifp, in nx_netif_mit_init() argument
159 #pragma unused(nif) in nx_netif_mit_init()
284 skoid_create(&mit->mit_skoid, SKOID_DNODE(nif->nif_skoid), oid_name, 0); in nx_netif_mit_init()
/xnu-12377.1.9/bsd/skywalk/nexus/
H A Dnexus_traffic_rule_eth.c496 struct nx_netif *nif; in eth_traffic_rule_notify() local
509 nif = NA(ifp)->nifna_netif; in eth_traffic_rule_notify()
510 if (!NX_LLINK_PROV(nif->nif_nx)) { in eth_traffic_rule_notify()
516 qset = nx_netif_find_qset(nif, ntre->ntre_ra.ras_qset_id); in eth_traffic_rule_notify()
517 err = nx_netif_notify_steering_info(nif, qset, in eth_traffic_rule_notify()
H A Dnexus_traffic_rule_inet.c754 struct nx_netif *nif; in inet_traffic_rule_notify() local
767 nif = NA(ifp)->nifna_netif; in inet_traffic_rule_notify()
768 if (!NX_LLINK_PROV(nif->nif_nx)) { in inet_traffic_rule_notify()
774 qset = nx_netif_find_qset(nif, ntri->ntri_ra.ras_qset_id); in inet_traffic_rule_notify()
780 err = nx_netif_notify_steering_info(nif, qset, in inet_traffic_rule_notify()
/xnu-12377.1.9/bsd/net/
H A Dif_redirect.c287 redirect_enqueue_pkt(struct nx_netif *nif, struct __kern_packet *pkt, in redirect_enqueue_pkt() argument
290 ifnet_t ifp = nif->nif_ifp; in redirect_enqueue_pkt()
294 if (NX_LLINK_PROV(nif->nif_nx) && in redirect_enqueue_pkt()
304 qset = nx_netif_find_qset(nif, qset_id); in redirect_enqueue_pkt()
317 redirect_enqueue_mbuf(struct nx_netif *nif, struct mbuf *m, in redirect_enqueue_mbuf() argument
320 return ifnet_enqueue_mbuf(nif->nif_ifp, m, flush, drop); in redirect_enqueue_mbuf()
327 struct nx_netif *nif; in redirect_tx_submit() local
347 nif = dev_nifna->nifna_netif; in redirect_tx_submit()
348 nifs = &nif->nif_stats; in redirect_tx_submit()
370 err = redirect_enqueue_pkt(nif, pkt, FALSE, &drop); in redirect_tx_submit()
[all …]
/xnu-12377.1.9/bsd/skywalk/nexus/flowswitch/flow/
H A Dflow_entry.c366 struct nx_netif *nif; in flow_qset_select_dynamic() local
423 nif = NX_NETIF_PRIVATE(fsw->fsw_dev_ch->ch_na->na_nx); in flow_qset_select_dynamic()
425 fe->fe_qset = nx_netif_find_qset(nif, qset_id); in flow_qset_select_dynamic()
439 struct nx_netif *nif; in flow_entry_alloc() local
536 nif = NX_NETIF_PRIVATE(dev_na->na_nx); in flow_entry_alloc()
537 if (NX_LLINK_PROV(nif->nif_nx) && in flow_entry_alloc()
543 fe->fe_qset = nx_netif_find_qset(nif, req->nfr_qset_id); in flow_entry_alloc()
/xnu-12377.1.9/bsd/skywalk/nexus/flowswitch/
H A Dfsw.c335 struct nx_netif *nif = NA(ifp)->nifna_netif; in fsw_tso_setup() local
348 if (((nif->nif_hwassist & IFNET_TSO_IPV4) != 0) || in fsw_tso_setup()
349 ((nif->nif_hwassist & IFNET_TSO_IPV6) != 0)) { in fsw_tso_setup()