1*d4514f0bSApple OSS Distributions /*
2*d4514f0bSApple OSS Distributions * Copyright (c) 2016-2021 Apple Inc. All rights reserved.
3*d4514f0bSApple OSS Distributions *
4*d4514f0bSApple OSS Distributions * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5*d4514f0bSApple OSS Distributions *
6*d4514f0bSApple OSS Distributions * This file contains Original Code and/or Modifications of Original Code
7*d4514f0bSApple OSS Distributions * as defined in and that are subject to the Apple Public Source License
8*d4514f0bSApple OSS Distributions * Version 2.0 (the 'License'). You may not use this file except in
9*d4514f0bSApple OSS Distributions * compliance with the License. The rights granted to you under the License
10*d4514f0bSApple OSS Distributions * may not be used to create, or enable the creation or redistribution of,
11*d4514f0bSApple OSS Distributions * unlawful or unlicensed copies of an Apple operating system, or to
12*d4514f0bSApple OSS Distributions * circumvent, violate, or enable the circumvention or violation of, any
13*d4514f0bSApple OSS Distributions * terms of an Apple operating system software license agreement.
14*d4514f0bSApple OSS Distributions *
15*d4514f0bSApple OSS Distributions * Please obtain a copy of the License at
16*d4514f0bSApple OSS Distributions * http://www.opensource.apple.com/apsl/ and read it before using this file.
17*d4514f0bSApple OSS Distributions *
18*d4514f0bSApple OSS Distributions * The Original Code and all software distributed under the License are
19*d4514f0bSApple OSS Distributions * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20*d4514f0bSApple OSS Distributions * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21*d4514f0bSApple OSS Distributions * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22*d4514f0bSApple OSS Distributions * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23*d4514f0bSApple OSS Distributions * Please see the License for the specific language governing rights and
24*d4514f0bSApple OSS Distributions * limitations under the License.
25*d4514f0bSApple OSS Distributions *
26*d4514f0bSApple OSS Distributions * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27*d4514f0bSApple OSS Distributions */
28*d4514f0bSApple OSS Distributions
29*d4514f0bSApple OSS Distributions /*
30*d4514f0bSApple OSS Distributions * The migration of flow queue between the different states is summarised in
31*d4514f0bSApple OSS Distributions * the below state diagram. (RFC 8290)
32*d4514f0bSApple OSS Distributions *
33*d4514f0bSApple OSS Distributions * +-----------------+ +------------------+
34*d4514f0bSApple OSS Distributions * | | Empty | |
35*d4514f0bSApple OSS Distributions * | Empty |<---------------+ Old +----+
36*d4514f0bSApple OSS Distributions * | | | | |
37*d4514f0bSApple OSS Distributions * +-------+---------+ +------------------+ |
38*d4514f0bSApple OSS Distributions * | ^ ^ |Credits
39*d4514f0bSApple OSS Distributions * |Arrival | | |Exhausted
40*d4514f0bSApple OSS Distributions * v | | |
41*d4514f0bSApple OSS Distributions * +-----------------+ | | |
42*d4514f0bSApple OSS Distributions * | | Empty or | | |
43*d4514f0bSApple OSS Distributions * | New +-------------------+ +-------+
44*d4514f0bSApple OSS Distributions * | | Credits Exhausted
45*d4514f0bSApple OSS Distributions * +-----------------+
46*d4514f0bSApple OSS Distributions *
47*d4514f0bSApple OSS Distributions * In this implementation of FQ-CODEL, flow queue is a dynamically allocated
48*d4514f0bSApple OSS Distributions * object. An active flow queue goes through the above cycle of state
49*d4514f0bSApple OSS Distributions * transitions very often. To avoid the cost of frequent flow queue object
50*d4514f0bSApple OSS Distributions * allocation/free, this implementation retains the flow queue object in
51*d4514f0bSApple OSS Distributions * [Empty] state on an Empty flow queue list with an active reference in flow
52*d4514f0bSApple OSS Distributions * queue hash table. The flow queue objects on the Empty flow queue list have
53*d4514f0bSApple OSS Distributions * an associated age and are purged accordingly.
54*d4514f0bSApple OSS Distributions */
55*d4514f0bSApple OSS Distributions
56*d4514f0bSApple OSS Distributions #include <sys/cdefs.h>
57*d4514f0bSApple OSS Distributions #include <sys/param.h>
58*d4514f0bSApple OSS Distributions #include <sys/mbuf.h>
59*d4514f0bSApple OSS Distributions #include <sys/socket.h>
60*d4514f0bSApple OSS Distributions #include <sys/sockio.h>
61*d4514f0bSApple OSS Distributions #include <sys/systm.h>
62*d4514f0bSApple OSS Distributions #include <sys/syslog.h>
63*d4514f0bSApple OSS Distributions #include <sys/proc.h>
64*d4514f0bSApple OSS Distributions #include <sys/errno.h>
65*d4514f0bSApple OSS Distributions #include <sys/kernel.h>
66*d4514f0bSApple OSS Distributions #include <sys/kauth.h>
67*d4514f0bSApple OSS Distributions #include <sys/sdt.h>
68*d4514f0bSApple OSS Distributions #include <kern/zalloc.h>
69*d4514f0bSApple OSS Distributions #include <netinet/in.h>
70*d4514f0bSApple OSS Distributions
71*d4514f0bSApple OSS Distributions #include <net/classq/classq.h>
72*d4514f0bSApple OSS Distributions #include <net/classq/if_classq.h>
73*d4514f0bSApple OSS Distributions #include <net/pktsched/pktsched.h>
74*d4514f0bSApple OSS Distributions #include <net/pktsched/pktsched_fq_codel.h>
75*d4514f0bSApple OSS Distributions #include <net/classq/classq_fq_codel.h>
76*d4514f0bSApple OSS Distributions #include <net/droptap.h>
77*d4514f0bSApple OSS Distributions
78*d4514f0bSApple OSS Distributions #include <netinet/tcp_var.h>
79*d4514f0bSApple OSS Distributions
80*d4514f0bSApple OSS Distributions #define FQ_ZONE_MAX (32 * 1024) /* across all interfaces */
81*d4514f0bSApple OSS Distributions
82*d4514f0bSApple OSS Distributions #define DTYPE_NODROP 0 /* no drop */
83*d4514f0bSApple OSS Distributions #define DTYPE_FORCED 1 /* a "forced" drop */
84*d4514f0bSApple OSS Distributions #define DTYPE_EARLY 2 /* an "unforced" (early) drop */
85*d4514f0bSApple OSS Distributions
86*d4514f0bSApple OSS Distributions static uint32_t pkt_compressor = 1;
87*d4514f0bSApple OSS Distributions static uint64_t l4s_ce_threshold = 0; /* in usec */
88*d4514f0bSApple OSS Distributions static uint32_t l4s_local_ce_report = 0;
89*d4514f0bSApple OSS Distributions static uint64_t pkt_pacing_leeway = 0; /* in usec */
90*d4514f0bSApple OSS Distributions static uint64_t max_pkt_pacing_interval = 3 * NSEC_PER_SEC;
91*d4514f0bSApple OSS Distributions static uint64_t l4s_min_delay_threshold = 20 * NSEC_PER_MSEC; /* 20 ms */
92*d4514f0bSApple OSS Distributions #if (DEBUG || DEVELOPMENT)
93*d4514f0bSApple OSS Distributions SYSCTL_NODE(_net_classq, OID_AUTO, flow_q, CTLFLAG_RW | CTLFLAG_LOCKED,
94*d4514f0bSApple OSS Distributions 0, "FQ-CODEL parameters");
95*d4514f0bSApple OSS Distributions
96*d4514f0bSApple OSS Distributions SYSCTL_UINT(_net_classq_flow_q, OID_AUTO, pkt_compressor,
97*d4514f0bSApple OSS Distributions CTLFLAG_RW | CTLFLAG_LOCKED, &pkt_compressor, 0, "enable pkt compression");
98*d4514f0bSApple OSS Distributions
99*d4514f0bSApple OSS Distributions SYSCTL_QUAD(_net_classq, OID_AUTO, l4s_ce_threshold,
100*d4514f0bSApple OSS Distributions CTLFLAG_RW | CTLFLAG_LOCKED, &l4s_ce_threshold,
101*d4514f0bSApple OSS Distributions "L4S CE threshold");
102*d4514f0bSApple OSS Distributions
103*d4514f0bSApple OSS Distributions SYSCTL_UINT(_net_classq_flow_q, OID_AUTO, l4s_local_ce_report,
104*d4514f0bSApple OSS Distributions CTLFLAG_RW | CTLFLAG_LOCKED, &l4s_local_ce_report, 0,
105*d4514f0bSApple OSS Distributions "enable L4S local CE report");
106*d4514f0bSApple OSS Distributions
107*d4514f0bSApple OSS Distributions SYSCTL_QUAD(_net_classq_flow_q, OID_AUTO, pkt_pacing_leeway,
108*d4514f0bSApple OSS Distributions CTLFLAG_RW | CTLFLAG_LOCKED, &pkt_pacing_leeway, "packet pacing leeway");
109*d4514f0bSApple OSS Distributions
110*d4514f0bSApple OSS Distributions SYSCTL_QUAD(_net_classq_flow_q, OID_AUTO, max_pkt_pacing_interval,
111*d4514f0bSApple OSS Distributions CTLFLAG_RW | CTLFLAG_LOCKED, &max_pkt_pacing_interval, "max packet pacing interval");
112*d4514f0bSApple OSS Distributions
113*d4514f0bSApple OSS Distributions SYSCTL_QUAD(_net_classq_flow_q, OID_AUTO, l4s_min_delay_threshold,
114*d4514f0bSApple OSS Distributions CTLFLAG_RW | CTLFLAG_LOCKED, &l4s_min_delay_threshold, "l4s min delay threshold");
115*d4514f0bSApple OSS Distributions #endif /* (DEBUG || DEVELOPMENT) */
116*d4514f0bSApple OSS Distributions
117*d4514f0bSApple OSS Distributions void
fq_codel_init(void)118*d4514f0bSApple OSS Distributions fq_codel_init(void)
119*d4514f0bSApple OSS Distributions {
120*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_AON_FLOW_HIGH_DELAY == 0x8300004);
121*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_AON_THROTTLE == 0x8300008);
122*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_AON_FLOW_OVERWHELMING == 0x830000c);
123*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_AON_FLOW_DQ_STALL == 0x8300010);
124*d4514f0bSApple OSS Distributions
125*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_FLOW_ENQUEUE == 0x8310004);
126*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_FLOW_DEQUEUE == 0x8310008);
127*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_FLOW_CTL == 0x831000c);
128*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_FLOW_ALLOC == 0x8310010);
129*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_FLOW_DESTROY == 0x8310014);
130*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_FLOW_REPORT_CE == 0x8310018);
131*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_STATS_GET_QLEN == 0x831001c);
132*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_TX_NOT_READY == 0x8310020);
133*d4514f0bSApple OSS Distributions _CASSERT(AQM_KTRACE_TX_PACEMAKER == 0x8310024);
134*d4514f0bSApple OSS Distributions }
135*d4514f0bSApple OSS Distributions
136*d4514f0bSApple OSS Distributions fq_t *
fq_alloc(classq_pkt_type_t ptype)137*d4514f0bSApple OSS Distributions fq_alloc(classq_pkt_type_t ptype)
138*d4514f0bSApple OSS Distributions {
139*d4514f0bSApple OSS Distributions fq_t *fq = NULL;
140*d4514f0bSApple OSS Distributions
141*d4514f0bSApple OSS Distributions fq = kalloc_type(fq_t, Z_WAITOK_ZERO);
142*d4514f0bSApple OSS Distributions if (ptype == QP_MBUF) {
143*d4514f0bSApple OSS Distributions MBUFQ_INIT(&fq->fq_mbufq);
144*d4514f0bSApple OSS Distributions }
145*d4514f0bSApple OSS Distributions #if SKYWALK
146*d4514f0bSApple OSS Distributions else {
147*d4514f0bSApple OSS Distributions VERIFY(ptype == QP_PACKET);
148*d4514f0bSApple OSS Distributions KPKTQ_INIT(&fq->fq_kpktq);
149*d4514f0bSApple OSS Distributions }
150*d4514f0bSApple OSS Distributions #endif /* SKYWALK */
151*d4514f0bSApple OSS Distributions CLASSQ_PKT_INIT(&fq->fq_dq_head);
152*d4514f0bSApple OSS Distributions CLASSQ_PKT_INIT(&fq->fq_dq_tail);
153*d4514f0bSApple OSS Distributions fq->fq_in_dqlist = false;
154*d4514f0bSApple OSS Distributions
155*d4514f0bSApple OSS Distributions return fq;
156*d4514f0bSApple OSS Distributions }
157*d4514f0bSApple OSS Distributions
158*d4514f0bSApple OSS Distributions void
fq_destroy(fq_t * fq,classq_pkt_type_t ptype)159*d4514f0bSApple OSS Distributions fq_destroy(fq_t *fq, classq_pkt_type_t ptype)
160*d4514f0bSApple OSS Distributions {
161*d4514f0bSApple OSS Distributions VERIFY(!fq->fq_in_dqlist);
162*d4514f0bSApple OSS Distributions VERIFY(fq_empty(fq, ptype));
163*d4514f0bSApple OSS Distributions VERIFY(!(fq->fq_flags & (FQF_NEW_FLOW | FQF_OLD_FLOW |
164*d4514f0bSApple OSS Distributions FQF_EMPTY_FLOW)));
165*d4514f0bSApple OSS Distributions VERIFY(fq->fq_bytes == 0);
166*d4514f0bSApple OSS Distributions kfree_type(fq_t, fq);
167*d4514f0bSApple OSS Distributions }
168*d4514f0bSApple OSS Distributions
169*d4514f0bSApple OSS Distributions static inline void
fq_detect_dequeue_stall(fq_if_t * fqs,fq_t * flowq,fq_if_classq_t * fq_cl,u_int64_t * now)170*d4514f0bSApple OSS Distributions fq_detect_dequeue_stall(fq_if_t *fqs, fq_t *flowq, fq_if_classq_t *fq_cl,
171*d4514f0bSApple OSS Distributions u_int64_t *now)
172*d4514f0bSApple OSS Distributions {
173*d4514f0bSApple OSS Distributions u_int64_t maxgetqtime, update_interval;
174*d4514f0bSApple OSS Distributions if (FQ_IS_DELAY_HIGH(flowq) || flowq->fq_getqtime == 0 ||
175*d4514f0bSApple OSS Distributions fq_empty(flowq, fqs->fqs_ptype) ||
176*d4514f0bSApple OSS Distributions flowq->fq_bytes < FQ_MIN_FC_THRESHOLD_BYTES) {
177*d4514f0bSApple OSS Distributions return;
178*d4514f0bSApple OSS Distributions }
179*d4514f0bSApple OSS Distributions
180*d4514f0bSApple OSS Distributions update_interval = FQ_UPDATE_INTERVAL(flowq);
181*d4514f0bSApple OSS Distributions maxgetqtime = flowq->fq_getqtime + update_interval;
182*d4514f0bSApple OSS Distributions if ((*now) > maxgetqtime) {
183*d4514f0bSApple OSS Distributions /*
184*d4514f0bSApple OSS Distributions * there was no dequeue in an update interval worth of
185*d4514f0bSApple OSS Distributions * time. It means that the queue is stalled.
186*d4514f0bSApple OSS Distributions */
187*d4514f0bSApple OSS Distributions FQ_SET_DELAY_HIGH(flowq);
188*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_dequeue_stall++;
189*d4514f0bSApple OSS Distributions os_log_error(OS_LOG_DEFAULT, "%s:num: %d, "
190*d4514f0bSApple OSS Distributions "scidx: %d, flow: 0x%x, iface: %s grp: %hhu", __func__,
191*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_dequeue_stall, flowq->fq_sc_index,
192*d4514f0bSApple OSS Distributions flowq->fq_flowhash, if_name(fqs->fqs_ifq->ifcq_ifp),
193*d4514f0bSApple OSS Distributions FQ_GROUP(flowq)->fqg_index);
194*d4514f0bSApple OSS Distributions KDBG(AQM_KTRACE_AON_FLOW_DQ_STALL, flowq->fq_flowhash,
195*d4514f0bSApple OSS Distributions AQM_KTRACE_FQ_GRP_SC_IDX(flowq), flowq->fq_bytes,
196*d4514f0bSApple OSS Distributions (*now) - flowq->fq_getqtime);
197*d4514f0bSApple OSS Distributions }
198*d4514f0bSApple OSS Distributions }
199*d4514f0bSApple OSS Distributions
200*d4514f0bSApple OSS Distributions void
fq_head_drop(fq_if_t * fqs,fq_t * fq)201*d4514f0bSApple OSS Distributions fq_head_drop(fq_if_t *fqs, fq_t *fq)
202*d4514f0bSApple OSS Distributions {
203*d4514f0bSApple OSS Distributions pktsched_pkt_t pkt;
204*d4514f0bSApple OSS Distributions volatile uint32_t *__single pkt_flags;
205*d4514f0bSApple OSS Distributions uint64_t *__single pkt_timestamp;
206*d4514f0bSApple OSS Distributions struct ifclassq *ifq = fqs->fqs_ifq;
207*d4514f0bSApple OSS Distributions
208*d4514f0bSApple OSS Distributions _PKTSCHED_PKT_INIT(&pkt);
209*d4514f0bSApple OSS Distributions fq_getq_flow_internal(fqs, fq, &pkt);
210*d4514f0bSApple OSS Distributions if (pkt.pktsched_pkt_mbuf == NULL) {
211*d4514f0bSApple OSS Distributions return;
212*d4514f0bSApple OSS Distributions }
213*d4514f0bSApple OSS Distributions
214*d4514f0bSApple OSS Distributions pktsched_get_pkt_vars(&pkt, &pkt_flags, &pkt_timestamp, NULL, NULL,
215*d4514f0bSApple OSS Distributions NULL, NULL, NULL);
216*d4514f0bSApple OSS Distributions
217*d4514f0bSApple OSS Distributions *pkt_timestamp = 0;
218*d4514f0bSApple OSS Distributions switch (pkt.pktsched_ptype) {
219*d4514f0bSApple OSS Distributions case QP_MBUF:
220*d4514f0bSApple OSS Distributions *pkt_flags &= ~PKTF_PRIV_GUARDED;
221*d4514f0bSApple OSS Distributions break;
222*d4514f0bSApple OSS Distributions #if SKYWALK
223*d4514f0bSApple OSS Distributions case QP_PACKET:
224*d4514f0bSApple OSS Distributions /* sanity check */
225*d4514f0bSApple OSS Distributions ASSERT((*pkt_flags & ~PKT_F_COMMON_MASK) == 0);
226*d4514f0bSApple OSS Distributions break;
227*d4514f0bSApple OSS Distributions #endif /* SKYWALK */
228*d4514f0bSApple OSS Distributions default:
229*d4514f0bSApple OSS Distributions VERIFY(0);
230*d4514f0bSApple OSS Distributions /* NOTREACHED */
231*d4514f0bSApple OSS Distributions __builtin_unreachable();
232*d4514f0bSApple OSS Distributions }
233*d4514f0bSApple OSS Distributions
234*d4514f0bSApple OSS Distributions IFCQ_DROP_ADD(ifq, 1, pktsched_get_pkt_len(&pkt));
235*d4514f0bSApple OSS Distributions IFCQ_CONVERT_LOCK(ifq);
236*d4514f0bSApple OSS Distributions pktsched_free_pkt(&pkt);
237*d4514f0bSApple OSS Distributions }
238*d4514f0bSApple OSS Distributions
239*d4514f0bSApple OSS Distributions
240*d4514f0bSApple OSS Distributions static int
fq_compressor(fq_if_t * fqs,fq_t * fq,fq_if_classq_t * fq_cl,pktsched_pkt_t * pkt)241*d4514f0bSApple OSS Distributions fq_compressor(fq_if_t *fqs, fq_t *fq, fq_if_classq_t *fq_cl,
242*d4514f0bSApple OSS Distributions pktsched_pkt_t *pkt)
243*d4514f0bSApple OSS Distributions {
244*d4514f0bSApple OSS Distributions classq_pkt_type_t ptype = fqs->fqs_ptype;
245*d4514f0bSApple OSS Distributions uint32_t comp_gencnt = 0;
246*d4514f0bSApple OSS Distributions uint64_t *__single pkt_timestamp;
247*d4514f0bSApple OSS Distributions uint64_t old_timestamp = 0;
248*d4514f0bSApple OSS Distributions uint32_t old_pktlen = 0;
249*d4514f0bSApple OSS Distributions struct ifclassq *ifq = fqs->fqs_ifq;
250*d4514f0bSApple OSS Distributions
251*d4514f0bSApple OSS Distributions if (__improbable(pkt_compressor == 0)) {
252*d4514f0bSApple OSS Distributions return 0;
253*d4514f0bSApple OSS Distributions }
254*d4514f0bSApple OSS Distributions
255*d4514f0bSApple OSS Distributions pktsched_get_pkt_vars(pkt, NULL, &pkt_timestamp, NULL, NULL, NULL,
256*d4514f0bSApple OSS Distributions &comp_gencnt, NULL);
257*d4514f0bSApple OSS Distributions
258*d4514f0bSApple OSS Distributions if (comp_gencnt == 0) {
259*d4514f0bSApple OSS Distributions return 0;
260*d4514f0bSApple OSS Distributions }
261*d4514f0bSApple OSS Distributions
262*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_pkts_compressible++;
263*d4514f0bSApple OSS Distributions
264*d4514f0bSApple OSS Distributions if (fq_empty(fq, fqs->fqs_ptype)) {
265*d4514f0bSApple OSS Distributions return 0;
266*d4514f0bSApple OSS Distributions }
267*d4514f0bSApple OSS Distributions
268*d4514f0bSApple OSS Distributions if (ptype == QP_MBUF) {
269*d4514f0bSApple OSS Distributions struct mbuf *m = MBUFQ_LAST(&fq->fq_mbufq);
270*d4514f0bSApple OSS Distributions
271*d4514f0bSApple OSS Distributions if (comp_gencnt != m->m_pkthdr.comp_gencnt) {
272*d4514f0bSApple OSS Distributions return 0;
273*d4514f0bSApple OSS Distributions }
274*d4514f0bSApple OSS Distributions
275*d4514f0bSApple OSS Distributions /* If we got until here, we should merge/replace the segment */
276*d4514f0bSApple OSS Distributions MBUFQ_REMOVE(&fq->fq_mbufq, m);
277*d4514f0bSApple OSS Distributions old_pktlen = m_pktlen(m);
278*d4514f0bSApple OSS Distributions old_timestamp = m->m_pkthdr.pkt_timestamp;
279*d4514f0bSApple OSS Distributions
280*d4514f0bSApple OSS Distributions IFCQ_CONVERT_LOCK(fqs->fqs_ifq);
281*d4514f0bSApple OSS Distributions
282*d4514f0bSApple OSS Distributions if (__improbable(droptap_verbose > 0)) {
283*d4514f0bSApple OSS Distributions droptap_output_mbuf(m, DROP_REASON_AQM_COMPRESSED, NULL, 0, 0,
284*d4514f0bSApple OSS Distributions fqs->fqs_ifq->ifcq_ifp);
285*d4514f0bSApple OSS Distributions }
286*d4514f0bSApple OSS Distributions
287*d4514f0bSApple OSS Distributions m_freem(m);
288*d4514f0bSApple OSS Distributions }
289*d4514f0bSApple OSS Distributions #if SKYWALK
290*d4514f0bSApple OSS Distributions else {
291*d4514f0bSApple OSS Distributions struct __kern_packet *kpkt = KPKTQ_LAST(&fq->fq_kpktq);
292*d4514f0bSApple OSS Distributions
293*d4514f0bSApple OSS Distributions if (comp_gencnt != kpkt->pkt_comp_gencnt) {
294*d4514f0bSApple OSS Distributions return 0;
295*d4514f0bSApple OSS Distributions }
296*d4514f0bSApple OSS Distributions
297*d4514f0bSApple OSS Distributions /* If we got until here, we should merge/replace the segment */
298*d4514f0bSApple OSS Distributions KPKTQ_REMOVE(&fq->fq_kpktq, kpkt);
299*d4514f0bSApple OSS Distributions old_pktlen = kpkt->pkt_length;
300*d4514f0bSApple OSS Distributions old_timestamp = kpkt->pkt_timestamp;
301*d4514f0bSApple OSS Distributions
302*d4514f0bSApple OSS Distributions IFCQ_CONVERT_LOCK(fqs->fqs_ifq);
303*d4514f0bSApple OSS Distributions
304*d4514f0bSApple OSS Distributions if (__improbable(droptap_verbose > 0)) {
305*d4514f0bSApple OSS Distributions droptap_output_packet(SK_PKT2PH(kpkt), DROP_REASON_AQM_COMPRESSED, NULL, 0, 0,
306*d4514f0bSApple OSS Distributions fqs->fqs_ifq->ifcq_ifp, 0, NULL, -1, NULL, 0, 0);
307*d4514f0bSApple OSS Distributions }
308*d4514f0bSApple OSS Distributions
309*d4514f0bSApple OSS Distributions struct kern_pbufpool *pp =
310*d4514f0bSApple OSS Distributions __DECONST(struct kern_pbufpool *, ((struct __kern_quantum *)kpkt)->qum_pp);
311*d4514f0bSApple OSS Distributions pp_free_packet(pp, (uint64_t)kpkt);
312*d4514f0bSApple OSS Distributions }
313*d4514f0bSApple OSS Distributions #endif /* SKYWALK */
314*d4514f0bSApple OSS Distributions
315*d4514f0bSApple OSS Distributions fq->fq_bytes -= old_pktlen;
316*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_byte_cnt -= old_pktlen;
317*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_pkt_cnt--;
318*d4514f0bSApple OSS Distributions IFCQ_DEC_LEN(ifq);
319*d4514f0bSApple OSS Distributions IFCQ_DEC_BYTES(ifq, old_pktlen);
320*d4514f0bSApple OSS Distributions
321*d4514f0bSApple OSS Distributions FQ_GRP_DEC_LEN(fq);
322*d4514f0bSApple OSS Distributions FQ_GRP_DEC_BYTES(fq, old_pktlen);
323*d4514f0bSApple OSS Distributions
324*d4514f0bSApple OSS Distributions *pkt_timestamp = old_timestamp;
325*d4514f0bSApple OSS Distributions
326*d4514f0bSApple OSS Distributions return CLASSQEQ_COMPRESSED;
327*d4514f0bSApple OSS Distributions }
328*d4514f0bSApple OSS Distributions
329*d4514f0bSApple OSS Distributions int
fq_addq(fq_if_t * fqs,fq_if_group_t * fq_grp,pktsched_pkt_t * pkt,fq_if_classq_t * fq_cl)330*d4514f0bSApple OSS Distributions fq_addq(fq_if_t *fqs, fq_if_group_t *fq_grp, pktsched_pkt_t *pkt,
331*d4514f0bSApple OSS Distributions fq_if_classq_t *fq_cl)
332*d4514f0bSApple OSS Distributions {
333*d4514f0bSApple OSS Distributions int droptype = DTYPE_NODROP, fc_adv = 0, ret = CLASSQEQ_SUCCESS;
334*d4514f0bSApple OSS Distributions u_int64_t now;
335*d4514f0bSApple OSS Distributions fq_t *fq = NULL;
336*d4514f0bSApple OSS Distributions uint64_t *__single pkt_timestamp;
337*d4514f0bSApple OSS Distributions volatile uint32_t *__single pkt_flags;
338*d4514f0bSApple OSS Distributions uint32_t pkt_flowid, cnt;
339*d4514f0bSApple OSS Distributions uint8_t pkt_proto, pkt_flowsrc;
340*d4514f0bSApple OSS Distributions fq_tfc_type_t tfc_type = FQ_TFC_C;
341*d4514f0bSApple OSS Distributions
342*d4514f0bSApple OSS Distributions cnt = pkt->pktsched_pcnt;
343*d4514f0bSApple OSS Distributions pktsched_get_pkt_vars(pkt, &pkt_flags, &pkt_timestamp, &pkt_flowid,
344*d4514f0bSApple OSS Distributions &pkt_flowsrc, &pkt_proto, NULL, NULL);
345*d4514f0bSApple OSS Distributions
346*d4514f0bSApple OSS Distributions /*
347*d4514f0bSApple OSS Distributions * XXX Not walking the chain to set this flag on every packet.
348*d4514f0bSApple OSS Distributions * This flag is only used for debugging. Nothing is affected if it's
349*d4514f0bSApple OSS Distributions * not set.
350*d4514f0bSApple OSS Distributions */
351*d4514f0bSApple OSS Distributions switch (pkt->pktsched_ptype) {
352*d4514f0bSApple OSS Distributions case QP_MBUF:
353*d4514f0bSApple OSS Distributions /* See comments in <rdar://problem/14040693> */
354*d4514f0bSApple OSS Distributions VERIFY(!(*pkt_flags & PKTF_PRIV_GUARDED));
355*d4514f0bSApple OSS Distributions *pkt_flags |= PKTF_PRIV_GUARDED;
356*d4514f0bSApple OSS Distributions break;
357*d4514f0bSApple OSS Distributions #if SKYWALK
358*d4514f0bSApple OSS Distributions case QP_PACKET:
359*d4514f0bSApple OSS Distributions /* sanity check */
360*d4514f0bSApple OSS Distributions ASSERT((*pkt_flags & ~PKT_F_COMMON_MASK) == 0);
361*d4514f0bSApple OSS Distributions break;
362*d4514f0bSApple OSS Distributions #endif /* SKYWALK */
363*d4514f0bSApple OSS Distributions default:
364*d4514f0bSApple OSS Distributions VERIFY(0);
365*d4514f0bSApple OSS Distributions /* NOTREACHED */
366*d4514f0bSApple OSS Distributions __builtin_unreachable();
367*d4514f0bSApple OSS Distributions }
368*d4514f0bSApple OSS Distributions
369*d4514f0bSApple OSS Distributions if (ifclassq_enable_l4s) {
370*d4514f0bSApple OSS Distributions tfc_type = pktsched_is_pkt_l4s(pkt) ? FQ_TFC_L4S : FQ_TFC_C;
371*d4514f0bSApple OSS Distributions }
372*d4514f0bSApple OSS Distributions
373*d4514f0bSApple OSS Distributions /*
374*d4514f0bSApple OSS Distributions * Timestamps for every packet must be set prior to entering this path.
375*d4514f0bSApple OSS Distributions */
376*d4514f0bSApple OSS Distributions now = *pkt_timestamp;
377*d4514f0bSApple OSS Distributions ASSERT(now > 0);
378*d4514f0bSApple OSS Distributions
379*d4514f0bSApple OSS Distributions /* find the flowq for this packet */
380*d4514f0bSApple OSS Distributions fq = fq_if_hash_pkt(fqs, fq_grp, pkt_flowid, pktsched_get_pkt_svc(pkt),
381*d4514f0bSApple OSS Distributions now, true, tfc_type);
382*d4514f0bSApple OSS Distributions if (__improbable(fq == NULL)) {
383*d4514f0bSApple OSS Distributions DTRACE_IP1(memfail__drop, fq_if_t *, fqs);
384*d4514f0bSApple OSS Distributions /* drop the packet if we could not allocate a flow queue */
385*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_drop_memfailure += cnt;
386*d4514f0bSApple OSS Distributions return CLASSQEQ_DROP;
387*d4514f0bSApple OSS Distributions }
388*d4514f0bSApple OSS Distributions VERIFY(fq->fq_group == fq_grp);
389*d4514f0bSApple OSS Distributions VERIFY(fqs->fqs_ptype == pkt->pktsched_ptype);
390*d4514f0bSApple OSS Distributions
391*d4514f0bSApple OSS Distributions KDBG(AQM_KTRACE_STATS_FLOW_ENQUEUE, fq->fq_flowhash,
392*d4514f0bSApple OSS Distributions AQM_KTRACE_FQ_GRP_SC_IDX(fq),
393*d4514f0bSApple OSS Distributions fq->fq_bytes, pktsched_get_pkt_len(pkt));
394*d4514f0bSApple OSS Distributions
395*d4514f0bSApple OSS Distributions fq_detect_dequeue_stall(fqs, fq, fq_cl, &now);
396*d4514f0bSApple OSS Distributions
397*d4514f0bSApple OSS Distributions /*
398*d4514f0bSApple OSS Distributions * Skip the dropping part if it's L4S. Flow control or ECN marking decision
399*d4514f0bSApple OSS Distributions * will be made at dequeue time.
400*d4514f0bSApple OSS Distributions */
401*d4514f0bSApple OSS Distributions if (ifclassq_enable_l4s && tfc_type == FQ_TFC_L4S) {
402*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_l4s_pkts += cnt;
403*d4514f0bSApple OSS Distributions droptype = DTYPE_NODROP;
404*d4514f0bSApple OSS Distributions }
405*d4514f0bSApple OSS Distributions
406*d4514f0bSApple OSS Distributions if (__improbable(FQ_IS_DELAY_HIGH(fq) || FQ_IS_OVERWHELMING(fq))) {
407*d4514f0bSApple OSS Distributions if ((fq->fq_flags & FQF_FLOWCTL_CAPABLE) &&
408*d4514f0bSApple OSS Distributions (*pkt_flags & PKTF_FLOW_ADV)) {
409*d4514f0bSApple OSS Distributions fc_adv = 1;
410*d4514f0bSApple OSS Distributions /*
411*d4514f0bSApple OSS Distributions * If the flow is suspended or it is not
412*d4514f0bSApple OSS Distributions * TCP/QUIC, drop the chain.
413*d4514f0bSApple OSS Distributions */
414*d4514f0bSApple OSS Distributions if ((pkt_proto != IPPROTO_TCP) &&
415*d4514f0bSApple OSS Distributions (pkt_proto != IPPROTO_QUIC)) {
416*d4514f0bSApple OSS Distributions droptype = DTYPE_EARLY;
417*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_drop_early += cnt;
418*d4514f0bSApple OSS Distributions IFCQ_DROP_ADD(fqs->fqs_ifq, cnt, pktsched_get_pkt_len(pkt));
419*d4514f0bSApple OSS Distributions }
420*d4514f0bSApple OSS Distributions DTRACE_IP6(flow__adv, fq_if_t *, fqs,
421*d4514f0bSApple OSS Distributions fq_if_classq_t *, fq_cl, fq_t *, fq,
422*d4514f0bSApple OSS Distributions int, droptype, pktsched_pkt_t *, pkt,
423*d4514f0bSApple OSS Distributions uint32_t, cnt);
424*d4514f0bSApple OSS Distributions } else {
425*d4514f0bSApple OSS Distributions /*
426*d4514f0bSApple OSS Distributions * Need to drop packets to make room for the new
427*d4514f0bSApple OSS Distributions * ones. Try to drop from the head of the queue
428*d4514f0bSApple OSS Distributions * instead of the latest packets.
429*d4514f0bSApple OSS Distributions */
430*d4514f0bSApple OSS Distributions if (!fq_empty(fq, fqs->fqs_ptype)) {
431*d4514f0bSApple OSS Distributions uint32_t i;
432*d4514f0bSApple OSS Distributions
433*d4514f0bSApple OSS Distributions for (i = 0; i < cnt; i++) {
434*d4514f0bSApple OSS Distributions fq_head_drop(fqs, fq);
435*d4514f0bSApple OSS Distributions }
436*d4514f0bSApple OSS Distributions droptype = DTYPE_NODROP;
437*d4514f0bSApple OSS Distributions } else {
438*d4514f0bSApple OSS Distributions droptype = DTYPE_EARLY;
439*d4514f0bSApple OSS Distributions }
440*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_drop_early += cnt;
441*d4514f0bSApple OSS Distributions
442*d4514f0bSApple OSS Distributions DTRACE_IP6(no__flow__adv, fq_if_t *, fqs,
443*d4514f0bSApple OSS Distributions fq_if_classq_t *, fq_cl, fq_t *, fq,
444*d4514f0bSApple OSS Distributions int, droptype, pktsched_pkt_t *, pkt,
445*d4514f0bSApple OSS Distributions uint32_t, cnt);
446*d4514f0bSApple OSS Distributions }
447*d4514f0bSApple OSS Distributions }
448*d4514f0bSApple OSS Distributions
449*d4514f0bSApple OSS Distributions /* Set the return code correctly */
450*d4514f0bSApple OSS Distributions if (__improbable(fc_adv == 1 && droptype != DTYPE_FORCED)) {
451*d4514f0bSApple OSS Distributions if (fq_if_add_fcentry(fqs, pkt, pkt_flowsrc, fq, fq_cl)) {
452*d4514f0bSApple OSS Distributions fq->fq_flags |= FQF_FLOWCTL_ON;
453*d4514f0bSApple OSS Distributions /* deliver flow control advisory error */
454*d4514f0bSApple OSS Distributions if (droptype == DTYPE_NODROP) {
455*d4514f0bSApple OSS Distributions ret = CLASSQEQ_SUCCESS_FC;
456*d4514f0bSApple OSS Distributions } else {
457*d4514f0bSApple OSS Distributions /* dropped due to flow control */
458*d4514f0bSApple OSS Distributions ret = CLASSQEQ_DROP_FC;
459*d4514f0bSApple OSS Distributions }
460*d4514f0bSApple OSS Distributions } else {
461*d4514f0bSApple OSS Distributions /*
462*d4514f0bSApple OSS Distributions * if we could not flow control the flow, it is
463*d4514f0bSApple OSS Distributions * better to drop
464*d4514f0bSApple OSS Distributions */
465*d4514f0bSApple OSS Distributions droptype = DTYPE_FORCED;
466*d4514f0bSApple OSS Distributions ret = CLASSQEQ_DROP_FC;
467*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_flow_control_fail++;
468*d4514f0bSApple OSS Distributions }
469*d4514f0bSApple OSS Distributions DTRACE_IP3(fc__ret, fq_if_t *, fqs, int, droptype, int, ret);
470*d4514f0bSApple OSS Distributions }
471*d4514f0bSApple OSS Distributions
472*d4514f0bSApple OSS Distributions /*
473*d4514f0bSApple OSS Distributions * If the queue length hits the queue limit, drop a chain with the
474*d4514f0bSApple OSS Distributions * same number of packets from the front of the queue for a flow with
475*d4514f0bSApple OSS Distributions * maximum number of bytes. This will penalize heavy and unresponsive
476*d4514f0bSApple OSS Distributions * flows. It will also avoid a tail drop.
477*d4514f0bSApple OSS Distributions */
478*d4514f0bSApple OSS Distributions if (__improbable(droptype == DTYPE_NODROP &&
479*d4514f0bSApple OSS Distributions fq_if_at_drop_limit(fqs))) {
480*d4514f0bSApple OSS Distributions uint32_t i;
481*d4514f0bSApple OSS Distributions
482*d4514f0bSApple OSS Distributions if (fqs->fqs_large_flow == fq) {
483*d4514f0bSApple OSS Distributions /*
484*d4514f0bSApple OSS Distributions * Drop from the head of the current fq. Since a
485*d4514f0bSApple OSS Distributions * new packet will be added to the tail, it is ok
486*d4514f0bSApple OSS Distributions * to leave fq in place.
487*d4514f0bSApple OSS Distributions */
488*d4514f0bSApple OSS Distributions DTRACE_IP5(large__flow, fq_if_t *, fqs,
489*d4514f0bSApple OSS Distributions fq_if_classq_t *, fq_cl, fq_t *, fq,
490*d4514f0bSApple OSS Distributions pktsched_pkt_t *, pkt, uint32_t, cnt);
491*d4514f0bSApple OSS Distributions
492*d4514f0bSApple OSS Distributions for (i = 0; i < cnt; i++) {
493*d4514f0bSApple OSS Distributions fq_head_drop(fqs, fq);
494*d4514f0bSApple OSS Distributions }
495*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_drop_overflow += cnt;
496*d4514f0bSApple OSS Distributions
497*d4514f0bSApple OSS Distributions /*
498*d4514f0bSApple OSS Distributions * TCP and QUIC will react to the loss of those head dropped pkts
499*d4514f0bSApple OSS Distributions * and adjust send rate.
500*d4514f0bSApple OSS Distributions */
501*d4514f0bSApple OSS Distributions if ((fq->fq_flags & FQF_FLOWCTL_CAPABLE) &&
502*d4514f0bSApple OSS Distributions (*pkt_flags & PKTF_FLOW_ADV) &&
503*d4514f0bSApple OSS Distributions (pkt_proto != IPPROTO_TCP) &&
504*d4514f0bSApple OSS Distributions (pkt_proto != IPPROTO_QUIC)) {
505*d4514f0bSApple OSS Distributions if (fq_if_add_fcentry(fqs, pkt, pkt_flowsrc, fq, fq_cl)) {
506*d4514f0bSApple OSS Distributions fq->fq_flags |= FQF_FLOWCTL_ON;
507*d4514f0bSApple OSS Distributions FQ_SET_OVERWHELMING(fq);
508*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_overwhelming++;
509*d4514f0bSApple OSS Distributions /* deliver flow control advisory error */
510*d4514f0bSApple OSS Distributions ret = CLASSQEQ_SUCCESS_FC;
511*d4514f0bSApple OSS Distributions }
512*d4514f0bSApple OSS Distributions }
513*d4514f0bSApple OSS Distributions } else {
514*d4514f0bSApple OSS Distributions if (fqs->fqs_large_flow == NULL) {
515*d4514f0bSApple OSS Distributions droptype = DTYPE_FORCED;
516*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_drop_overflow += cnt;
517*d4514f0bSApple OSS Distributions ret = CLASSQEQ_DROP;
518*d4514f0bSApple OSS Distributions
519*d4514f0bSApple OSS Distributions DTRACE_IP5(no__large__flow, fq_if_t *, fqs,
520*d4514f0bSApple OSS Distributions fq_if_classq_t *, fq_cl, fq_t *, fq,
521*d4514f0bSApple OSS Distributions pktsched_pkt_t *, pkt, uint32_t, cnt);
522*d4514f0bSApple OSS Distributions
523*d4514f0bSApple OSS Distributions /*
524*d4514f0bSApple OSS Distributions * if this fq was freshly created and there
525*d4514f0bSApple OSS Distributions * is nothing to enqueue, move it to empty list
526*d4514f0bSApple OSS Distributions */
527*d4514f0bSApple OSS Distributions if (fq_empty(fq, fqs->fqs_ptype) &&
528*d4514f0bSApple OSS Distributions !(fq->fq_flags & (FQF_NEW_FLOW |
529*d4514f0bSApple OSS Distributions FQF_OLD_FLOW))) {
530*d4514f0bSApple OSS Distributions fq_if_move_to_empty_flow(fqs, fq_cl,
531*d4514f0bSApple OSS Distributions fq, now);
532*d4514f0bSApple OSS Distributions fq = NULL;
533*d4514f0bSApple OSS Distributions }
534*d4514f0bSApple OSS Distributions } else {
535*d4514f0bSApple OSS Distributions DTRACE_IP5(different__large__flow,
536*d4514f0bSApple OSS Distributions fq_if_t *, fqs, fq_if_classq_t *, fq_cl,
537*d4514f0bSApple OSS Distributions fq_t *, fq, pktsched_pkt_t *, pkt,
538*d4514f0bSApple OSS Distributions uint32_t, cnt);
539*d4514f0bSApple OSS Distributions
540*d4514f0bSApple OSS Distributions for (i = 0; i < cnt; i++) {
541*d4514f0bSApple OSS Distributions fq_if_drop_packet(fqs, now);
542*d4514f0bSApple OSS Distributions }
543*d4514f0bSApple OSS Distributions }
544*d4514f0bSApple OSS Distributions }
545*d4514f0bSApple OSS Distributions }
546*d4514f0bSApple OSS Distributions
547*d4514f0bSApple OSS Distributions fq_cl->fcl_flags &= ~FCL_PACED;
548*d4514f0bSApple OSS Distributions
549*d4514f0bSApple OSS Distributions if (__probable(droptype == DTYPE_NODROP)) {
550*d4514f0bSApple OSS Distributions uint32_t chain_len = pktsched_get_pkt_len(pkt);
551*d4514f0bSApple OSS Distributions int ret_compress = 0;
552*d4514f0bSApple OSS Distributions
553*d4514f0bSApple OSS Distributions /*
554*d4514f0bSApple OSS Distributions * We do not compress if we are enqueuing a chain.
555*d4514f0bSApple OSS Distributions * Traversing the chain to look for acks would defeat the
556*d4514f0bSApple OSS Distributions * purpose of batch enqueueing.
557*d4514f0bSApple OSS Distributions */
558*d4514f0bSApple OSS Distributions if (cnt == 1) {
559*d4514f0bSApple OSS Distributions ret_compress = fq_compressor(fqs, fq, fq_cl, pkt);
560*d4514f0bSApple OSS Distributions if (ret_compress == CLASSQEQ_COMPRESSED) {
561*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_pkts_compressed++;
562*d4514f0bSApple OSS Distributions }
563*d4514f0bSApple OSS Distributions }
564*d4514f0bSApple OSS Distributions DTRACE_IP5(fq_enqueue, fq_if_t *, fqs, fq_if_classq_t *, fq_cl,
565*d4514f0bSApple OSS Distributions fq_t *, fq, pktsched_pkt_t *, pkt, uint32_t, cnt);
566*d4514f0bSApple OSS Distributions fq_enqueue(fq, pkt->pktsched_pkt, pkt->pktsched_tail, cnt,
567*d4514f0bSApple OSS Distributions pkt->pktsched_ptype);
568*d4514f0bSApple OSS Distributions
569*d4514f0bSApple OSS Distributions fq->fq_bytes += chain_len;
570*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_byte_cnt += chain_len;
571*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_pkt_cnt += cnt;
572*d4514f0bSApple OSS Distributions
573*d4514f0bSApple OSS Distributions /*
574*d4514f0bSApple OSS Distributions * check if this queue will qualify to be the next
575*d4514f0bSApple OSS Distributions * victim queue
576*d4514f0bSApple OSS Distributions */
577*d4514f0bSApple OSS Distributions fq_if_is_flow_heavy(fqs, fq);
578*d4514f0bSApple OSS Distributions } else {
579*d4514f0bSApple OSS Distributions DTRACE_IP3(fq_drop, fq_if_t *, fqs, int, droptype, int, ret);
580*d4514f0bSApple OSS Distributions return (ret != CLASSQEQ_SUCCESS) ? ret : CLASSQEQ_DROP;
581*d4514f0bSApple OSS Distributions }
582*d4514f0bSApple OSS Distributions
583*d4514f0bSApple OSS Distributions /*
584*d4514f0bSApple OSS Distributions * If the queue is not currently active, add it to the end of new
585*d4514f0bSApple OSS Distributions * flows list for that service class.
586*d4514f0bSApple OSS Distributions */
587*d4514f0bSApple OSS Distributions if ((fq->fq_flags & (FQF_NEW_FLOW | FQF_OLD_FLOW)) == 0) {
588*d4514f0bSApple OSS Distributions VERIFY(STAILQ_NEXT(fq, fq_actlink) == NULL);
589*d4514f0bSApple OSS Distributions STAILQ_INSERT_TAIL(&fq_cl->fcl_new_flows, fq, fq_actlink);
590*d4514f0bSApple OSS Distributions fq->fq_flags |= FQF_NEW_FLOW;
591*d4514f0bSApple OSS Distributions
592*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_newflows_cnt++;
593*d4514f0bSApple OSS Distributions
594*d4514f0bSApple OSS Distributions fq->fq_deficit = fq_cl->fcl_quantum;
595*d4514f0bSApple OSS Distributions }
596*d4514f0bSApple OSS Distributions return ret;
597*d4514f0bSApple OSS Distributions }
598*d4514f0bSApple OSS Distributions
599*d4514f0bSApple OSS Distributions void
fq_getq_flow_internal(fq_if_t * fqs,fq_t * fq,pktsched_pkt_t * pkt)600*d4514f0bSApple OSS Distributions fq_getq_flow_internal(fq_if_t *fqs, fq_t *fq, pktsched_pkt_t *pkt)
601*d4514f0bSApple OSS Distributions {
602*d4514f0bSApple OSS Distributions classq_pkt_t p = CLASSQ_PKT_INITIALIZER(p);
603*d4514f0bSApple OSS Distributions uint32_t plen;
604*d4514f0bSApple OSS Distributions fq_if_classq_t *fq_cl;
605*d4514f0bSApple OSS Distributions struct ifclassq *ifq = fqs->fqs_ifq;
606*d4514f0bSApple OSS Distributions
607*d4514f0bSApple OSS Distributions fq_dequeue(fq, &p, fqs->fqs_ptype);
608*d4514f0bSApple OSS Distributions if (p.cp_ptype == QP_INVALID) {
609*d4514f0bSApple OSS Distributions VERIFY(p.cp_mbuf == NULL);
610*d4514f0bSApple OSS Distributions return;
611*d4514f0bSApple OSS Distributions }
612*d4514f0bSApple OSS Distributions
613*d4514f0bSApple OSS Distributions fq->fq_next_tx_time = FQ_INVALID_TX_TS;
614*d4514f0bSApple OSS Distributions
615*d4514f0bSApple OSS Distributions pktsched_pkt_encap(pkt, &p);
616*d4514f0bSApple OSS Distributions plen = pktsched_get_pkt_len(pkt);
617*d4514f0bSApple OSS Distributions
618*d4514f0bSApple OSS Distributions VERIFY(fq->fq_bytes >= plen);
619*d4514f0bSApple OSS Distributions fq->fq_bytes -= plen;
620*d4514f0bSApple OSS Distributions
621*d4514f0bSApple OSS Distributions fq_cl = &FQ_CLASSQ(fq);
622*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_byte_cnt -= plen;
623*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_pkt_cnt--;
624*d4514f0bSApple OSS Distributions fq_cl->fcl_flags &= ~FCL_PACED;
625*d4514f0bSApple OSS Distributions
626*d4514f0bSApple OSS Distributions IFCQ_DEC_LEN(ifq);
627*d4514f0bSApple OSS Distributions IFCQ_DEC_BYTES(ifq, plen);
628*d4514f0bSApple OSS Distributions
629*d4514f0bSApple OSS Distributions FQ_GRP_DEC_LEN(fq);
630*d4514f0bSApple OSS Distributions FQ_GRP_DEC_BYTES(fq, plen);
631*d4514f0bSApple OSS Distributions
632*d4514f0bSApple OSS Distributions /* Reset getqtime so that we don't count idle times */
633*d4514f0bSApple OSS Distributions if (fq_empty(fq, fqs->fqs_ptype)) {
634*d4514f0bSApple OSS Distributions fq->fq_getqtime = 0;
635*d4514f0bSApple OSS Distributions }
636*d4514f0bSApple OSS Distributions }
637*d4514f0bSApple OSS Distributions
638*d4514f0bSApple OSS Distributions /*
639*d4514f0bSApple OSS Distributions * fq_get_next_tx_time returns FQ_INVALID_TX_TS when there is no tx time in fq
640*d4514f0bSApple OSS Distributions */
641*d4514f0bSApple OSS Distributions static uint64_t
fq_get_next_tx_time(fq_if_t * fqs,fq_t * fq)642*d4514f0bSApple OSS Distributions fq_get_next_tx_time(fq_if_t *fqs, fq_t *fq)
643*d4514f0bSApple OSS Distributions {
644*d4514f0bSApple OSS Distributions uint64_t tx_time = FQ_INVALID_TX_TS;
645*d4514f0bSApple OSS Distributions
646*d4514f0bSApple OSS Distributions /*
647*d4514f0bSApple OSS Distributions * Check the cached value in fq
648*d4514f0bSApple OSS Distributions */
649*d4514f0bSApple OSS Distributions if (fq->fq_next_tx_time != FQ_INVALID_TX_TS) {
650*d4514f0bSApple OSS Distributions return fq->fq_next_tx_time;
651*d4514f0bSApple OSS Distributions }
652*d4514f0bSApple OSS Distributions
653*d4514f0bSApple OSS Distributions switch (fqs->fqs_ptype) {
654*d4514f0bSApple OSS Distributions case QP_MBUF: {
655*d4514f0bSApple OSS Distributions struct mbuf *m;
656*d4514f0bSApple OSS Distributions if ((m = MBUFQ_FIRST(&fq->fq_mbufq)) != NULL) {
657*d4514f0bSApple OSS Distributions struct m_tag *tag;
658*d4514f0bSApple OSS Distributions tag = m_tag_locate(m, KERNEL_MODULE_TAG_ID,
659*d4514f0bSApple OSS Distributions KERNEL_TAG_TYPE_AQM);
660*d4514f0bSApple OSS Distributions if (tag != NULL) {
661*d4514f0bSApple OSS Distributions tx_time = *(uint64_t *)tag->m_tag_data;
662*d4514f0bSApple OSS Distributions }
663*d4514f0bSApple OSS Distributions }
664*d4514f0bSApple OSS Distributions break;
665*d4514f0bSApple OSS Distributions }
666*d4514f0bSApple OSS Distributions case QP_PACKET: {
667*d4514f0bSApple OSS Distributions struct __kern_packet *p = KPKTQ_FIRST(&fq->fq_kpktq);
668*d4514f0bSApple OSS Distributions if (__probable(p != NULL)) {
669*d4514f0bSApple OSS Distributions tx_time = __packet_get_tx_timestamp(SK_PKT2PH(p));
670*d4514f0bSApple OSS Distributions }
671*d4514f0bSApple OSS Distributions break;
672*d4514f0bSApple OSS Distributions }
673*d4514f0bSApple OSS Distributions default:
674*d4514f0bSApple OSS Distributions VERIFY(0);
675*d4514f0bSApple OSS Distributions /* NOTREACHED */
676*d4514f0bSApple OSS Distributions __builtin_unreachable();
677*d4514f0bSApple OSS Distributions }
678*d4514f0bSApple OSS Distributions
679*d4514f0bSApple OSS Distributions /*
680*d4514f0bSApple OSS Distributions * Cache the tx time in fq. The cache will be clear after dequeue or drop
681*d4514f0bSApple OSS Distributions * from the fq.
682*d4514f0bSApple OSS Distributions */
683*d4514f0bSApple OSS Distributions fq->fq_next_tx_time = tx_time;
684*d4514f0bSApple OSS Distributions
685*d4514f0bSApple OSS Distributions return tx_time;
686*d4514f0bSApple OSS Distributions }
687*d4514f0bSApple OSS Distributions
688*d4514f0bSApple OSS Distributions /*
689*d4514f0bSApple OSS Distributions * fq_tx_time_ready returns true if the fq is empty so that it doesn't
690*d4514f0bSApple OSS Distributions * affect caller logics that handles empty flow.
691*d4514f0bSApple OSS Distributions */
692*d4514f0bSApple OSS Distributions boolean_t
fq_tx_time_ready(fq_if_t * fqs,fq_t * fq,uint64_t now,uint64_t * ready_time)693*d4514f0bSApple OSS Distributions fq_tx_time_ready(fq_if_t *fqs, fq_t *fq, uint64_t now, uint64_t *ready_time)
694*d4514f0bSApple OSS Distributions {
695*d4514f0bSApple OSS Distributions uint64_t pkt_tx_time;
696*d4514f0bSApple OSS Distributions fq_if_classq_t *fq_cl = &FQ_CLASSQ(fq);
697*d4514f0bSApple OSS Distributions
698*d4514f0bSApple OSS Distributions if (!ifclassq_enable_pacing || !ifclassq_enable_l4s || fq->fq_tfc_type != FQ_TFC_L4S) {
699*d4514f0bSApple OSS Distributions return TRUE;
700*d4514f0bSApple OSS Distributions }
701*d4514f0bSApple OSS Distributions
702*d4514f0bSApple OSS Distributions pkt_tx_time = fq_get_next_tx_time(fqs, fq);
703*d4514f0bSApple OSS Distributions if (ready_time != NULL) {
704*d4514f0bSApple OSS Distributions *ready_time = pkt_tx_time;
705*d4514f0bSApple OSS Distributions }
706*d4514f0bSApple OSS Distributions
707*d4514f0bSApple OSS Distributions if (pkt_tx_time <= now + pkt_pacing_leeway ||
708*d4514f0bSApple OSS Distributions pkt_tx_time == FQ_INVALID_TX_TS) {
709*d4514f0bSApple OSS Distributions return TRUE;
710*d4514f0bSApple OSS Distributions }
711*d4514f0bSApple OSS Distributions
712*d4514f0bSApple OSS Distributions /*
713*d4514f0bSApple OSS Distributions * Ignore the tx time if it's scheduled too far in the future
714*d4514f0bSApple OSS Distributions */
715*d4514f0bSApple OSS Distributions if (pkt_tx_time > max_pkt_pacing_interval + now) {
716*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_ignore_tx_time++;
717*d4514f0bSApple OSS Distributions return TRUE;
718*d4514f0bSApple OSS Distributions }
719*d4514f0bSApple OSS Distributions
720*d4514f0bSApple OSS Distributions ASSERT(pkt_tx_time != FQ_INVALID_TX_TS);
721*d4514f0bSApple OSS Distributions KDBG(AQM_KTRACE_TX_NOT_READY, fq->fq_flowhash,
722*d4514f0bSApple OSS Distributions AQM_KTRACE_FQ_GRP_SC_IDX(fq), now, pkt_tx_time);
723*d4514f0bSApple OSS Distributions return FALSE;
724*d4514f0bSApple OSS Distributions }
725*d4514f0bSApple OSS Distributions
726*d4514f0bSApple OSS Distributions void
fq_getq_flow(fq_if_t * fqs,fq_t * fq,pktsched_pkt_t * pkt,uint64_t now)727*d4514f0bSApple OSS Distributions fq_getq_flow(fq_if_t *fqs, fq_t *fq, pktsched_pkt_t *pkt, uint64_t now)
728*d4514f0bSApple OSS Distributions {
729*d4514f0bSApple OSS Distributions fq_if_classq_t *fq_cl = &FQ_CLASSQ(fq);
730*d4514f0bSApple OSS Distributions int64_t qdelay = 0;
731*d4514f0bSApple OSS Distributions volatile uint32_t *__single pkt_flags;
732*d4514f0bSApple OSS Distributions uint64_t *__single pkt_timestamp, pkt_tx_time = 0, pacing_delay = 0;
733*d4514f0bSApple OSS Distributions uint64_t fq_min_delay_threshold = FQ_TARGET_DELAY(fq);
734*d4514f0bSApple OSS Distributions uint8_t pkt_flowsrc;
735*d4514f0bSApple OSS Distributions boolean_t l4s_pkt;
736*d4514f0bSApple OSS Distributions
737*d4514f0bSApple OSS Distributions fq_getq_flow_internal(fqs, fq, pkt);
738*d4514f0bSApple OSS Distributions if (pkt->pktsched_ptype == QP_INVALID) {
739*d4514f0bSApple OSS Distributions VERIFY(pkt->pktsched_pkt_mbuf == NULL);
740*d4514f0bSApple OSS Distributions return;
741*d4514f0bSApple OSS Distributions }
742*d4514f0bSApple OSS Distributions
743*d4514f0bSApple OSS Distributions pktsched_get_pkt_vars(pkt, &pkt_flags, &pkt_timestamp, NULL, &pkt_flowsrc,
744*d4514f0bSApple OSS Distributions NULL, NULL, &pkt_tx_time);
745*d4514f0bSApple OSS Distributions l4s_pkt = pktsched_is_pkt_l4s(pkt);
746*d4514f0bSApple OSS Distributions if (ifclassq_enable_pacing && ifclassq_enable_l4s) {
747*d4514f0bSApple OSS Distributions if (pkt_tx_time > *pkt_timestamp) {
748*d4514f0bSApple OSS Distributions pacing_delay = pkt_tx_time - *pkt_timestamp;
749*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_paced_pkts++;
750*d4514f0bSApple OSS Distributions DTRACE_SKYWALK3(aqm__pacing__delta, uint64_t, now - pkt_tx_time,
751*d4514f0bSApple OSS Distributions fq_if_t *, fqs, fq_t *, fq);
752*d4514f0bSApple OSS Distributions }
753*d4514f0bSApple OSS Distributions #if (DEVELOPMENT || DEBUG)
754*d4514f0bSApple OSS Distributions else if (pkt_tx_time != 0) {
755*d4514f0bSApple OSS Distributions DTRACE_SKYWALK5(aqm__miss__pacing__delay, uint64_t, *pkt_timestamp,
756*d4514f0bSApple OSS Distributions uint64_t, pkt_tx_time, uint64_t, now, fq_if_t *,
757*d4514f0bSApple OSS Distributions fqs, fq_t *, fq);
758*d4514f0bSApple OSS Distributions }
759*d4514f0bSApple OSS Distributions #endif // (DEVELOPMENT || DEBUG)
760*d4514f0bSApple OSS Distributions }
761*d4514f0bSApple OSS Distributions
762*d4514f0bSApple OSS Distributions /* this will compute qdelay in nanoseconds */
763*d4514f0bSApple OSS Distributions if (now > *pkt_timestamp) {
764*d4514f0bSApple OSS Distributions qdelay = now - *pkt_timestamp;
765*d4514f0bSApple OSS Distributions }
766*d4514f0bSApple OSS Distributions
767*d4514f0bSApple OSS Distributions /* Update min/max/avg qdelay for the respective class */
768*d4514f0bSApple OSS Distributions if (fq_cl->fcl_stat.fcl_min_qdelay == 0 ||
769*d4514f0bSApple OSS Distributions (qdelay > 0 && (u_int64_t)qdelay < fq_cl->fcl_stat.fcl_min_qdelay)) {
770*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_min_qdelay = qdelay;
771*d4514f0bSApple OSS Distributions }
772*d4514f0bSApple OSS Distributions
773*d4514f0bSApple OSS Distributions if (fq_cl->fcl_stat.fcl_max_qdelay == 0 ||
774*d4514f0bSApple OSS Distributions (qdelay > 0 && (u_int64_t)qdelay > fq_cl->fcl_stat.fcl_max_qdelay)) {
775*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_max_qdelay = qdelay;
776*d4514f0bSApple OSS Distributions }
777*d4514f0bSApple OSS Distributions
778*d4514f0bSApple OSS Distributions uint64_t num_dequeues = fq_cl->fcl_stat.fcl_dequeue;
779*d4514f0bSApple OSS Distributions
780*d4514f0bSApple OSS Distributions if (num_dequeues == 0) {
781*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_avg_qdelay = qdelay;
782*d4514f0bSApple OSS Distributions } else if (qdelay > 0) {
783*d4514f0bSApple OSS Distributions uint64_t res = 0;
784*d4514f0bSApple OSS Distributions if (os_add_overflow(num_dequeues, 1, &res)) {
785*d4514f0bSApple OSS Distributions /* Reset the dequeue num and dequeue bytes */
786*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_dequeue = num_dequeues = 0;
787*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_dequeue_bytes = 0;
788*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_avg_qdelay = qdelay;
789*d4514f0bSApple OSS Distributions os_log_info(OS_LOG_DEFAULT, "%s: dequeue num overflow, "
790*d4514f0bSApple OSS Distributions "flow: 0x%x, iface: %s", __func__, fq->fq_flowhash,
791*d4514f0bSApple OSS Distributions if_name(fqs->fqs_ifq->ifcq_ifp));
792*d4514f0bSApple OSS Distributions } else {
793*d4514f0bSApple OSS Distributions uint64_t product = 0;
794*d4514f0bSApple OSS Distributions if (os_mul_overflow(fq_cl->fcl_stat.fcl_avg_qdelay,
795*d4514f0bSApple OSS Distributions num_dequeues, &product) || os_add_overflow(product, qdelay, &res)) {
796*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_avg_qdelay = qdelay;
797*d4514f0bSApple OSS Distributions } else {
798*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_avg_qdelay = res /
799*d4514f0bSApple OSS Distributions (num_dequeues + 1);
800*d4514f0bSApple OSS Distributions }
801*d4514f0bSApple OSS Distributions }
802*d4514f0bSApple OSS Distributions }
803*d4514f0bSApple OSS Distributions
804*d4514f0bSApple OSS Distributions fq->fq_pkts_since_last_report++;
805*d4514f0bSApple OSS Distributions if (ifclassq_enable_l4s && l4s_pkt) {
806*d4514f0bSApple OSS Distributions /*
807*d4514f0bSApple OSS Distributions * A safe guard to make sure that L4S is not going to build a huge
808*d4514f0bSApple OSS Distributions * queue if we encounter unexpected problems (for eg., if ACKs don't
809*d4514f0bSApple OSS Distributions * arrive in timely manner due to congestion in reverse path).
810*d4514f0bSApple OSS Distributions */
811*d4514f0bSApple OSS Distributions fq_min_delay_threshold = l4s_min_delay_threshold;
812*d4514f0bSApple OSS Distributions
813*d4514f0bSApple OSS Distributions if ((l4s_ce_threshold != 0 && qdelay > l4s_ce_threshold + pacing_delay) ||
814*d4514f0bSApple OSS Distributions (l4s_ce_threshold == 0 && qdelay > FQ_TARGET_DELAY(fq) + pacing_delay)) {
815*d4514f0bSApple OSS Distributions DTRACE_SKYWALK4(aqm__mark__ce, uint64_t, qdelay, uint64_t, pacing_delay,
816*d4514f0bSApple OSS Distributions fq_if_t *, fqs, fq_t *, fq);
817*d4514f0bSApple OSS Distributions KDBG(AQM_KTRACE_STATS_FLOW_REPORT_CE, fq->fq_flowhash,
818*d4514f0bSApple OSS Distributions AQM_KTRACE_FQ_GRP_SC_IDX(fq), qdelay, pacing_delay);
819*d4514f0bSApple OSS Distributions /*
820*d4514f0bSApple OSS Distributions * The packet buffer that pktsched_mark_ecn writes to can be pageable.
821*d4514f0bSApple OSS Distributions * Since it is not safe to write to pageable memory while preemption
822*d4514f0bSApple OSS Distributions * is disabled, convert the spin lock into mutex.
823*d4514f0bSApple OSS Distributions */
824*d4514f0bSApple OSS Distributions IFCQ_CONVERT_LOCK(fqs->fqs_ifq);
825*d4514f0bSApple OSS Distributions if (__improbable(l4s_local_ce_report != 0) &&
826*d4514f0bSApple OSS Distributions (*pkt_flags & PKTF_FLOW_ADV) != 0 &&
827*d4514f0bSApple OSS Distributions fq_if_report_ce(fqs, pkt, 1, fq->fq_pkts_since_last_report)) {
828*d4514f0bSApple OSS Distributions fq->fq_pkts_since_last_report = 0;
829*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_ce_reported++;
830*d4514f0bSApple OSS Distributions } else if (pktsched_mark_ecn(pkt) == 0) {
831*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_ce_marked++;
832*d4514f0bSApple OSS Distributions } else {
833*d4514f0bSApple OSS Distributions fq_cl->fcl_stat.fcl_ce_mark_failures++;
834*d4514f0bSApple OSS Distributions }
835*d4514f0bSApple OSS Distributions }
836*d4514f0bSApple OSS Distributions }
837*d4514f0bSApple OSS Distributions
838*d4514f0bSApple OSS Distributions ASSERT(pacing_delay <= INT64_MAX);
839*d4514f0bSApple OSS Distributions qdelay = MAX(0, qdelay - (int64_t)pacing_delay);
840*d4514f0bSApple OSS Distributions if (fq->fq_min_qdelay == 0 ||
841*d4514f0bSApple OSS Distributions (u_int64_t)qdelay < fq->fq_min_qdelay) {
842*d4514f0bSApple OSS Distributions fq->fq_min_qdelay = qdelay;
843*d4514f0bSApple OSS Distributions }
844*d4514f0bSApple OSS Distributions
845*d4514f0bSApple OSS Distributions if (now >= fq->fq_updatetime) {
846*d4514f0bSApple OSS Distributions if (fq->fq_min_qdelay > fq_min_delay_threshold) {
847*d4514f0bSApple OSS Distributions if (!FQ_IS_DELAY_HIGH(fq)) {
848*d4514f0bSApple OSS Distributions FQ_SET_DELAY_HIGH(fq);
849*d4514f0bSApple OSS Distributions }
850*d4514f0bSApple OSS Distributions } else {
851*d4514f0bSApple OSS Distributions FQ_CLEAR_DELAY_HIGH(fq);
852*d4514f0bSApple OSS Distributions }
853*d4514f0bSApple OSS Distributions /* Reset measured queue delay and update time */
854*d4514f0bSApple OSS Distributions fq->fq_updatetime = now + FQ_UPDATE_INTERVAL(fq);
855*d4514f0bSApple OSS Distributions fq->fq_min_qdelay = 0;
856*d4514f0bSApple OSS Distributions }
857*d4514f0bSApple OSS Distributions
858*d4514f0bSApple OSS Distributions if (fqs->fqs_large_flow != fq || !fq_if_almost_at_drop_limit(fqs)) {
859*d4514f0bSApple OSS Distributions FQ_CLEAR_OVERWHELMING(fq);
860*d4514f0bSApple OSS Distributions }
861*d4514f0bSApple OSS Distributions if (!FQ_IS_DELAY_HIGH(fq) || fq_empty(fq, fqs->fqs_ptype)) {
862*d4514f0bSApple OSS Distributions FQ_CLEAR_DELAY_HIGH(fq);
863*d4514f0bSApple OSS Distributions }
864*d4514f0bSApple OSS Distributions
865*d4514f0bSApple OSS Distributions if ((fq->fq_flags & FQF_FLOWCTL_ON) &&
866*d4514f0bSApple OSS Distributions !FQ_IS_DELAY_HIGH(fq) && !FQ_IS_OVERWHELMING(fq)) {
867*d4514f0bSApple OSS Distributions fq_if_flow_feedback(fqs, fq, fq_cl);
868*d4514f0bSApple OSS Distributions }
869*d4514f0bSApple OSS Distributions
870*d4514f0bSApple OSS Distributions if (fq_empty(fq, fqs->fqs_ptype)) {
871*d4514f0bSApple OSS Distributions /* Reset getqtime so that we don't count idle times */
872*d4514f0bSApple OSS Distributions fq->fq_getqtime = 0;
873*d4514f0bSApple OSS Distributions } else {
874*d4514f0bSApple OSS Distributions fq->fq_getqtime = now;
875*d4514f0bSApple OSS Distributions }
876*d4514f0bSApple OSS Distributions fq_if_is_flow_heavy(fqs, fq);
877*d4514f0bSApple OSS Distributions
878*d4514f0bSApple OSS Distributions *pkt_timestamp = 0;
879*d4514f0bSApple OSS Distributions switch (pkt->pktsched_ptype) {
880*d4514f0bSApple OSS Distributions case QP_MBUF:
881*d4514f0bSApple OSS Distributions *pkt_flags &= ~PKTF_PRIV_GUARDED;
882*d4514f0bSApple OSS Distributions break;
883*d4514f0bSApple OSS Distributions #if SKYWALK
884*d4514f0bSApple OSS Distributions case QP_PACKET:
885*d4514f0bSApple OSS Distributions /* sanity check */
886*d4514f0bSApple OSS Distributions ASSERT((*pkt_flags & ~PKT_F_COMMON_MASK) == 0);
887*d4514f0bSApple OSS Distributions break;
888*d4514f0bSApple OSS Distributions #endif /* SKYWALK */
889*d4514f0bSApple OSS Distributions default:
890*d4514f0bSApple OSS Distributions VERIFY(0);
891*d4514f0bSApple OSS Distributions /* NOTREACHED */
892*d4514f0bSApple OSS Distributions __builtin_unreachable();
893*d4514f0bSApple OSS Distributions }
894*d4514f0bSApple OSS Distributions }
895