1 // SPDX-License-Identifier: GPL-2.0+
2 /* Microchip Sparx5 Switch driver
3  *
4  * Copyright (c) 2021 Microchip Technology Inc. and its subsidiaries.
5  */
6 
7 #include "sparx5_main_regs.h"
8 #include "sparx5_main.h"
9 
10 #define XTR_EOF_0     ntohl((__force __be32)0x80000000u)
11 #define XTR_EOF_1     ntohl((__force __be32)0x80000001u)
12 #define XTR_EOF_2     ntohl((__force __be32)0x80000002u)
13 #define XTR_EOF_3     ntohl((__force __be32)0x80000003u)
14 #define XTR_PRUNED    ntohl((__force __be32)0x80000004u)
15 #define XTR_ABORT     ntohl((__force __be32)0x80000005u)
16 #define XTR_ESCAPE    ntohl((__force __be32)0x80000006u)
17 #define XTR_NOT_READY ntohl((__force __be32)0x80000007u)
18 
19 #define XTR_VALID_BYTES(x)      (4 - ((x) & 3))
20 
21 #define INJ_TIMEOUT_NS 50000
22 
sparx5_xtr_flush(struct sparx5 * sparx5,u8 grp)23 void sparx5_xtr_flush(struct sparx5 *sparx5, u8 grp)
24 {
25 	/* Start flush */
26 	spx5_wr(QS_XTR_FLUSH_FLUSH_SET(BIT(grp)), sparx5, QS_XTR_FLUSH);
27 
28 	/* Allow to drain */
29 	mdelay(1);
30 
31 	/* All Queues normal */
32 	spx5_wr(0, sparx5, QS_XTR_FLUSH);
33 }
34 
sparx5_ifh_parse(struct sparx5 * sparx5,u32 * ifh,struct frame_info * info)35 void sparx5_ifh_parse(struct sparx5 *sparx5, u32 *ifh, struct frame_info *info)
36 {
37 	u8 *xtr_hdr = (u8 *)ifh;
38 
39 	/* FWD is bit 45-72 (28 bits), but we only read the 27 LSB for now */
40 	u32 fwd =
41 		((u32)xtr_hdr[27] << 24) |
42 		((u32)xtr_hdr[28] << 16) |
43 		((u32)xtr_hdr[29] <<  8) |
44 		((u32)xtr_hdr[30] <<  0);
45 	fwd = (fwd >> 5);
46 	info->src_port = spx5_field_get(GENMASK(is_sparx5(sparx5) ? 7 : 6, 1),
47 					fwd);
48 
49 	/*
50 	 * Bit 270-271 are occasionally unexpectedly set by the hardware,
51 	 * clear bits before extracting timestamp
52 	 */
53 	info->timestamp =
54 		((u64)(xtr_hdr[2] & GENMASK(5, 0)) << 24) |
55 		((u64)xtr_hdr[3] << 16) |
56 		((u64)xtr_hdr[4] <<  8) |
57 		((u64)xtr_hdr[5] <<  0);
58 }
59 
sparx5_xtr_grp(struct sparx5 * sparx5,u8 grp,bool byte_swap)60 static void sparx5_xtr_grp(struct sparx5 *sparx5, u8 grp, bool byte_swap)
61 {
62 	bool eof_flag = false, pruned_flag = false, abort_flag = false;
63 	struct net_device *netdev;
64 	struct sparx5_port *port;
65 	struct frame_info fi;
66 	int i, byte_cnt = 0;
67 	struct sk_buff *skb;
68 	u32 ifh[IFH_LEN];
69 	u32 *rxbuf;
70 
71 	/* Get IFH */
72 	for (i = 0; i < IFH_LEN; i++)
73 		ifh[i] = spx5_rd(sparx5, QS_XTR_RD(grp));
74 
75 	/* Decode IFH (what's needed) */
76 	sparx5_ifh_parse(sparx5, ifh, &fi);
77 
78 	/* Map to port netdev */
79 	port = fi.src_port < sparx5->data->consts->n_ports ?
80 		sparx5->ports[fi.src_port] : NULL;
81 	if (!port || !port->ndev) {
82 		dev_err(sparx5->dev, "Data on inactive port %d\n", fi.src_port);
83 		sparx5_xtr_flush(sparx5, grp);
84 		return;
85 	}
86 
87 	/* Have netdev, get skb */
88 	netdev = port->ndev;
89 	skb = netdev_alloc_skb(netdev, netdev->mtu + ETH_HLEN);
90 	if (!skb) {
91 		sparx5_xtr_flush(sparx5, grp);
92 		dev_err(sparx5->dev, "No skb allocated\n");
93 		netdev->stats.rx_dropped++;
94 		return;
95 	}
96 	rxbuf = (u32 *)skb->data;
97 
98 	/* Now, pull frame data */
99 	while (!eof_flag) {
100 		u32 val = spx5_rd(sparx5, QS_XTR_RD(grp));
101 		u32 cmp = val;
102 
103 		if (byte_swap)
104 			cmp = ntohl((__force __be32)val);
105 
106 		switch (cmp) {
107 		case XTR_NOT_READY:
108 			break;
109 		case XTR_ABORT:
110 			/* No accompanying data */
111 			abort_flag = true;
112 			eof_flag = true;
113 			break;
114 		case XTR_EOF_0:
115 		case XTR_EOF_1:
116 		case XTR_EOF_2:
117 		case XTR_EOF_3:
118 			/* This assumes STATUS_WORD_POS == 1, Status
119 			 * just after last data
120 			 */
121 			if (!byte_swap)
122 				val = ntohl((__force __be32)val);
123 			byte_cnt -= (4 - XTR_VALID_BYTES(val));
124 			eof_flag = true;
125 			break;
126 		case XTR_PRUNED:
127 			/* But get the last 4 bytes as well */
128 			eof_flag = true;
129 			pruned_flag = true;
130 			fallthrough;
131 		case XTR_ESCAPE:
132 			*rxbuf = spx5_rd(sparx5, QS_XTR_RD(grp));
133 			byte_cnt += 4;
134 			rxbuf++;
135 			break;
136 		default:
137 			*rxbuf = val;
138 			byte_cnt += 4;
139 			rxbuf++;
140 		}
141 	}
142 
143 	if (abort_flag || pruned_flag || !eof_flag) {
144 		netdev_err(netdev, "Discarded frame: abort:%d pruned:%d eof:%d\n",
145 			   abort_flag, pruned_flag, eof_flag);
146 		kfree_skb(skb);
147 		netdev->stats.rx_dropped++;
148 		return;
149 	}
150 
151 	/* Everything we see on an interface that is in the HW bridge
152 	 * has already been forwarded
153 	 */
154 	if (test_bit(port->portno, sparx5->bridge_mask))
155 		skb->offload_fwd_mark = 1;
156 
157 	/* Finish up skb */
158 	skb_put(skb, byte_cnt - ETH_FCS_LEN);
159 	eth_skb_pad(skb);
160 	sparx5_ptp_rxtstamp(sparx5, skb, fi.timestamp);
161 	skb->protocol = eth_type_trans(skb, netdev);
162 	netdev->stats.rx_bytes += skb->len;
163 	netdev->stats.rx_packets++;
164 	netif_rx(skb);
165 }
166 
sparx5_inject(struct sparx5 * sparx5,u32 * ifh,struct sk_buff * skb,struct net_device * ndev)167 static int sparx5_inject(struct sparx5 *sparx5,
168 			 u32 *ifh,
169 			 struct sk_buff *skb,
170 			 struct net_device *ndev)
171 {
172 	int grp = INJ_QUEUE;
173 	u32 val, w, count;
174 	u8 *buf;
175 
176 	val = spx5_rd(sparx5, QS_INJ_STATUS);
177 	if (!(QS_INJ_STATUS_FIFO_RDY_GET(val) & BIT(grp))) {
178 		pr_err_ratelimited("Injection: Queue not ready: 0x%lx\n",
179 				   QS_INJ_STATUS_FIFO_RDY_GET(val));
180 		return -EBUSY;
181 	}
182 
183 	/* Indicate SOF */
184 	spx5_wr(QS_INJ_CTRL_SOF_SET(1) |
185 		QS_INJ_CTRL_GAP_SIZE_SET(1),
186 		sparx5, QS_INJ_CTRL(grp));
187 
188 	/* Write the IFH to the chip. */
189 	for (w = 0; w < IFH_LEN; w++)
190 		spx5_wr(ifh[w], sparx5, QS_INJ_WR(grp));
191 
192 	/* Write words, round up */
193 	count = DIV_ROUND_UP(skb->len, 4);
194 	buf = skb->data;
195 	for (w = 0; w < count; w++, buf += 4) {
196 		val = get_unaligned((const u32 *)buf);
197 		spx5_wr(val, sparx5, QS_INJ_WR(grp));
198 	}
199 
200 	/* Add padding */
201 	while (w < (60 / 4)) {
202 		spx5_wr(0, sparx5, QS_INJ_WR(grp));
203 		w++;
204 	}
205 
206 	/* Indicate EOF and valid bytes in last word */
207 	spx5_wr(QS_INJ_CTRL_GAP_SIZE_SET(1) |
208 		QS_INJ_CTRL_VLD_BYTES_SET(skb->len < 60 ? 0 : skb->len % 4) |
209 		QS_INJ_CTRL_EOF_SET(1),
210 		sparx5, QS_INJ_CTRL(grp));
211 
212 	/* Add dummy CRC */
213 	spx5_wr(0, sparx5, QS_INJ_WR(grp));
214 	w++;
215 
216 	val = spx5_rd(sparx5, QS_INJ_STATUS);
217 	if (QS_INJ_STATUS_WMARK_REACHED_GET(val) & BIT(grp)) {
218 		struct sparx5_port *port = netdev_priv(ndev);
219 
220 		pr_err_ratelimited("Injection: Watermark reached: 0x%lx\n",
221 				   QS_INJ_STATUS_WMARK_REACHED_GET(val));
222 		netif_stop_queue(ndev);
223 		hrtimer_start(&port->inj_timer, INJ_TIMEOUT_NS,
224 			      HRTIMER_MODE_REL);
225 	}
226 
227 	return NETDEV_TX_OK;
228 }
229 
sparx5_port_xmit_impl(struct sk_buff * skb,struct net_device * dev)230 netdev_tx_t sparx5_port_xmit_impl(struct sk_buff *skb, struct net_device *dev)
231 {
232 	struct net_device_stats *stats = &dev->stats;
233 	struct sparx5_port *port = netdev_priv(dev);
234 	struct sparx5 *sparx5 = port->sparx5;
235 	const struct sparx5_ops *ops;
236 	u32 ifh[IFH_LEN];
237 	netdev_tx_t ret;
238 
239 	ops = sparx5->data->ops;
240 
241 	memset(ifh, 0, IFH_LEN * 4);
242 	sparx5_set_port_ifh(sparx5, ifh, port->portno);
243 
244 	if (sparx5->ptp && skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) {
245 		if (sparx5_ptp_txtstamp_request(port, skb) < 0)
246 			return NETDEV_TX_BUSY;
247 
248 		sparx5_set_port_ifh_rew_op(ifh, SPARX5_SKB_CB(skb)->rew_op);
249 		sparx5_set_port_ifh_pdu_type(sparx5, ifh,
250 					     SPARX5_SKB_CB(skb)->pdu_type);
251 		sparx5_set_port_ifh_pdu_w16_offset(sparx5, ifh,
252 						   SPARX5_SKB_CB(skb)->pdu_w16_offset);
253 		sparx5_set_port_ifh_timestamp(sparx5, ifh,
254 					      SPARX5_SKB_CB(skb)->ts_id);
255 	}
256 
257 	skb_tx_timestamp(skb);
258 	spin_lock(&sparx5->tx_lock);
259 	if (sparx5->fdma_irq > 0)
260 		ret = ops->fdma_xmit(sparx5, ifh, skb, dev);
261 	else
262 		ret = sparx5_inject(sparx5, ifh, skb, dev);
263 	spin_unlock(&sparx5->tx_lock);
264 
265 	if (ret == -EBUSY)
266 		goto busy;
267 	if (ret < 0)
268 		goto drop;
269 
270 	if (!is_sparx5(sparx5))
271 		/* When lan969x and TX_OK, stats and SKB consumption is handled
272 		 * in the TX completion loop, so dont go any further.
273 		 */
274 		return NETDEV_TX_OK;
275 
276 	stats->tx_bytes += skb->len;
277 	stats->tx_packets++;
278 	sparx5->tx.packets++;
279 
280 	if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP &&
281 	    SPARX5_SKB_CB(skb)->rew_op == IFH_REW_OP_TWO_STEP_PTP)
282 		return NETDEV_TX_OK;
283 
284 	dev_consume_skb_any(skb);
285 	return NETDEV_TX_OK;
286 drop:
287 	stats->tx_dropped++;
288 	sparx5->tx.dropped++;
289 	dev_kfree_skb_any(skb);
290 	return NETDEV_TX_OK;
291 busy:
292 	if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP &&
293 	    SPARX5_SKB_CB(skb)->rew_op == IFH_REW_OP_TWO_STEP_PTP)
294 		sparx5_ptp_txtstamp_release(port, skb);
295 	return NETDEV_TX_BUSY;
296 }
297 
sparx5_injection_timeout(struct hrtimer * tmr)298 static enum hrtimer_restart sparx5_injection_timeout(struct hrtimer *tmr)
299 {
300 	struct sparx5_port *port = container_of(tmr, struct sparx5_port,
301 						inj_timer);
302 	int grp = INJ_QUEUE;
303 	u32 val;
304 
305 	val = spx5_rd(port->sparx5, QS_INJ_STATUS);
306 	if (QS_INJ_STATUS_WMARK_REACHED_GET(val) & BIT(grp)) {
307 		pr_err_ratelimited("Injection: Reset watermark count\n");
308 		/* Reset Watermark count to restart */
309 		spx5_rmw(DSM_DEV_TX_STOP_WM_CFG_DEV_TX_CNT_CLR_SET(1),
310 			 DSM_DEV_TX_STOP_WM_CFG_DEV_TX_CNT_CLR,
311 			 port->sparx5,
312 			 DSM_DEV_TX_STOP_WM_CFG(port->portno));
313 	}
314 	netif_wake_queue(port->ndev);
315 	return HRTIMER_NORESTART;
316 }
317 
sparx5_manual_injection_mode(struct sparx5 * sparx5)318 int sparx5_manual_injection_mode(struct sparx5 *sparx5)
319 {
320 	const int byte_swap = 1;
321 	int portno;
322 
323 	/* Change mode to manual extraction and injection */
324 	spx5_wr(QS_XTR_GRP_CFG_MODE_SET(1) |
325 		QS_XTR_GRP_CFG_STATUS_WORD_POS_SET(1) |
326 		QS_XTR_GRP_CFG_BYTE_SWAP_SET(byte_swap),
327 		sparx5, QS_XTR_GRP_CFG(XTR_QUEUE));
328 	spx5_wr(QS_INJ_GRP_CFG_MODE_SET(1) |
329 		QS_INJ_GRP_CFG_BYTE_SWAP_SET(byte_swap),
330 		sparx5, QS_INJ_GRP_CFG(INJ_QUEUE));
331 
332 	/* CPU ports capture setup */
333 	for (portno = sparx5_get_internal_port(sparx5, SPX5_PORT_CPU_0);
334 	     portno <= sparx5_get_internal_port(sparx5, SPX5_PORT_CPU_1);
335 	     portno++) {
336 		/* ASM CPU port: No preamble, IFH, enable padding */
337 		spx5_wr(ASM_PORT_CFG_PAD_ENA_SET(1) |
338 			ASM_PORT_CFG_NO_PREAMBLE_ENA_SET(1) |
339 			ASM_PORT_CFG_INJ_FORMAT_CFG_SET(1), /* 1 = IFH */
340 			sparx5, ASM_PORT_CFG(portno));
341 
342 		/* Reset WM cnt to unclog queued frames */
343 		spx5_rmw(DSM_DEV_TX_STOP_WM_CFG_DEV_TX_CNT_CLR_SET(1),
344 			 DSM_DEV_TX_STOP_WM_CFG_DEV_TX_CNT_CLR,
345 			 sparx5,
346 			 DSM_DEV_TX_STOP_WM_CFG(portno));
347 
348 		/* Set Disassembler Stop Watermark level */
349 		spx5_rmw(DSM_DEV_TX_STOP_WM_CFG_DEV_TX_STOP_WM_SET(0),
350 			 DSM_DEV_TX_STOP_WM_CFG_DEV_TX_STOP_WM,
351 			 sparx5,
352 			 DSM_DEV_TX_STOP_WM_CFG(portno));
353 
354 		/* Enable Disassembler buffer underrun watchdog
355 		 */
356 		spx5_rmw(DSM_BUF_CFG_UNDERFLOW_WATCHDOG_DIS_SET(0),
357 			 DSM_BUF_CFG_UNDERFLOW_WATCHDOG_DIS,
358 			 sparx5,
359 			 DSM_BUF_CFG(portno));
360 	}
361 	return 0;
362 }
363 
sparx5_xtr_handler(int irq,void * _sparx5)364 irqreturn_t sparx5_xtr_handler(int irq, void *_sparx5)
365 {
366 	struct sparx5 *s5 = _sparx5;
367 	int poll = 64;
368 
369 	/* Check data in queue */
370 	while (spx5_rd(s5, QS_XTR_DATA_PRESENT) & BIT(XTR_QUEUE) && poll-- > 0)
371 		sparx5_xtr_grp(s5, XTR_QUEUE, false);
372 
373 	return IRQ_HANDLED;
374 }
375 
sparx5_port_inj_timer_setup(struct sparx5_port * port)376 void sparx5_port_inj_timer_setup(struct sparx5_port *port)
377 {
378 	hrtimer_init(&port->inj_timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
379 	port->inj_timer.function = sparx5_injection_timeout;
380 }
381