txrx.c 60.7 KB
Newer Older
1
/*
2
 * Copyright (c) 2012-2017 Qualcomm Atheros, Inc.
3
 * Copyright (c) 2018, The Linux Foundation. All rights reserved.
4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21
 *
 * Permission to use, copy, modify, and/or distribute this software for any
 * purpose with or without fee is hereby granted, provided that the above
 * copyright notice and this permission notice appear in all copies.
 *
 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
 */

#include <linux/etherdevice.h>
#include <net/ieee80211_radiotap.h>
#include <linux/if_arp.h>
#include <linux/moduleparam.h>
22 23 24
#include <linux/ip.h>
#include <linux/ipv6.h>
#include <net/ipv6.h>
25
#include <linux/prefetch.h>
26 27 28 29

#include "wil6210.h"
#include "wmi.h"
#include "txrx.h"
V
Vladimir Kondratiev 已提交
30
#include "trace.h"
31
#include "txrx_edma.h"
32 33

static bool rtap_include_phy_info;
34
module_param(rtap_include_phy_info, bool, 0444);
35 36 37
MODULE_PARM_DESC(rtap_include_phy_info,
		 " Include PHY info in the radiotap header, default - no");

38
bool rx_align_2;
39
module_param(rx_align_2, bool, 0444);
40 41
MODULE_PARM_DESC(rx_align_2, " align Rx buffers on 4*n+2, default - no");

L
Lior David 已提交
42 43 44 45
bool rx_large_buf;
module_param(rx_large_buf, bool, 0444);
MODULE_PARM_DESC(rx_large_buf, " allocate 8KB RX buffers, default - no");

46 47 48 49 50
static inline uint wil_rx_snaplen(void)
{
	return rx_align_2 ? 6 : 0;
}

51 52
/* wil_ring_wmark_low - low watermark for available descriptor space */
static inline int wil_ring_wmark_low(struct wil_ring *ring)
53
{
54
	return ring->size / 8;
55 56
}

57 58
/* wil_ring_wmark_high - high watermark for available descriptor space */
static inline int wil_ring_wmark_high(struct wil_ring *ring)
59
{
60
	return ring->size / 4;
61 62
}

D
Dedy Lansky 已提交
63
/* returns true if num avail descriptors is lower than wmark_low */
64
static inline int wil_ring_avail_low(struct wil_ring *ring)
D
Dedy Lansky 已提交
65
{
66
	return wil_ring_avail_tx(ring) < wil_ring_wmark_low(ring);
D
Dedy Lansky 已提交
67 68 69
}

/* returns true if num avail descriptors is higher than wmark_high */
70
static inline int wil_ring_avail_high(struct wil_ring *ring)
D
Dedy Lansky 已提交
71
{
72
	return wil_ring_avail_tx(ring) > wil_ring_wmark_high(ring);
D
Dedy Lansky 已提交
73 74
}

75 76 77 78 79 80 81
/* returns true when all tx vrings are empty */
bool wil_is_tx_idle(struct wil6210_priv *wil)
{
	int i;
	unsigned long data_comp_to;

	for (i = 0; i < WIL6210_MAX_TX_RINGS; i++) {
82 83 84 85
		struct wil_ring *vring = &wil->ring_tx[i];
		int vring_index = vring - wil->ring_tx;
		struct wil_ring_tx_data *txdata =
			&wil->ring_tx_data[vring_index];
86 87 88 89 90 91 92 93 94 95 96

		spin_lock(&txdata->lock);

		if (!vring->va || !txdata->enabled) {
			spin_unlock(&txdata->lock);
			continue;
		}

		data_comp_to = jiffies + msecs_to_jiffies(
					WIL_DATA_COMPLETION_TO_MS);
		if (test_bit(wil_status_napi_en, wil->status)) {
97
			while (!wil_ring_is_empty(vring)) {
98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120
				if (time_after(jiffies, data_comp_to)) {
					wil_dbg_pm(wil,
						   "TO waiting for idle tx\n");
					spin_unlock(&txdata->lock);
					return false;
				}
				wil_dbg_ratelimited(wil,
						    "tx vring is not empty -> NAPI\n");
				spin_unlock(&txdata->lock);
				napi_synchronize(&wil->napi_tx);
				msleep(20);
				spin_lock(&txdata->lock);
				if (!vring->va || !txdata->enabled)
					break;
			}
		}

		spin_unlock(&txdata->lock);
	}

	return true;
}

121
static int wil_vring_alloc(struct wil6210_priv *wil, struct wil_ring *vring)
122 123 124 125 126
{
	struct device *dev = wil_to_dev(wil);
	size_t sz = vring->size * sizeof(vring->va[0]);
	uint i;

127
	wil_dbg_misc(wil, "vring_alloc:\n");
128

129 130 131 132
	BUILD_BUG_ON(sizeof(vring->va[0]) != 32);

	vring->swhead = 0;
	vring->swtail = 0;
133
	vring->ctx = kcalloc(vring->size, sizeof(vring->ctx[0]), GFP_KERNEL);
134 135 136 137
	if (!vring->ctx) {
		vring->va = NULL;
		return -ENOMEM;
	}
138

139
	/* vring->va should be aligned on its size rounded up to power of 2
140 141 142 143
	 * This is granted by the dma_alloc_coherent.
	 *
	 * HW has limitation that all vrings addresses must share the same
	 * upper 16 msb bits part of 48 bits address. To workaround that,
144 145
	 * if we are using more than 32 bit addresses switch to 32 bit
	 * allocation before allocating vring memory.
146 147 148 149
	 *
	 * There's no check for the return value of dma_set_mask_and_coherent,
	 * since we assume if we were able to set the mask during
	 * initialization in this system it will not fail if we set it again
150
	 */
151
	if (wil->dma_addr_size > 32)
152 153
		dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32));

154 155 156 157 158 159
	vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL);
	if (!vring->va) {
		kfree(vring->ctx);
		vring->ctx = NULL;
		return -ENOMEM;
	}
160

161 162 163
	if (wil->dma_addr_size > 32)
		dma_set_mask_and_coherent(dev,
					  DMA_BIT_MASK(wil->dma_addr_size));
164

165 166 167 168 169
	/* initially, all descriptors are SW owned
	 * For Tx and Rx, ownership bit is at the same location, thus
	 * we can use any
	 */
	for (i = 0; i < vring->size; i++) {
170 171
		volatile struct vring_tx_desc *_d =
			&vring->va[i].tx.legacy;
172

173
		_d->dma.status = TX_DMA_STATUS_DU;
174 175
	}

176 177
	wil_dbg_misc(wil, "vring[%d] 0x%p:%pad 0x%p\n", vring->size,
		     vring->va, &vring->pa, vring->ctx);
178 179 180 181

	return 0;
}

182
static void wil_txdesc_unmap(struct device *dev, union wil_tx_desc *desc,
183 184
			     struct wil_ctx *ctx)
{
185
	struct vring_tx_desc *d = &desc->legacy;
186 187
	dma_addr_t pa = wil_desc_addr(&d->dma.addr);
	u16 dmalen = le16_to_cpu(d->dma.length);
188

189 190 191 192 193 194 195 196 197 198 199 200
	switch (ctx->mapped_as) {
	case wil_mapped_as_single:
		dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE);
		break;
	case wil_mapped_as_page:
		dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE);
		break;
	default:
		break;
	}
}

201
static void wil_vring_free(struct wil6210_priv *wil, struct wil_ring *vring)
202 203 204 205
{
	struct device *dev = wil_to_dev(wil);
	size_t sz = vring->size * sizeof(vring->va[0]);

206
	lockdep_assert_held(&wil->mutex);
207
	if (!vring->is_rx) {
208
		int vring_index = vring - wil->ring_tx;
209 210 211 212 213 214 215 216 217 218

		wil_dbg_misc(wil, "free Tx vring %d [%d] 0x%p:%pad 0x%p\n",
			     vring_index, vring->size, vring->va,
			     &vring->pa, vring->ctx);
	} else {
		wil_dbg_misc(wil, "free Rx vring [%d] 0x%p:%pad 0x%p\n",
			     vring->size, vring->va,
			     &vring->pa, vring->ctx);
	}

219
	while (!wil_ring_is_empty(vring)) {
220
		dma_addr_t pa;
221
		u16 dmalen;
222
		struct wil_ctx *ctx;
223

224
		if (!vring->is_rx) {
225 226
			struct vring_tx_desc dd, *d = &dd;
			volatile struct vring_tx_desc *_d =
227
					&vring->va[vring->swtail].tx.legacy;
228

229
			ctx = &vring->ctx[vring->swtail];
230 231 232 233
			if (!ctx) {
				wil_dbg_txrx(wil,
					     "ctx(%d) was already completed\n",
					     vring->swtail);
234
				vring->swtail = wil_ring_next_tail(vring);
235 236
				continue;
			}
237
			*d = *_d;
238
			wil_txdesc_unmap(dev, (union wil_tx_desc *)d, ctx);
239 240
			if (ctx->skb)
				dev_kfree_skb_any(ctx->skb);
241
			vring->swtail = wil_ring_next_tail(vring);
242
		} else { /* rx */
243 244
			struct vring_rx_desc dd, *d = &dd;
			volatile struct vring_rx_desc *_d =
245
				&vring->va[vring->swhead].rx.legacy;
246

247
			ctx = &vring->ctx[vring->swhead];
248 249
			*d = *_d;
			pa = wil_desc_addr(&d->dma.addr);
250 251
			dmalen = le16_to_cpu(d->dma.length);
			dma_unmap_single(dev, pa, dmalen, DMA_FROM_DEVICE);
252
			kfree_skb(ctx->skb);
253
			wil_ring_advance_head(vring, 1);
254 255 256 257 258 259 260 261 262 263 264 265 266 267
		}
	}
	dma_free_coherent(dev, sz, (void *)vring->va, vring->pa);
	kfree(vring->ctx);
	vring->pa = 0;
	vring->va = NULL;
	vring->ctx = NULL;
}

/**
 * Allocate one skb for Rx VRING
 *
 * Safe to call from IRQ
 */
268
static int wil_vring_alloc_skb(struct wil6210_priv *wil, struct wil_ring *vring,
269 270 271
			       u32 i, int headroom)
{
	struct device *dev = wil_to_dev(wil);
L
Lior David 已提交
272
	unsigned int sz = wil->rx_buf_len + ETH_HLEN + wil_rx_snaplen();
273
	struct vring_rx_desc dd, *d = &dd;
274
	volatile struct vring_rx_desc *_d = &vring->va[i].rx.legacy;
275 276
	dma_addr_t pa;
	struct sk_buff *skb = dev_alloc_skb(sz + headroom);
277

278 279 280 281 282 283 284 285 286 287 288 289
	if (unlikely(!skb))
		return -ENOMEM;

	skb_reserve(skb, headroom);
	skb_put(skb, sz);

	pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE);
	if (unlikely(dma_mapping_error(dev, pa))) {
		kfree_skb(skb);
		return -ENOMEM;
	}

290
	d->dma.d0 = RX_DMA_D0_CMD_DMA_RT | RX_DMA_D0_CMD_DMA_IT;
291
	wil_desc_addr_set(&d->dma.addr, pa);
292 293 294 295
	/* ip_length don't care */
	/* b11 don't care */
	/* error don't care */
	d->dma.status = 0; /* BIT(0) should be 0 for HW_OWNED */
296
	d->dma.length = cpu_to_le16(sz);
297
	*_d = *d;
298
	vring->ctx[i].skb = skb;
299 300 301 302 303 304 305 306 307 308 309 310 311 312

	return 0;
}

/**
 * Adds radiotap header
 *
 * Any error indicated as "Bad FCS"
 *
 * Vendor data for 04:ce:14-1 (Wilocity-1) consists of:
 *  - Rx descriptor: 32 bytes
 *  - Phy info
 */
static void wil_rx_add_radiotap_header(struct wil6210_priv *wil,
313
				       struct sk_buff *skb)
314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335
{
	struct wil6210_rtap {
		struct ieee80211_radiotap_header rthdr;
		/* fields should be in the order of bits in rthdr.it_present */
		/* flags */
		u8 flags;
		/* channel */
		__le16 chnl_freq __aligned(2);
		__le16 chnl_flags;
		/* MCS */
		u8 mcs_present;
		u8 mcs_flags;
		u8 mcs_index;
	} __packed;
	struct wil6210_rtap_vendor {
		struct wil6210_rtap rtap;
		/* vendor */
		u8 vendor_oui[3] __aligned(2);
		u8 vendor_ns;
		__le16 vendor_skip;
		u8 vendor_data[0];
	} __packed;
336
	struct vring_rx_desc *d = wil_skb_rxdesc(skb);
337 338 339 340
	struct wil6210_rtap_vendor *rtap_vendor;
	int rtap_len = sizeof(struct wil6210_rtap);
	int phy_length = 0; /* phy info header size, bytes */
	static char phy_data[128];
341
	struct ieee80211_channel *ch = wil->monitor_chandef.chan;
342 343 344 345 346 347 348 349 350 351 352 353 354

	if (rtap_include_phy_info) {
		rtap_len = sizeof(*rtap_vendor) + sizeof(*d);
		/* calculate additional length */
		if (d->dma.status & RX_DMA_STATUS_PHY_INFO) {
			/**
			 * PHY info starts from 8-byte boundary
			 * there are 8-byte lines, last line may be partially
			 * written (HW bug), thus FW configures for last line
			 * to be excessive. Driver skips this last line.
			 */
			int len = min_t(int, 8 + sizeof(phy_data),
					wil_rxdesc_phy_length(d));
355

356 357 358
			if (len > 8) {
				void *p = skb_tail_pointer(skb);
				void *pa = PTR_ALIGN(p, 8);
359

360 361 362 363 364 365 366 367 368 369 370
				if (skb_tailroom(skb) >= len + (pa - p)) {
					phy_length = len - 8;
					memcpy(phy_data, pa, phy_length);
				}
			}
		}
		rtap_len += phy_length;
	}

	if (skb_headroom(skb) < rtap_len &&
	    pskb_expand_head(skb, rtap_len, 0, GFP_ATOMIC)) {
371
		wil_err(wil, "Unable to expand headroom to %d\n", rtap_len);
372 373 374
		return;
	}

375
	rtap_vendor = skb_push(skb, rtap_len);
376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410
	memset(rtap_vendor, 0, rtap_len);

	rtap_vendor->rtap.rthdr.it_version = PKTHDR_RADIOTAP_VERSION;
	rtap_vendor->rtap.rthdr.it_len = cpu_to_le16(rtap_len);
	rtap_vendor->rtap.rthdr.it_present = cpu_to_le32(
			(1 << IEEE80211_RADIOTAP_FLAGS) |
			(1 << IEEE80211_RADIOTAP_CHANNEL) |
			(1 << IEEE80211_RADIOTAP_MCS));
	if (d->dma.status & RX_DMA_STATUS_ERROR)
		rtap_vendor->rtap.flags |= IEEE80211_RADIOTAP_F_BADFCS;

	rtap_vendor->rtap.chnl_freq = cpu_to_le16(ch ? ch->center_freq : 58320);
	rtap_vendor->rtap.chnl_flags = cpu_to_le16(0);

	rtap_vendor->rtap.mcs_present = IEEE80211_RADIOTAP_MCS_HAVE_MCS;
	rtap_vendor->rtap.mcs_flags = 0;
	rtap_vendor->rtap.mcs_index = wil_rxdesc_mcs(d);

	if (rtap_include_phy_info) {
		rtap_vendor->rtap.rthdr.it_present |= cpu_to_le32(1 <<
				IEEE80211_RADIOTAP_VENDOR_NAMESPACE);
		/* OUI for Wilocity 04:ce:14 */
		rtap_vendor->vendor_oui[0] = 0x04;
		rtap_vendor->vendor_oui[1] = 0xce;
		rtap_vendor->vendor_oui[2] = 0x14;
		rtap_vendor->vendor_ns = 1;
		/* Rx descriptor + PHY data  */
		rtap_vendor->vendor_skip = cpu_to_le16(sizeof(*d) +
						       phy_length);
		memcpy(rtap_vendor->vendor_data, (void *)d, sizeof(*d));
		memcpy(rtap_vendor->vendor_data + sizeof(*d), phy_data,
		       phy_length);
	}
}

411
static bool wil_is_rx_idle(struct wil6210_priv *wil)
412 413
{
	struct vring_rx_desc *_d;
414
	struct wil_ring *ring = &wil->ring_rx;
415

416
	_d = (struct vring_rx_desc *)&ring->va[ring->swhead].rx.legacy;
417 418 419 420 421 422
	if (_d->dma.status & RX_DMA_STATUS_DU)
		return false;

	return true;
}

423 424 425
/**
 * reap 1 frame from @swhead
 *
426 427
 * Rx descriptor copied to skb->cb
 *
428 429 430
 * Safe to call from IRQ
 */
static struct sk_buff *wil_vring_reap_rx(struct wil6210_priv *wil,
431
					 struct wil_ring *vring)
432 433
{
	struct device *dev = wil_to_dev(wil);
434 435
	struct wil6210_vif *vif;
	struct net_device *ndev;
436 437
	volatile struct vring_rx_desc *_d;
	struct vring_rx_desc *d;
438 439
	struct sk_buff *skb;
	dma_addr_t pa;
440
	unsigned int snaplen = wil_rx_snaplen();
L
Lior David 已提交
441
	unsigned int sz = wil->rx_buf_len + ETH_HLEN + snaplen;
442
	u16 dmalen;
443
	u8 ftype;
444
	int cid, mid;
445
	int i;
446 447
	struct wil_net_stats *stats;

448 449
	BUILD_BUG_ON(sizeof(struct vring_rx_desc) > sizeof(skb->cb));

450
again:
451
	if (unlikely(wil_ring_is_empty(vring)))
452 453
		return NULL;

454
	i = (int)vring->swhead;
455
	_d = &vring->va[i].rx.legacy;
456
	if (unlikely(!(_d->dma.status & RX_DMA_STATUS_DU))) {
457 458 459 460
		/* it is not error, we just reached end of Rx done area */
		return NULL;
	}

461 462
	skb = vring->ctx[i].skb;
	vring->ctx[i].skb = NULL;
463
	wil_ring_advance_head(vring, 1);
464 465
	if (!skb) {
		wil_err(wil, "No Rx skb at [%d]\n", i);
466
		goto again;
467
	}
468 469 470 471
	d = wil_skb_rxdesc(skb);
	*d = *_d;
	pa = wil_desc_addr(&d->dma.addr);

472
	dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE);
473 474
	dmalen = le16_to_cpu(d->dma.length);

475 476
	trace_wil6210_rx(i, d);
	wil_dbg_txrx(wil, "Rx[%3d] : %d bytes\n", i, dmalen);
477
	wil_hex_dump_txrx("RxD ", DUMP_PREFIX_NONE, 32, 4,
478
			  (const void *)d, sizeof(*d), false);
479

480
	cid = wil_rxdesc_cid(d);
481 482 483 484 485 486 487 488 489 490
	mid = wil_rxdesc_mid(d);
	vif = wil->vifs[mid];

	if (unlikely(!vif)) {
		wil_dbg_txrx(wil, "skipped RX descriptor with invalid mid %d",
			     mid);
		kfree_skb(skb);
		goto again;
	}
	ndev = vif_to_ndev(vif);
491 492
	stats = &wil->sta[cid].stats;

493
	if (unlikely(dmalen > sz)) {
494
		wil_err(wil, "Rx size too large: %d bytes!\n", dmalen);
495
		stats->rx_large_frame++;
496
		kfree_skb(skb);
497
		goto again;
498
	}
499
	skb_trim(skb, dmalen);
500

501 502
	prefetch(skb->data);

503 504 505
	wil_hex_dump_txrx("Rx ", DUMP_PREFIX_OFFSET, 16, 1,
			  skb->data, skb_headlen(skb), false);

506
	stats->last_mcs_rx = wil_rxdesc_mcs(d);
507 508
	if (stats->last_mcs_rx < ARRAY_SIZE(stats->rx_per_mcs))
		stats->rx_per_mcs[stats->last_mcs_rx]++;
509 510 511

	/* use radiotap header only if required */
	if (ndev->type == ARPHRD_IEEE80211_RADIOTAP)
512
		wil_rx_add_radiotap_header(wil, skb);
513 514 515 516

	/* no extra checks if in sniffer mode */
	if (ndev->type != ARPHRD_ETHER)
		return skb;
517
	/* Non-data frames may be delivered through Rx DMA channel (ex: BAR)
518 519 520
	 * Driver should recognize it by frame type, that is found
	 * in Rx descriptor. If type is not data, it is 802.11 frame as is
	 */
521
	ftype = wil_rxdesc_ftype(d) << 2;
522
	if (unlikely(ftype != IEEE80211_FTYPE_DATA)) {
523 524 525 526 527 528 529
		u8 fc1 = wil_rxdesc_fc1(d);
		int tid = wil_rxdesc_tid(d);
		u16 seq = wil_rxdesc_seq(d);

		wil_dbg_txrx(wil,
			     "Non-data frame FC[7:0] 0x%02x MID %d CID %d TID %d Seq 0x%03x\n",
			     fc1, mid, cid, tid, seq);
530
		stats->rx_non_data_frame++;
531 532 533 534
		if (wil_is_back_req(fc1)) {
			wil_dbg_txrx(wil,
				     "BAR: MID %d CID %d TID %d Seq 0x%03x\n",
				     mid, cid, tid, seq);
535
			wil_rx_bar(wil, vif, cid, tid, seq);
536 537 538 539 540 541 542 543 544 545 546 547
		} else {
			/* print again all info. One can enable only this
			 * without overhead for printing every Rx frame
			 */
			wil_dbg_txrx(wil,
				     "Unhandled non-data frame FC[7:0] 0x%02x MID %d CID %d TID %d Seq 0x%03x\n",
				     fc1, mid, cid, tid, seq);
			wil_hex_dump_txrx("RxD ", DUMP_PREFIX_NONE, 32, 4,
					  (const void *)d, sizeof(*d), false);
			wil_hex_dump_txrx("Rx ", DUMP_PREFIX_OFFSET, 16, 1,
					  skb->data, skb_headlen(skb), false);
		}
548
		kfree_skb(skb);
549
		goto again;
550 551
	}

552
	if (unlikely(skb->len < ETH_HLEN + snaplen)) {
553
		wil_err(wil, "Short frame, len = %d\n", skb->len);
554
		stats->rx_short_frame++;
555
		kfree_skb(skb);
556
		goto again;
557 558
	}

559 560 561 562
	/* L4 IDENT is on when HW calculated checksum, check status
	 * and in case of error drop the packet
	 * higher stack layers will handle retransmission (if required)
	 */
563
	if (likely(d->dma.status & RX_DMA_STATUS_L4I)) {
564
		/* L4 protocol identified, csum calculated */
565
		if (likely((d->dma.error & RX_DMA_ERROR_L4_ERR) == 0))
566
			skb->ip_summed = CHECKSUM_UNNECESSARY;
567 568 569 570 571
		/* If HW reports bad checksum, let IP stack re-check it
		 * For example, HW don't understand Microsoft IP stack that
		 * mis-calculates TCP checksum - if it should be 0x0,
		 * it writes 0xffff in violation of RFC 1624
		 */
572 573
	}

574 575 576 577 578 579 580 581 582 583 584
	if (snaplen) {
		/* Packet layout
		 * +-------+-------+---------+------------+------+
		 * | SA(6) | DA(6) | SNAP(6) | ETHTYPE(2) | DATA |
		 * +-------+-------+---------+------------+------+
		 * Need to remove SNAP, shifting SA and DA forward
		 */
		memmove(skb->data + snaplen, skb->data, 2 * ETH_ALEN);
		skb_pull(skb, snaplen);
	}

585 586 587 588 589 590
	return skb;
}

/**
 * allocate and fill up to @count buffers in rx ring
 * buffers posted at @swtail
591 592 593 594 595
 * Note: we have a single RX queue for servicing all VIFs, but we
 * allocate skbs with headroom according to main interface only. This
 * means it will not work with monitor interface together with other VIFs.
 * Currently we only support monitor interface on its own without other VIFs,
 * and we will need to fix this code once we add support.
596 597 598
 */
static int wil_rx_refill(struct wil6210_priv *wil, int count)
{
599
	struct net_device *ndev = wil->main_ndev;
600
	struct wil_ring *v = &wil->ring_rx;
601 602 603 604 605
	u32 next_tail;
	int rc = 0;
	int headroom = ndev->type == ARPHRD_IEEE80211_RADIOTAP ?
			WIL6210_RTAP_SIZE : 0;

606 607 608
	for (; next_tail = wil_ring_next_tail(v),
	     (next_tail != v->swhead) && (count-- > 0);
	     v->swtail = next_tail) {
609
		rc = wil_vring_alloc_skb(wil, v, v->swtail, headroom);
610
		if (unlikely(rc)) {
611 612
			wil_err_ratelimited(wil, "Error %d in rx refill[%d]\n",
					    rc, v->swtail);
613 614 615
			break;
		}
	}
616 617 618 619 620 621

	/* make sure all writes to descriptors (shared memory) are done before
	 * committing them to HW
	 */
	wmb();

622
	wil_w(wil, v->hwtail, v->swtail);
623 624 625 626

	return rc;
}

627 628 629 630 631 632 633 634 635
/**
 * reverse_memcmp - Compare two areas of memory, in reverse order
 * @cs: One area of memory
 * @ct: Another area of memory
 * @count: The size of the area.
 *
 * Cut'n'paste from original memcmp (see lib/string.c)
 * with minimal modifications
 */
636
int reverse_memcmp(const void *cs, const void *ct, size_t count)
637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680
{
	const unsigned char *su1, *su2;
	int res = 0;

	for (su1 = cs + count - 1, su2 = ct + count - 1; count > 0;
	     --su1, --su2, count--) {
		res = *su1 - *su2;
		if (res)
			break;
	}
	return res;
}

static int wil_rx_crypto_check(struct wil6210_priv *wil, struct sk_buff *skb)
{
	struct vring_rx_desc *d = wil_skb_rxdesc(skb);
	int cid = wil_rxdesc_cid(d);
	int tid = wil_rxdesc_tid(d);
	int key_id = wil_rxdesc_key_id(d);
	int mc = wil_rxdesc_mcast(d);
	struct wil_sta_info *s = &wil->sta[cid];
	struct wil_tid_crypto_rx *c = mc ? &s->group_crypto_rx :
				      &s->tid_crypto_rx[tid];
	struct wil_tid_crypto_rx_single *cc = &c->key_id[key_id];
	const u8 *pn = (u8 *)&d->mac.pn_15_0;

	if (!cc->key_set) {
		wil_err_ratelimited(wil,
				    "Key missing. CID %d TID %d MCast %d KEY_ID %d\n",
				    cid, tid, mc, key_id);
		return -EINVAL;
	}

	if (reverse_memcmp(pn, cc->pn, IEEE80211_GCMP_PN_LEN) <= 0) {
		wil_err_ratelimited(wil,
				    "Replay attack. CID %d TID %d MCast %d KEY_ID %d PN %6phN last %6phN\n",
				    cid, tid, mc, key_id, pn, cc->pn);
		return -EINVAL;
	}
	memcpy(cc->pn, pn, IEEE80211_GCMP_PN_LEN);

	return 0;
}

681 682 683 684 685 686 687 688 689 690 691 692 693 694 695
static int wil_rx_error_check(struct wil6210_priv *wil, struct sk_buff *skb,
			      struct wil_net_stats *stats)
{
	struct vring_rx_desc *d = wil_skb_rxdesc(skb);

	if ((d->dma.status & RX_DMA_STATUS_ERROR) &&
	    (d->dma.error & RX_DMA_ERROR_MIC)) {
		stats->rx_mic_error++;
		wil_dbg_txrx(wil, "MIC error, dropping packet\n");
		return -EFAULT;
	}

	return 0;
}

696 697 698 699 700 701 702 703 704
static void wil_get_netif_rx_params(struct sk_buff *skb, int *cid,
				    int *security)
{
	struct vring_rx_desc *d = wil_skb_rxdesc(skb);

	*cid = wil_rxdesc_cid(d); /* always 0..7, no need to check */
	*security = wil_rxdesc_security(d);
}

705 706
/*
 * Pass Rx packet to the netif. Update statistics.
V
Vladimir Kondratiev 已提交
707
 * Called in softirq context (NAPI poll).
708
 */
V
Vladimir Kondratiev 已提交
709
void wil_netif_rx_any(struct sk_buff *skb, struct net_device *ndev)
710
{
711
	gro_result_t rc = GRO_NORMAL;
712
	struct wil6210_vif *vif = ndev_to_vif(ndev);
713
	struct wil6210_priv *wil = ndev_to_wil(ndev);
714
	struct wireless_dev *wdev = vif_to_wdev(vif);
715
	unsigned int len = skb->len;
716 717
	int cid;
	int security;
718 719 720 721 722
	struct ethhdr *eth = (void *)skb->data;
	/* here looking for DA, not A1, thus Rxdesc's 'mcast' indication
	 * is not suitable, need to look at data
	 */
	int mcast = is_multicast_ether_addr(eth->h_dest);
723
	struct wil_net_stats *stats;
724 725 726 727 728 729 730 731
	struct sk_buff *xmit_skb = NULL;
	static const char * const gro_res_str[] = {
		[GRO_MERGED]		= "GRO_MERGED",
		[GRO_MERGED_FREE]	= "GRO_MERGED_FREE",
		[GRO_HELD]		= "GRO_HELD",
		[GRO_NORMAL]		= "GRO_NORMAL",
		[GRO_DROP]		= "GRO_DROP",
	};
732

733 734 735 736
	wil->txrx_ops.get_netif_rx_params(skb, &cid, &security);

	stats = &wil->sta[cid].stats;

737 738 739 740 741 742 743 744
	if (ndev->features & NETIF_F_RXHASH)
		/* fake L4 to ensure it won't be re-calculated later
		 * set hash to any non-zero value to activate rps
		 * mechanism, core will be chosen according
		 * to user-level rps configuration.
		 */
		skb_set_hash(skb, 1, PKT_HASH_TYPE_L4);

745 746
	skb_orphan(skb);

747
	if (security && (wil->txrx_ops.rx_crypto_check(wil, skb) != 0)) {
748 749 750 751 752 753
		rc = GRO_DROP;
		dev_kfree_skb(skb);
		stats->rx_replay++;
		goto stats;
	}

754 755 756 757 758 759
	/* check errors reported by HW and update statistics */
	if (unlikely(wil->txrx_ops.rx_error_check(wil, skb, stats))) {
		dev_kfree_skb(skb);
		return;
	}

760
	if (wdev->iftype == NL80211_IFTYPE_AP && !vif->ap_isolate) {
761 762 763 764 765 766
		if (mcast) {
			/* send multicast frames both to higher layers in
			 * local net stack and back to the wireless medium
			 */
			xmit_skb = skb_copy(skb, GFP_ATOMIC);
		} else {
767 768
			int xmit_cid = wil_find_cid(wil, vif->mid,
						    eth->h_dest);
769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793

			if (xmit_cid >= 0) {
				/* The destination station is associated to
				 * this AP (in this VLAN), so send the frame
				 * directly to it and do not pass it to local
				 * net stack.
				 */
				xmit_skb = skb;
				skb = NULL;
			}
		}
	}
	if (xmit_skb) {
		/* Send to wireless media and increase priority by 256 to
		 * keep the received priority instead of reclassifying
		 * the frame (see cfg80211_classify8021d).
		 */
		xmit_skb->dev = ndev;
		xmit_skb->priority += 256;
		xmit_skb->protocol = htons(ETH_P_802_3);
		skb_reset_network_header(xmit_skb);
		skb_reset_mac_header(xmit_skb);
		wil_dbg_txrx(wil, "Rx -> Tx %d bytes\n", len);
		dev_queue_xmit(xmit_skb);
	}
794

795 796
	if (skb) { /* deliver to local stack */
		skb->protocol = eth_type_trans(skb, ndev);
797
		skb->dev = ndev;
798 799 800 801
		rc = napi_gro_receive(&wil->napi_rx, skb);
		wil_dbg_txrx(wil, "Rx complete %d bytes => %s\n",
			     len, gro_res_str[rc]);
	}
802
stats:
803
	/* statistics. rc set to GRO_NORMAL for AP bridging */
V
Vladimir Kondratiev 已提交
804 805 806 807 808
	if (unlikely(rc == GRO_DROP)) {
		ndev->stats.rx_dropped++;
		stats->rx_dropped++;
		wil_dbg_txrx(wil, "Rx drop %d bytes\n", len);
	} else {
809
		ndev->stats.rx_packets++;
810
		stats->rx_packets++;
811
		ndev->stats.rx_bytes += len;
812
		stats->rx_bytes += len;
813 814
		if (mcast)
			ndev->stats.multicast++;
815
	}
816 817 818 819 820
}

/**
 * Proceed all completed skb's from Rx VRING
 *
V
Vladimir Kondratiev 已提交
821
 * Safe to call from NAPI poll, i.e. softirq with interrupts enabled
822
 */
V
Vladimir Kondratiev 已提交
823
void wil_rx_handle(struct wil6210_priv *wil, int *quota)
824
{
825 826
	struct net_device *ndev = wil->main_ndev;
	struct wireless_dev *wdev = ndev->ieee80211_ptr;
827
	struct wil_ring *v = &wil->ring_rx;
828 829
	struct sk_buff *skb;

830
	if (unlikely(!v->va)) {
831 832 833
		wil_err(wil, "Rx IRQ while Rx not yet initialized\n");
		return;
	}
834
	wil_dbg_txrx(wil, "rx_handle\n");
V
Vladimir Kondratiev 已提交
835 836
	while ((*quota > 0) && (NULL != (skb = wil_vring_reap_rx(wil, v)))) {
		(*quota)--;
837

838 839
		/* monitor is currently supported on main interface only */
		if (wdev->iftype == NL80211_IFTYPE_MONITOR) {
840 841 842 843 844
			skb->dev = ndev;
			skb_reset_mac_header(skb);
			skb->ip_summed = CHECKSUM_UNNECESSARY;
			skb->pkt_type = PACKET_OTHERHOST;
			skb->protocol = htons(ETH_P_802_2);
V
Vladimir Kondratiev 已提交
845
			wil_netif_rx_any(skb, ndev);
846
		} else {
847
			wil_rx_reorder(wil, skb);
848 849 850 851 852
		}
	}
	wil_rx_refill(wil, v->size);
}

L
Lior David 已提交
853 854 855 856 857 858 859 860 861 862 863 864 865 866
static void wil_rx_buf_len_init(struct wil6210_priv *wil)
{
	wil->rx_buf_len = rx_large_buf ?
		WIL_MAX_ETH_MTU : TXRX_BUF_LEN_DEFAULT - WIL_MAX_MPDU_OVERHEAD;
	if (mtu_max > wil->rx_buf_len) {
		/* do not allow RX buffers to be smaller than mtu_max, for
		 * backward compatibility (mtu_max parameter was also used
		 * to support receiving large packets)
		 */
		wil_info(wil, "Override RX buffer to mtu_max(%d)\n", mtu_max);
		wil->rx_buf_len = mtu_max;
	}
}

867
static int wil_rx_init(struct wil6210_priv *wil, u16 size)
868
{
869
	struct wil_ring *vring = &wil->ring_rx;
870 871
	int rc;

872
	wil_dbg_misc(wil, "rx_init\n");
873

874 875 876 877 878
	if (vring->va) {
		wil_err(wil, "Rx ring already allocated\n");
		return -EINVAL;
	}

L
Lior David 已提交
879 880
	wil_rx_buf_len_init(wil);

881
	vring->size = size;
882
	vring->is_rx = true;
883 884 885 886
	rc = wil_vring_alloc(wil, vring);
	if (rc)
		return rc;

887
	rc = wmi_rx_chain_add(wil, vring);
888 889 890 891 892 893 894 895 896
	if (rc)
		goto err_free;

	rc = wil_rx_refill(wil, vring->size);
	if (rc)
		goto err_free;

	return 0;
 err_free:
897
	wil_vring_free(wil, vring);
898 899 900 901

	return rc;
}

902
static void wil_rx_fini(struct wil6210_priv *wil)
903
{
904
	struct wil_ring *vring = &wil->ring_rx;
905

906
	wil_dbg_misc(wil, "rx_fini\n");
907

908
	if (vring->va)
909
		wil_vring_free(wil, vring);
910 911
}

912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935
static int wil_tx_desc_map(union wil_tx_desc *desc, dma_addr_t pa,
			   u32 len, int vring_index)
{
	struct vring_tx_desc *d = &desc->legacy;

	wil_desc_addr_set(&d->dma.addr, pa);
	d->dma.ip_length = 0;
	/* 0..6: mac_length; 7:ip_version 0-IP6 1-IP4*/
	d->dma.b11 = 0/*14 | BIT(7)*/;
	d->dma.error = 0;
	d->dma.status = 0; /* BIT(0) should be 0 for HW_OWNED */
	d->dma.length = cpu_to_le16((u16)len);
	d->dma.d0 = (vring_index << DMA_CFG_DESC_TX_0_QID_POS);
	d->mac.d[0] = 0;
	d->mac.d[1] = 0;
	d->mac.d[2] = 0;
	d->mac.ucode_cmd = 0;
	/* translation type:  0 - bypass; 1 - 802.3; 2 - native wifi */
	d->mac.d[2] = BIT(MAC_CFG_DESC_TX_2_SNAP_HDR_INSERTION_EN_POS) |
		      (1 << MAC_CFG_DESC_TX_2_L2_TRANSLATION_TYPE_POS);

	return 0;
}

936
void wil_tx_data_init(struct wil_ring_tx_data *txdata)
937 938 939 940 941 942 943 944 945 946 947
{
	spin_lock_bh(&txdata->lock);
	txdata->dot1x_open = 0;
	txdata->enabled = 0;
	txdata->idle = 0;
	txdata->last_idle = 0;
	txdata->begin = 0;
	txdata->agg_wsize = 0;
	txdata->agg_timeout = 0;
	txdata->agg_amsdu = 0;
	txdata->addba_in_progress = false;
948
	txdata->mid = U8_MAX;
949 950 951
	spin_unlock_bh(&txdata->lock);
}

952 953
static int wil_vring_init_tx(struct wil6210_vif *vif, int id, int size,
			     int cid, int tid)
954
{
955
	struct wil6210_priv *wil = vif_to_wil(vif);
956 957 958 959 960
	int rc;
	struct wmi_vring_cfg_cmd cmd = {
		.action = cpu_to_le32(WMI_VRING_CMD_ADD),
		.vring_cfg = {
			.tx_sw_ring = {
961
				.max_mpdu_size =
962
					cpu_to_le16(wil_mtu2macbuf(mtu_max)),
963
				.ring_size = cpu_to_le16(size),
964 965
			},
			.ringid = id,
966
			.cidxtid = mk_cidxtid(cid, tid),
967 968 969
			.encap_trans_type = WMI_VRING_ENC_TYPE_802_3,
			.mac_ctrl = 0,
			.to_resolution = 0,
V
Vladimir Kondratiev 已提交
970
			.agg_max_wsize = 0,
971 972 973 974 975 976 977
			.schd_params = {
				.priority = cpu_to_le16(0),
				.timeslot_us = cpu_to_le16(0xfff),
			},
		},
	};
	struct {
L
Lior David 已提交
978
		struct wmi_cmd_hdr wmi;
979
		struct wmi_vring_cfg_done_event cmd;
980 981 982
	} __packed reply = {
		.cmd = {.status = WMI_FW_STATUS_FAILURE},
	};
983 984
	struct wil_ring *vring = &wil->ring_tx[id];
	struct wil_ring_tx_data *txdata = &wil->ring_tx_data[id];
985

986
	wil_dbg_misc(wil, "vring_init_tx: max_mpdu_size %d\n",
987
		     cmd.vring_cfg.tx_sw_ring.max_mpdu_size);
988
	lockdep_assert_held(&wil->mutex);
989

990 991 992 993 994 995
	if (vring->va) {
		wil_err(wil, "Tx ring [%d] already allocated\n", id);
		rc = -EINVAL;
		goto out;
	}

996
	wil_tx_data_init(txdata);
997
	vring->is_rx = false;
998 999 1000 1001 1002
	vring->size = size;
	rc = wil_vring_alloc(wil, vring);
	if (rc)
		goto out;

1003 1004
	wil->ring2cid_tid[id][0] = cid;
	wil->ring2cid_tid[id][1] = tid;
1005

1006 1007
	cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa);

1008
	if (!vif->privacy)
1009
		txdata->dot1x_open = true;
1010
	rc = wmi_call(wil, WMI_VRING_CFG_CMDID, vif->mid, &cmd, sizeof(cmd),
1011 1012 1013 1014
		      WMI_VRING_CFG_DONE_EVENTID, &reply, sizeof(reply), 100);
	if (rc)
		goto out_free;

1015
	if (reply.cmd.status != WMI_FW_STATUS_SUCCESS) {
1016 1017
		wil_err(wil, "Tx config failed, status 0x%02x\n",
			reply.cmd.status);
1018
		rc = -EINVAL;
1019 1020 1021
		goto out_free;
	}

1022 1023
	spin_lock_bh(&txdata->lock);
	vring->hwtail = le32_to_cpu(reply.cmd.tx_vring_tail_ptr);
1024
	txdata->mid = vif->mid;
1025
	txdata->enabled = 1;
1026 1027
	spin_unlock_bh(&txdata->lock);

1028
	if (txdata->dot1x_open && (agg_wsize >= 0))
1029
		wil_addba_tx_request(wil, id, agg_wsize);
1030

1031 1032
	return 0;
 out_free:
1033
	spin_lock_bh(&txdata->lock);
1034 1035
	txdata->dot1x_open = false;
	txdata->enabled = 0;
1036
	spin_unlock_bh(&txdata->lock);
1037
	wil_vring_free(wil, vring);
1038 1039
	wil->ring2cid_tid[id][0] = WIL6210_MAX_CID;
	wil->ring2cid_tid[id][1] = 0;
1040

1041 1042 1043 1044 1045
 out:

	return rc;
}

1046
int wil_vring_init_bcast(struct wil6210_vif *vif, int id, int size)
1047
{
1048
	struct wil6210_priv *wil = vif_to_wil(vif);
1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062
	int rc;
	struct wmi_bcast_vring_cfg_cmd cmd = {
		.action = cpu_to_le32(WMI_VRING_CMD_ADD),
		.vring_cfg = {
			.tx_sw_ring = {
				.max_mpdu_size =
					cpu_to_le16(wil_mtu2macbuf(mtu_max)),
				.ring_size = cpu_to_le16(size),
			},
			.ringid = id,
			.encap_trans_type = WMI_VRING_ENC_TYPE_802_3,
		},
	};
	struct {
L
Lior David 已提交
1063
		struct wmi_cmd_hdr wmi;
1064
		struct wmi_vring_cfg_done_event cmd;
1065 1066 1067
	} __packed reply = {
		.cmd = {.status = WMI_FW_STATUS_FAILURE},
	};
1068 1069
	struct wil_ring *vring = &wil->ring_tx[id];
	struct wil_ring_tx_data *txdata = &wil->ring_tx_data[id];
1070

1071
	wil_dbg_misc(wil, "vring_init_bcast: max_mpdu_size %d\n",
1072
		     cmd.vring_cfg.tx_sw_ring.max_mpdu_size);
1073
	lockdep_assert_held(&wil->mutex);
1074 1075 1076 1077 1078 1079 1080

	if (vring->va) {
		wil_err(wil, "Tx ring [%d] already allocated\n", id);
		rc = -EINVAL;
		goto out;
	}

1081
	wil_tx_data_init(txdata);
1082
	vring->is_rx = false;
1083 1084 1085 1086 1087
	vring->size = size;
	rc = wil_vring_alloc(wil, vring);
	if (rc)
		goto out;

1088 1089
	wil->ring2cid_tid[id][0] = WIL6210_MAX_CID; /* CID */
	wil->ring2cid_tid[id][1] = 0; /* TID */
1090 1091 1092

	cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa);

1093
	if (!vif->privacy)
1094
		txdata->dot1x_open = true;
1095 1096
	rc = wmi_call(wil, WMI_BCAST_VRING_CFG_CMDID, vif->mid,
		      &cmd, sizeof(cmd),
1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107
		      WMI_VRING_CFG_DONE_EVENTID, &reply, sizeof(reply), 100);
	if (rc)
		goto out_free;

	if (reply.cmd.status != WMI_FW_STATUS_SUCCESS) {
		wil_err(wil, "Tx config failed, status 0x%02x\n",
			reply.cmd.status);
		rc = -EINVAL;
		goto out_free;
	}

1108 1109
	spin_lock_bh(&txdata->lock);
	vring->hwtail = le32_to_cpu(reply.cmd.tx_vring_tail_ptr);
1110
	txdata->mid = vif->mid;
1111
	txdata->enabled = 1;
1112
	spin_unlock_bh(&txdata->lock);
1113 1114 1115

	return 0;
 out_free:
1116
	spin_lock_bh(&txdata->lock);
1117 1118
	txdata->enabled = 0;
	txdata->dot1x_open = false;
1119
	spin_unlock_bh(&txdata->lock);
1120
	wil_vring_free(wil, vring);
1121 1122 1123 1124 1125
 out:

	return rc;
}

1126 1127 1128
static struct wil_ring *wil_find_tx_ucast(struct wil6210_priv *wil,
					  struct wil6210_vif *vif,
					  struct sk_buff *skb)
1129
{
1130 1131
	int i;
	struct ethhdr *eth = (void *)skb->data;
1132
	int cid = wil_find_cid(wil, vif->mid, eth->h_dest);
1133
	int min_ring_id = wil_get_min_tx_ring_id(wil);
1134 1135 1136

	if (cid < 0)
		return NULL;
1137

1138
	/* TODO: fix for multiple TID */
1139
	for (i = min_ring_id; i < ARRAY_SIZE(wil->ring2cid_tid); i++) {
1140 1141
		if (!wil->ring_tx_data[i].dot1x_open &&
		    skb->protocol != cpu_to_be16(ETH_P_PAE))
1142
			continue;
1143 1144 1145
		if (wil->ring2cid_tid[i][0] == cid) {
			struct wil_ring *v = &wil->ring_tx[i];
			struct wil_ring_tx_data *txdata = &wil->ring_tx_data[i];
1146

1147 1148
			wil_dbg_txrx(wil, "find_tx_ucast: (%pM) -> [%d]\n",
				     eth->h_dest, i);
1149
			if (v->va && txdata->enabled) {
1150 1151
				return v;
			} else {
1152 1153 1154
				wil_dbg_txrx(wil,
					     "find_tx_ucast: vring[%d] not valid\n",
					     i);
1155 1156 1157 1158
				return NULL;
			}
		}
	}
1159 1160 1161 1162

	return NULL;
}

1163 1164
static int wil_tx_ring(struct wil6210_priv *wil, struct wil6210_vif *vif,
		       struct wil_ring *ring, struct sk_buff *skb);
1165

1166 1167 1168
static struct wil_ring *wil_find_tx_ring_sta(struct wil6210_priv *wil,
					     struct wil6210_vif *vif,
					     struct sk_buff *skb)
1169
{
1170
	struct wil_ring *ring;
1171 1172
	int i;
	u8 cid;
1173
	struct wil_ring_tx_data  *txdata;
1174
	int min_ring_id = wil_get_min_tx_ring_id(wil);
1175 1176 1177

	/* In the STA mode, it is expected to have only 1 VRING
	 * for the AP we connected to.
1178
	 * find 1-st vring eligible for this skb and use it.
1179
	 */
1180
	for (i = min_ring_id; i < WIL6210_MAX_TX_RINGS; i++) {
1181 1182 1183
		ring = &wil->ring_tx[i];
		txdata = &wil->ring_tx_data[i];
		if (!ring->va || !txdata->enabled || txdata->mid != vif->mid)
1184 1185
			continue;

1186
		cid = wil->ring2cid_tid[i][0];
1187 1188 1189
		if (cid >= WIL6210_MAX_CID) /* skip BCAST */
			continue;

1190 1191
		if (!wil->ring_tx_data[i].dot1x_open &&
		    skb->protocol != cpu_to_be16(ETH_P_PAE))
1192
			continue;
1193 1194 1195

		wil_dbg_txrx(wil, "Tx -> ring %d\n", i);

1196
		return ring;
1197 1198
	}

1199
	wil_dbg_txrx(wil, "Tx while no rings active?\n");
1200 1201 1202 1203

	return NULL;
}

1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214
/* Use one of 2 strategies:
 *
 * 1. New (real broadcast):
 *    use dedicated broadcast vring
 * 2. Old (pseudo-DMS):
 *    Find 1-st vring and return it;
 *    duplicate skb and send it to other active vrings;
 *    in all cases override dest address to unicast peer's address
 * Use old strategy when new is not supported yet:
 *  - for PBSS
 */
1215 1216 1217
static struct wil_ring *wil_find_tx_bcast_1(struct wil6210_priv *wil,
					    struct wil6210_vif *vif,
					    struct sk_buff *skb)
V
Vladimir Kondratiev 已提交
1218
{
1219 1220 1221
	struct wil_ring *v;
	struct wil_ring_tx_data *txdata;
	int i = vif->bcast_ring;
1222

1223 1224
	if (i < 0)
		return NULL;
1225 1226
	v = &wil->ring_tx[i];
	txdata = &wil->ring_tx_data[i];
1227
	if (!v->va || !txdata->enabled)
1228
		return NULL;
1229 1230
	if (!wil->ring_tx_data[i].dot1x_open &&
	    skb->protocol != cpu_to_be16(ETH_P_PAE))
1231
		return NULL;
V
Vladimir Kondratiev 已提交
1232 1233 1234 1235

	return v;
}

1236 1237 1238 1239
static void wil_set_da_for_vring(struct wil6210_priv *wil,
				 struct sk_buff *skb, int vring_index)
{
	struct ethhdr *eth = (void *)skb->data;
1240
	int cid = wil->ring2cid_tid[vring_index][0];
1241 1242 1243 1244

	ether_addr_copy(eth->h_dest, wil->sta[cid].addr);
}

1245 1246 1247
static struct wil_ring *wil_find_tx_bcast_2(struct wil6210_priv *wil,
					    struct wil6210_vif *vif,
					    struct sk_buff *skb)
1248
{
1249
	struct wil_ring *v, *v2;
1250 1251 1252 1253 1254
	struct sk_buff *skb2;
	int i;
	u8 cid;
	struct ethhdr *eth = (void *)skb->data;
	char *src = eth->h_source;
1255
	struct wil_ring_tx_data *txdata, *txdata2;
1256
	int min_ring_id = wil_get_min_tx_ring_id(wil);
1257 1258

	/* find 1-st vring eligible for data */
1259
	for (i = min_ring_id; i < WIL6210_MAX_TX_RINGS; i++) {
1260 1261
		v = &wil->ring_tx[i];
		txdata = &wil->ring_tx_data[i];
1262
		if (!v->va || !txdata->enabled || txdata->mid != vif->mid)
1263 1264
			continue;

1265
		cid = wil->ring2cid_tid[i][0];
1266 1267
		if (cid >= WIL6210_MAX_CID) /* skip BCAST */
			continue;
1268 1269
		if (!wil->ring_tx_data[i].dot1x_open &&
		    skb->protocol != cpu_to_be16(ETH_P_PAE))
1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288
			continue;

		/* don't Tx back to source when re-routing Rx->Tx at the AP */
		if (0 == memcmp(wil->sta[cid].addr, src, ETH_ALEN))
			continue;

		goto found;
	}

	wil_dbg_txrx(wil, "Tx while no vrings active?\n");

	return NULL;

found:
	wil_dbg_txrx(wil, "BCAST -> ring %d\n", i);
	wil_set_da_for_vring(wil, skb, i);

	/* find other active vrings and duplicate skb for each */
	for (i++; i < WIL6210_MAX_TX_RINGS; i++) {
1289 1290
		v2 = &wil->ring_tx[i];
		txdata2 = &wil->ring_tx_data[i];
1291
		if (!v2->va || txdata2->mid != vif->mid)
1292
			continue;
1293
		cid = wil->ring2cid_tid[i][0];
1294 1295
		if (cid >= WIL6210_MAX_CID) /* skip BCAST */
			continue;
1296 1297
		if (!wil->ring_tx_data[i].dot1x_open &&
		    skb->protocol != cpu_to_be16(ETH_P_PAE))
1298 1299 1300 1301 1302 1303 1304 1305 1306
			continue;

		if (0 == memcmp(wil->sta[cid].addr, src, ETH_ALEN))
			continue;

		skb2 = skb_copy(skb, GFP_ATOMIC);
		if (skb2) {
			wil_dbg_txrx(wil, "BCAST DUP -> ring %d\n", i);
			wil_set_da_for_vring(wil, skb2, i);
1307
			wil_tx_ring(wil, vif, v2, skb2);
1308 1309 1310 1311 1312 1313 1314 1315
		} else {
			wil_err(wil, "skb_copy failed\n");
		}
	}

	return v;
}

1316 1317 1318
static inline
void wil_tx_desc_set_nr_frags(struct vring_tx_desc *d, int nr_frags)
{
1319
	d->mac.d[2] |= (nr_frags << MAC_CFG_DESC_TX_2_NUM_OF_DESCRIPTORS_POS);
1320 1321
}

1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332
/**
 * Sets the descriptor @d up for csum and/or TSO offloading. The corresponding
 * @skb is used to obtain the protocol and headers length.
 * @tso_desc_type is a descriptor type for TSO: 0 - a header, 1 - first data,
 * 2 - middle, 3 - last descriptor.
 */

static void wil_tx_desc_offload_setup_tso(struct vring_tx_desc *d,
					  struct sk_buff *skb,
					  int tso_desc_type, bool is_ipv4,
					  int tcp_hdr_len, int skb_net_hdr_len)
1333
{
1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364
	d->dma.b11 = ETH_HLEN; /* MAC header length */
	d->dma.b11 |= is_ipv4 << DMA_CFG_DESC_TX_OFFLOAD_CFG_L3T_IPV4_POS;

	d->dma.d0 |= (2 << DMA_CFG_DESC_TX_0_L4_TYPE_POS);
	/* L4 header len: TCP header length */
	d->dma.d0 |= (tcp_hdr_len & DMA_CFG_DESC_TX_0_L4_LENGTH_MSK);

	/* Setup TSO: bit and desc type */
	d->dma.d0 |= (BIT(DMA_CFG_DESC_TX_0_TCP_SEG_EN_POS)) |
		(tso_desc_type << DMA_CFG_DESC_TX_0_SEGMENT_BUF_DETAILS_POS);
	d->dma.d0 |= (is_ipv4 << DMA_CFG_DESC_TX_0_IPV4_CHECKSUM_EN_POS);

	d->dma.ip_length = skb_net_hdr_len;
	/* Enable TCP/UDP checksum */
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_TCP_UDP_CHECKSUM_EN_POS);
	/* Calculate pseudo-header */
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_PSEUDO_HEADER_CALC_EN_POS);
}

/**
 * Sets the descriptor @d up for csum. The corresponding
 * @skb is used to obtain the protocol and headers length.
 * Returns the protocol: 0 - not TCP, 1 - TCPv4, 2 - TCPv6.
 * Note, if d==NULL, the function only returns the protocol result.
 *
 * It is very similar to previous wil_tx_desc_offload_setup_tso. This
 * is "if unrolling" to optimize the critical path.
 */

static int wil_tx_desc_offload_setup(struct vring_tx_desc *d,
				     struct sk_buff *skb){
1365 1366 1367 1368 1369
	int protocol;

	if (skb->ip_summed != CHECKSUM_PARTIAL)
		return 0;

1370 1371
	d->dma.b11 = ETH_HLEN; /* MAC header length */

1372 1373 1374
	switch (skb->protocol) {
	case cpu_to_be16(ETH_P_IP):
		protocol = ip_hdr(skb)->protocol;
1375
		d->dma.b11 |= BIT(DMA_CFG_DESC_TX_OFFLOAD_CFG_L3T_IPV4_POS);
1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408
		break;
	case cpu_to_be16(ETH_P_IPV6):
		protocol = ipv6_hdr(skb)->nexthdr;
		break;
	default:
		return -EINVAL;
	}

	switch (protocol) {
	case IPPROTO_TCP:
		d->dma.d0 |= (2 << DMA_CFG_DESC_TX_0_L4_TYPE_POS);
		/* L4 header len: TCP header length */
		d->dma.d0 |=
		(tcp_hdrlen(skb) & DMA_CFG_DESC_TX_0_L4_LENGTH_MSK);
		break;
	case IPPROTO_UDP:
		/* L4 header len: UDP header length */
		d->dma.d0 |=
		(sizeof(struct udphdr) & DMA_CFG_DESC_TX_0_L4_LENGTH_MSK);
		break;
	default:
		return -EINVAL;
	}

	d->dma.ip_length = skb_network_header_len(skb);
	/* Enable TCP/UDP checksum */
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_TCP_UDP_CHECKSUM_EN_POS);
	/* Calculate pseudo-header */
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_PSEUDO_HEADER_CALC_EN_POS);

	return 0;
}

1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421
static inline void wil_tx_last_desc(struct vring_tx_desc *d)
{
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_CMD_EOP_POS) |
	      BIT(DMA_CFG_DESC_TX_0_CMD_MARK_WB_POS) |
	      BIT(DMA_CFG_DESC_TX_0_CMD_DMA_IT_POS);
}

static inline void wil_set_tx_desc_last_tso(volatile struct vring_tx_desc *d)
{
	d->dma.d0 |= wil_tso_type_lst <<
		  DMA_CFG_DESC_TX_0_SEGMENT_BUF_DETAILS_POS;
}

1422
static int __wil_tx_vring_tso(struct wil6210_priv *wil, struct wil6210_vif *vif,
1423
			      struct wil_ring *vring, struct sk_buff *skb)
1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442
{
	struct device *dev = wil_to_dev(wil);

	/* point to descriptors in shared memory */
	volatile struct vring_tx_desc *_desc = NULL, *_hdr_desc,
				      *_first_desc = NULL;

	/* pointers to shadow descriptors */
	struct vring_tx_desc desc_mem, hdr_desc_mem, first_desc_mem,
			     *d = &hdr_desc_mem, *hdr_desc = &hdr_desc_mem,
			     *first_desc = &first_desc_mem;

	/* pointer to shadow descriptors' context */
	struct wil_ctx *hdr_ctx, *first_ctx = NULL;

	int descs_used = 0; /* total number of used descriptors */
	int sg_desc_cnt = 0; /* number of descriptors for current mss*/

	u32 swhead = vring->swhead;
1443
	int used, avail = wil_ring_avail_tx(vring);
1444 1445 1446 1447
	int nr_frags = skb_shinfo(skb)->nr_frags;
	int min_desc_required = nr_frags + 1;
	int mss = skb_shinfo(skb)->gso_size;	/* payload size w/o headers */
	int f, len, hdrlen, headlen;
1448 1449
	int vring_index = vring - wil->ring_tx;
	struct wil_ring_tx_data *txdata = &wil->ring_tx_data[vring_index];
1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460
	uint i = swhead;
	dma_addr_t pa;
	const skb_frag_t *frag = NULL;
	int rem_data = mss;
	int lenmss;
	int hdr_compensation_need = true;
	int desc_tso_type = wil_tso_type_first;
	bool is_ipv4;
	int tcp_hdr_len;
	int skb_net_hdr_len;
	int gso_type;
H
Hamad Kadmany 已提交
1461
	int rc = -EINVAL;
1462

1463 1464
	wil_dbg_txrx(wil, "tx_vring_tso: %d bytes to vring %d\n", skb->len,
		     vring_index);
1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516

	if (unlikely(!txdata->enabled))
		return -EINVAL;

	/* A typical page 4K is 3-4 payloads, we assume each fragment
	 * is a full payload, that's how min_desc_required has been
	 * calculated. In real we might need more or less descriptors,
	 * this is the initial check only.
	 */
	if (unlikely(avail < min_desc_required)) {
		wil_err_ratelimited(wil,
				    "TSO: Tx ring[%2d] full. No space for %d fragments\n",
				    vring_index, min_desc_required);
		return -ENOMEM;
	}

	/* Header Length = MAC header len + IP header len + TCP header len*/
	hdrlen = ETH_HLEN +
		(int)skb_network_header_len(skb) +
		tcp_hdrlen(skb);

	gso_type = skb_shinfo(skb)->gso_type & (SKB_GSO_TCPV6 | SKB_GSO_TCPV4);
	switch (gso_type) {
	case SKB_GSO_TCPV4:
		/* TCP v4, zero out the IP length and IPv4 checksum fields
		 * as required by the offloading doc
		 */
		ip_hdr(skb)->tot_len = 0;
		ip_hdr(skb)->check = 0;
		is_ipv4 = true;
		break;
	case SKB_GSO_TCPV6:
		/* TCP v6, zero out the payload length */
		ipv6_hdr(skb)->payload_len = 0;
		is_ipv4 = false;
		break;
	default:
		/* other than TCPv4 or TCPv6 types are not supported for TSO.
		 * It is also illegal for both to be set simultaneously
		 */
		return -EINVAL;
	}

	if (skb->ip_summed != CHECKSUM_PARTIAL)
		return -EINVAL;

	/* tcp header length and skb network header length are fixed for all
	 * packet's descriptors - read then once here
	 */
	tcp_hdr_len = tcp_hdrlen(skb);
	skb_net_hdr_len = skb_network_header_len(skb);

1517
	_hdr_desc = &vring->va[i].tx.legacy;
1518 1519 1520 1521 1522 1523 1524

	pa = dma_map_single(dev, skb->data, hdrlen, DMA_TO_DEVICE);
	if (unlikely(dma_mapping_error(dev, pa))) {
		wil_err(wil, "TSO: Skb head DMA map error\n");
		goto err_exit;
	}

1525 1526
	wil->txrx_ops.tx_desc_map((union wil_tx_desc *)hdr_desc, pa,
				  hdrlen, vring_index);
1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553
	wil_tx_desc_offload_setup_tso(hdr_desc, skb, wil_tso_type_hdr, is_ipv4,
				      tcp_hdr_len, skb_net_hdr_len);
	wil_tx_last_desc(hdr_desc);

	vring->ctx[i].mapped_as = wil_mapped_as_single;
	hdr_ctx = &vring->ctx[i];

	descs_used++;
	headlen = skb_headlen(skb) - hdrlen;

	for (f = headlen ? -1 : 0; f < nr_frags; f++)  {
		if (headlen) {
			len = headlen;
			wil_dbg_txrx(wil, "TSO: process skb head, len %u\n",
				     len);
		} else {
			frag = &skb_shinfo(skb)->frags[f];
			len = frag->size;
			wil_dbg_txrx(wil, "TSO: frag[%d]: len %u\n", f, len);
		}

		while (len) {
			wil_dbg_txrx(wil,
				     "TSO: len %d, rem_data %d, descs_used %d\n",
				     len, rem_data, descs_used);

			if (descs_used == avail)  {
H
Hamad Kadmany 已提交
1554 1555 1556
				wil_err_ratelimited(wil, "TSO: ring overflow\n");
				rc = -ENOMEM;
				goto mem_error;
1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577
			}

			lenmss = min_t(int, rem_data, len);
			i = (swhead + descs_used) % vring->size;
			wil_dbg_txrx(wil, "TSO: lenmss %d, i %d\n", lenmss, i);

			if (!headlen) {
				pa = skb_frag_dma_map(dev, frag,
						      frag->size - len, lenmss,
						      DMA_TO_DEVICE);
				vring->ctx[i].mapped_as = wil_mapped_as_page;
			} else {
				pa = dma_map_single(dev,
						    skb->data +
						    skb_headlen(skb) - headlen,
						    lenmss,
						    DMA_TO_DEVICE);
				vring->ctx[i].mapped_as = wil_mapped_as_single;
				headlen -= lenmss;
			}

H
Hamad Kadmany 已提交
1578 1579 1580 1581
			if (unlikely(dma_mapping_error(dev, pa))) {
				wil_err(wil, "TSO: DMA map page error\n");
				goto mem_error;
			}
1582

1583
			_desc = &vring->va[i].tx.legacy;
1584 1585 1586 1587 1588 1589 1590 1591 1592

			if (!_first_desc) {
				_first_desc = _desc;
				first_ctx = &vring->ctx[i];
				d = first_desc;
			} else {
				d = &desc_mem;
			}

1593 1594
			wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d,
						  pa, lenmss, vring_index);
1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671
			wil_tx_desc_offload_setup_tso(d, skb, desc_tso_type,
						      is_ipv4, tcp_hdr_len,
						      skb_net_hdr_len);

			/* use tso_type_first only once */
			desc_tso_type = wil_tso_type_mid;

			descs_used++;  /* desc used so far */
			sg_desc_cnt++; /* desc used for this segment */
			len -= lenmss;
			rem_data -= lenmss;

			wil_dbg_txrx(wil,
				     "TSO: len %d, rem_data %d, descs_used %d, sg_desc_cnt %d,\n",
				     len, rem_data, descs_used, sg_desc_cnt);

			/* Close the segment if reached mss size or last frag*/
			if (rem_data == 0 || (f == nr_frags - 1 && len == 0)) {
				if (hdr_compensation_need) {
					/* first segment include hdr desc for
					 * release
					 */
					hdr_ctx->nr_frags = sg_desc_cnt;
					wil_tx_desc_set_nr_frags(first_desc,
								 sg_desc_cnt +
								 1);
					hdr_compensation_need = false;
				} else {
					wil_tx_desc_set_nr_frags(first_desc,
								 sg_desc_cnt);
				}
				first_ctx->nr_frags = sg_desc_cnt - 1;

				wil_tx_last_desc(d);

				/* first descriptor may also be the last
				 * for this mss - make sure not to copy
				 * it twice
				 */
				if (first_desc != d)
					*_first_desc = *first_desc;

				/*last descriptor will be copied at the end
				 * of this TS processing
				 */
				if (f < nr_frags - 1 || len > 0)
					*_desc = *d;

				rem_data = mss;
				_first_desc = NULL;
				sg_desc_cnt = 0;
			} else if (first_desc != d) /* update mid descriptor */
					*_desc = *d;
		}
	}

	/* first descriptor may also be the last.
	 * in this case d pointer is invalid
	 */
	if (_first_desc == _desc)
		d = first_desc;

	/* Last data descriptor */
	wil_set_tx_desc_last_tso(d);
	*_desc = *d;

	/* Fill the total number of descriptors in first desc (hdr)*/
	wil_tx_desc_set_nr_frags(hdr_desc, descs_used);
	*_hdr_desc = *hdr_desc;

	/* hold reference to skb
	 * to prevent skb release before accounting
	 * in case of immediate "tx done"
	 */
	vring->ctx[i].skb = skb_get(skb);

	/* performance monitoring */
1672 1673
	used = wil_ring_used_tx(vring);
	if (wil_val_in_range(wil->ring_idle_trsh,
1674 1675 1676 1677 1678 1679
			     used, used + descs_used)) {
		txdata->idle += get_cycles() - txdata->last_idle;
		wil_dbg_txrx(wil,  "Ring[%2d] not idle %d -> %d\n",
			     vring_index, used, used + descs_used);
	}

1680 1681 1682 1683 1684 1685 1686
	/* Make sure to advance the head only after descriptor update is done.
	 * This will prevent a race condition where the completion thread
	 * will see the DU bit set from previous run and will handle the
	 * skb before it was completed.
	 */
	wmb();

1687
	/* advance swhead */
1688
	wil_ring_advance_head(vring, descs_used);
H
Hamad Kadmany 已提交
1689
	wil_dbg_txrx(wil, "TSO: Tx swhead %d -> %d\n", swhead, vring->swhead);
1690 1691 1692 1693 1694 1695

	/* make sure all writes to descriptors (shared memory) are done before
	 * committing them to HW
	 */
	wmb();

1696
	wil_w(wil, vring->hwtail, vring->swhead);
1697 1698
	return 0;

H
Hamad Kadmany 已提交
1699
mem_error:
1700 1701 1702
	while (descs_used > 0) {
		struct wil_ctx *ctx;

1703
		i = (swhead + descs_used - 1) % vring->size;
1704 1705
		d = (struct vring_tx_desc *)&vring->va[i].tx.legacy;
		_desc = &vring->va[i].tx.legacy;
1706 1707 1708
		*d = *_desc;
		_desc->dma.status = TX_DMA_STATUS_DU;
		ctx = &vring->ctx[i];
1709
		wil_txdesc_unmap(dev, (union wil_tx_desc *)d, ctx);
1710 1711 1712 1713
		memset(ctx, 0, sizeof(*ctx));
		descs_used--;
	}
err_exit:
H
Hamad Kadmany 已提交
1714
	return rc;
1715 1716
}

1717 1718
static int __wil_tx_ring(struct wil6210_priv *wil, struct wil6210_vif *vif,
			 struct wil_ring *ring, struct sk_buff *skb)
1719 1720
{
	struct device *dev = wil_to_dev(wil);
1721 1722
	struct vring_tx_desc dd, *d = &dd;
	volatile struct vring_tx_desc *_d;
1723 1724
	u32 swhead = ring->swhead;
	int avail = wil_ring_avail_tx(ring);
1725
	int nr_frags = skb_shinfo(skb)->nr_frags;
1726
	uint f = 0;
1727 1728
	int ring_index = ring - wil->ring_tx;
	struct wil_ring_tx_data  *txdata = &wil->ring_tx_data[ring_index];
1729 1730
	uint i = swhead;
	dma_addr_t pa;
1731
	int used;
1732
	bool mcast = (ring_index == vif->bcast_ring);
1733
	uint len = skb_headlen(skb);
1734

1735
	wil_dbg_txrx(wil, "tx_ring: %d bytes to ring %d, nr_frags %d\n",
1736
		     skb->len, ring_index, nr_frags);
1737

1738 1739 1740
	if (unlikely(!txdata->enabled))
		return -EINVAL;

1741
	if (unlikely(avail < 1 + nr_frags)) {
1742
		wil_err_ratelimited(wil,
V
Vladimir Kondratiev 已提交
1743
				    "Tx ring[%2d] full. No space for %d fragments\n",
1744
				    ring_index, 1 + nr_frags);
1745 1746
		return -ENOMEM;
	}
1747
	_d = &ring->va[i].tx.legacy;
1748

1749
	pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE);
1750

1751
	wil_dbg_txrx(wil, "Tx[%2d] skb %d bytes 0x%p -> %pad\n", ring_index,
V
Vladimir Kondratiev 已提交
1752
		     skb_headlen(skb), skb->data, &pa);
1753
	wil_hex_dump_txrx("Tx ", DUMP_PREFIX_OFFSET, 16, 1,
1754 1755 1756 1757
			  skb->data, skb_headlen(skb), false);

	if (unlikely(dma_mapping_error(dev, pa)))
		return -EINVAL;
1758
	ring->ctx[i].mapped_as = wil_mapped_as_single;
1759
	/* 1-st segment */
1760 1761
	wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d, pa, len,
				   ring_index);
1762 1763
	if (unlikely(mcast)) {
		d->mac.d[0] |= BIT(MAC_CFG_DESC_TX_0_MCS_EN_POS); /* MCS 0 */
1764
		if (unlikely(len > WIL_BCAST_MCS0_LIMIT)) /* set MCS 1 */
1765 1766
			d->mac.d[0] |= (1 << MAC_CFG_DESC_TX_0_MCS_INDEX_POS);
	}
1767
	/* Process TCP/UDP checksum offloading */
1768
	if (unlikely(wil_tx_desc_offload_setup(d, skb))) {
V
Vladimir Kondratiev 已提交
1769
		wil_err(wil, "Tx[%2d] Failed to set cksum, drop packet\n",
1770
			ring_index);
1771 1772 1773
		goto dma_error;
	}

1774
	ring->ctx[i].nr_frags = nr_frags;
1775
	wil_tx_desc_set_nr_frags(d, nr_frags + 1);
1776

1777
	/* middle segments */
1778
	for (; f < nr_frags; f++) {
1779 1780 1781
		const struct skb_frag_struct *frag =
				&skb_shinfo(skb)->frags[f];
		int len = skb_frag_size(frag);
1782

1783
		*_d = *d;
1784
		wil_dbg_txrx(wil, "Tx[%2d] desc[%4d]\n", ring_index, i);
V
Vladimir Kondratiev 已提交
1785 1786
		wil_hex_dump_txrx("TxD ", DUMP_PREFIX_NONE, 32, 4,
				  (const void *)d, sizeof(*d), false);
1787 1788
		i = (swhead + f + 1) % ring->size;
		_d = &ring->va[i].tx.legacy;
1789
		pa = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag),
1790
				      DMA_TO_DEVICE);
H
Hamad Kadmany 已提交
1791 1792
		if (unlikely(dma_mapping_error(dev, pa))) {
			wil_err(wil, "Tx[%2d] failed to map fragment\n",
1793
				ring_index);
1794
			goto dma_error;
H
Hamad Kadmany 已提交
1795
		}
1796 1797 1798
		ring->ctx[i].mapped_as = wil_mapped_as_page;
		wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d,
					   pa, len, ring_index);
1799 1800 1801 1802
		/* no need to check return code -
		 * if it succeeded for 1-st descriptor,
		 * it will succeed here too
		 */
1803
		wil_tx_desc_offload_setup(d, skb);
1804 1805 1806
	}
	/* for the last seg only */
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_CMD_EOP_POS);
1807
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_CMD_MARK_WB_POS);
1808
	d->dma.d0 |= BIT(DMA_CFG_DESC_TX_0_CMD_DMA_IT_POS);
1809
	*_d = *d;
1810
	wil_dbg_txrx(wil, "Tx[%2d] desc[%4d]\n", ring_index, i);
V
Vladimir Kondratiev 已提交
1811 1812
	wil_hex_dump_txrx("TxD ", DUMP_PREFIX_NONE, 32, 4,
			  (const void *)d, sizeof(*d), false);
1813

1814 1815 1816 1817
	/* hold reference to skb
	 * to prevent skb release before accounting
	 * in case of immediate "tx done"
	 */
1818
	ring->ctx[i].skb = skb_get(skb);
1819

1820
	/* performance monitoring */
1821
	used = wil_ring_used_tx(ring);
1822
	if (wil_val_in_range(wil->ring_idle_trsh,
1823
			     used, used + nr_frags + 1)) {
1824
		txdata->idle += get_cycles() - txdata->last_idle;
1825
		wil_dbg_txrx(wil,  "Ring[%2d] not idle %d -> %d\n",
1826
			     ring_index, used, used + nr_frags + 1);
1827
	}
1828

1829 1830 1831 1832 1833 1834 1835
	/* Make sure to advance the head only after descriptor update is done.
	 * This will prevent a race condition where the completion thread
	 * will see the DU bit set from previous run and will handle the
	 * skb before it was completed.
	 */
	wmb();

1836
	/* advance swhead */
1837 1838 1839 1840
	wil_ring_advance_head(ring, nr_frags + 1);
	wil_dbg_txrx(wil, "Tx[%2d] swhead %d -> %d\n", ring_index, swhead,
		     ring->swhead);
	trace_wil6210_tx(ring_index, swhead, skb->len, nr_frags);
1841 1842 1843 1844 1845 1846

	/* make sure all writes to descriptors (shared memory) are done before
	 * committing them to HW
	 */
	wmb();

1847
	wil_w(wil, ring->hwtail, ring->swhead);
1848 1849 1850 1851

	return 0;
 dma_error:
	/* unmap what we have mapped */
1852 1853 1854
	nr_frags = f + 1; /* frags mapped + one for skb head */
	for (f = 0; f < nr_frags; f++) {
		struct wil_ctx *ctx;
1855

1856 1857 1858
		i = (swhead + f) % ring->size;
		ctx = &ring->ctx[i];
		_d = &ring->va[i].tx.legacy;
1859 1860
		*d = *_d;
		_d->dma.status = TX_DMA_STATUS_DU;
1861 1862 1863
		wil->txrx_ops.tx_desc_unmap(dev,
					    (union wil_tx_desc *)d,
					    ctx);
1864 1865

		memset(ctx, 0, sizeof(*ctx));
1866 1867 1868 1869 1870
	}

	return -EINVAL;
}

1871 1872
static int wil_tx_ring(struct wil6210_priv *wil, struct wil6210_vif *vif,
		       struct wil_ring *ring, struct sk_buff *skb)
1873
{
1874
	int ring_index = ring - wil->ring_tx;
1875
	struct wil_ring_tx_data *txdata = &wil->ring_tx_data[ring_index];
1876 1877 1878
	int rc;

	spin_lock(&txdata->lock);
1879

1880 1881 1882 1883 1884 1885 1886 1887 1888
	if (test_bit(wil_status_suspending, wil->status) ||
	    test_bit(wil_status_suspended, wil->status) ||
	    test_bit(wil_status_resuming, wil->status)) {
		wil_dbg_txrx(wil,
			     "suspend/resume in progress. drop packet\n");
		spin_unlock(&txdata->lock);
		return -EINVAL;
	}

1889 1890
	rc = (skb_is_gso(skb) ? wil->txrx_ops.tx_ring_tso : __wil_tx_ring)
	     (wil, vif, ring, skb);
1891

1892
	spin_unlock(&txdata->lock);
1893

1894 1895 1896
	return rc;
}

D
Dedy Lansky 已提交
1897 1898
/**
 * Check status of tx vrings and stop/wake net queues if needed
1899
 * It will start/stop net queues of a specific VIF net_device.
D
Dedy Lansky 已提交
1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914
 *
 * This function does one of two checks:
 * In case check_stop is true, will check if net queues need to be stopped. If
 * the conditions for stopping are met, netif_tx_stop_all_queues() is called.
 * In case check_stop is false, will check if net queues need to be waked. If
 * the conditions for waking are met, netif_tx_wake_all_queues() is called.
 * vring is the vring which is currently being modified by either adding
 * descriptors (tx) into it or removing descriptors (tx complete) from it. Can
 * be null when irrelevant (e.g. connect/disconnect events).
 *
 * The implementation is to stop net queues if modified vring has low
 * descriptor availability. Wake if all vrings are not in low descriptor
 * availability and modified vring has high descriptor availability.
 */
static inline void __wil_update_net_queues(struct wil6210_priv *wil,
1915
					   struct wil6210_vif *vif,
1916
					   struct wil_ring *ring,
D
Dedy Lansky 已提交
1917 1918 1919 1920
					   bool check_stop)
{
	int i;

1921 1922 1923
	if (unlikely(!vif))
		return;

1924
	if (ring)
1925
		wil_dbg_txrx(wil, "vring %d, mid %d, check_stop=%d, stopped=%d",
1926
			     (int)(ring - wil->ring_tx), vif->mid, check_stop,
1927
			     vif->net_queue_stopped);
D
Dedy Lansky 已提交
1928
	else
1929 1930
		wil_dbg_txrx(wil, "check_stop=%d, mid=%d, stopped=%d",
			     check_stop, vif->mid, vif->net_queue_stopped);
D
Dedy Lansky 已提交
1931

1932
	if (check_stop == vif->net_queue_stopped)
D
Dedy Lansky 已提交
1933 1934 1935 1936
		/* net queues already in desired state */
		return;

	if (check_stop) {
1937
		if (!ring || unlikely(wil_ring_avail_low(ring))) {
D
Dedy Lansky 已提交
1938
			/* not enough room in the vring */
1939 1940
			netif_tx_stop_all_queues(vif_to_ndev(vif));
			vif->net_queue_stopped = true;
D
Dedy Lansky 已提交
1941 1942 1943 1944 1945
			wil_dbg_txrx(wil, "netif_tx_stop called\n");
		}
		return;
	}

1946 1947 1948 1949 1950
	/* Do not wake the queues in suspend flow */
	if (test_bit(wil_status_suspending, wil->status) ||
	    test_bit(wil_status_suspended, wil->status))
		return;

D
Dedy Lansky 已提交
1951 1952
	/* check wake */
	for (i = 0; i < WIL6210_MAX_TX_RINGS; i++) {
1953 1954
		struct wil_ring *cur_ring = &wil->ring_tx[i];
		struct wil_ring_tx_data  *txdata = &wil->ring_tx_data[i];
D
Dedy Lansky 已提交
1955

1956 1957
		if (txdata->mid != vif->mid || !cur_ring->va ||
		    !txdata->enabled || cur_ring == ring)
D
Dedy Lansky 已提交
1958 1959
			continue;

1960 1961 1962
		if (wil_ring_avail_low(cur_ring)) {
			wil_dbg_txrx(wil, "ring %d full, can't wake\n",
				     (int)(cur_ring - wil->ring_tx));
D
Dedy Lansky 已提交
1963 1964 1965 1966
			return;
		}
	}

1967 1968
	if (!ring || wil_ring_avail_high(ring)) {
		/* enough room in the ring */
D
Dedy Lansky 已提交
1969
		wil_dbg_txrx(wil, "calling netif_tx_wake\n");
1970 1971
		netif_tx_wake_all_queues(vif_to_ndev(vif));
		vif->net_queue_stopped = false;
D
Dedy Lansky 已提交
1972 1973 1974
	}
}

1975
void wil_update_net_queues(struct wil6210_priv *wil, struct wil6210_vif *vif,
1976
			   struct wil_ring *ring, bool check_stop)
D
Dedy Lansky 已提交
1977 1978
{
	spin_lock(&wil->net_queue_lock);
1979
	__wil_update_net_queues(wil, vif, ring, check_stop);
D
Dedy Lansky 已提交
1980 1981 1982
	spin_unlock(&wil->net_queue_lock);
}

1983
void wil_update_net_queues_bh(struct wil6210_priv *wil, struct wil6210_vif *vif,
1984
			      struct wil_ring *ring, bool check_stop)
D
Dedy Lansky 已提交
1985 1986
{
	spin_lock_bh(&wil->net_queue_lock);
1987
	__wil_update_net_queues(wil, vif, ring, check_stop);
D
Dedy Lansky 已提交
1988 1989 1990
	spin_unlock_bh(&wil->net_queue_lock);
}

1991 1992
netdev_tx_t wil_start_xmit(struct sk_buff *skb, struct net_device *ndev)
{
1993 1994
	struct wil6210_vif *vif = ndev_to_vif(ndev);
	struct wil6210_priv *wil = vif_to_wil(vif);
1995
	struct ethhdr *eth = (void *)skb->data;
1996
	bool bcast = is_multicast_ether_addr(eth->h_dest);
1997
	struct wil_ring *ring;
1998
	static bool pr_once_fw;
1999 2000
	int rc;

2001
	wil_dbg_txrx(wil, "start_xmit\n");
2002
	if (unlikely(!test_bit(wil_status_fwready, wil->status))) {
2003 2004 2005 2006
		if (!pr_once_fw) {
			wil_err(wil, "FW not ready\n");
			pr_once_fw = true;
		}
2007 2008
		goto drop;
	}
2009 2010 2011
	if (unlikely(!test_bit(wil_vif_fwconnected, vif->status))) {
		wil_dbg_ratelimited(wil,
				    "VIF not connected, packet dropped\n");
2012 2013
		goto drop;
	}
2014
	if (unlikely(vif->wdev.iftype == NL80211_IFTYPE_MONITOR)) {
2015 2016 2017
		wil_err(wil, "Xmit in monitor mode not supported\n");
		goto drop;
	}
2018
	pr_once_fw = false;
2019 2020

	/* find vring */
2021
	if (vif->wdev.iftype == NL80211_IFTYPE_STATION && !vif->pbss) {
2022
		/* in STA mode (ESS), all to same VRING (to AP) */
2023
		ring = wil_find_tx_ring_sta(wil, vif, skb);
2024
	} else if (bcast) {
2025
		if (vif->pbss)
2026 2027 2028
			/* in pbss, no bcast VRING - duplicate skb in
			 * all stations VRINGs
			 */
2029
			ring = wil_find_tx_bcast_2(wil, vif, skb);
2030
		else if (vif->wdev.iftype == NL80211_IFTYPE_AP)
2031
			/* AP has a dedicated bcast VRING */
2032
			ring = wil_find_tx_bcast_1(wil, vif, skb);
2033 2034 2035 2036
		else
			/* unexpected combination, fallback to duplicating
			 * the skb in all stations VRINGs
			 */
2037
			ring = wil_find_tx_bcast_2(wil, vif, skb);
2038 2039
	} else {
		/* unicast, find specific VRING by dest. address */
2040
		ring = wil_find_tx_ucast(wil, vif, skb);
2041
	}
2042
	if (unlikely(!ring)) {
2043
		wil_dbg_txrx(wil, "No Tx RING found for %pM\n", eth->h_dest);
2044
		goto drop;
2045
	}
2046
	/* set up vring entry */
2047
	rc = wil_tx_ring(wil, vif, ring, skb);
2048

2049 2050
	switch (rc) {
	case 0:
D
Dedy Lansky 已提交
2051
		/* shall we stop net queues? */
2052
		wil_update_net_queues_bh(wil, vif, ring, true);
2053
		/* statistics will be updated on the tx_complete */
2054 2055 2056 2057 2058
		dev_kfree_skb_any(skb);
		return NETDEV_TX_OK;
	case -ENOMEM:
		return NETDEV_TX_BUSY;
	default:
2059
		break; /* goto drop; */
2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070
	}
 drop:
	ndev->stats.tx_dropped++;
	dev_kfree_skb_any(skb);

	return NET_XMIT_DROP;
}

/**
 * Clean up transmitted skb's from the Tx VRING
 *
V
Vladimir Kondratiev 已提交
2071 2072
 * Return number of descriptors cleared
 *
2073 2074
 * Safe to call from IRQ
 */
2075
int wil_tx_complete(struct wil6210_vif *vif, int ringid)
2076
{
2077 2078
	struct wil6210_priv *wil = vif_to_wil(vif);
	struct net_device *ndev = vif_to_ndev(vif);
2079
	struct device *dev = wil_to_dev(wil);
2080 2081
	struct wil_ring *vring = &wil->ring_tx[ringid];
	struct wil_ring_tx_data *txdata = &wil->ring_tx_data[ringid];
V
Vladimir Kondratiev 已提交
2082
	int done = 0;
2083
	int cid = wil->ring2cid_tid[ringid][0];
2084
	struct wil_net_stats *stats = NULL;
2085
	volatile struct vring_tx_desc *_d;
2086 2087
	int used_before_complete;
	int used_new;
2088

2089
	if (unlikely(!vring->va)) {
2090
		wil_err(wil, "Tx irq[%d]: vring not initialized\n", ringid);
V
Vladimir Kondratiev 已提交
2091
		return 0;
2092 2093
	}

2094
	if (unlikely(!txdata->enabled)) {
2095 2096 2097 2098
		wil_info(wil, "Tx irq[%d]: vring disabled\n", ringid);
		return 0;
	}

2099
	wil_dbg_txrx(wil, "tx_complete: (%d)\n", ringid);
2100

2101
	used_before_complete = wil_ring_used_tx(vring);
2102

2103 2104 2105
	if (cid < WIL6210_MAX_CID)
		stats = &wil->sta[cid].stats;

2106
	while (!wil_ring_is_empty(vring)) {
2107
		int new_swtail;
2108
		struct wil_ctx *ctx = &vring->ctx[vring->swtail];
2109 2110
		/**
		 * For the fragmented skb, HW will set DU bit only for the
2111 2112
		 * last fragment. look for it.
		 * In TSO the first DU will include hdr desc
2113 2114 2115
		 */
		int lf = (vring->swtail + ctx->nr_frags) % vring->size;
		/* TODO: check we are not past head */
2116

2117
		_d = &vring->va[lf].tx.legacy;
2118
		if (unlikely(!(_d->dma.status & TX_DMA_STATUS_DU)))
2119 2120
			break;

2121 2122 2123 2124
		new_swtail = (lf + 1) % vring->size;
		while (vring->swtail != new_swtail) {
			struct vring_tx_desc dd, *d = &dd;
			u16 dmalen;
2125 2126 2127 2128
			struct sk_buff *skb;

			ctx = &vring->ctx[vring->swtail];
			skb = ctx->skb;
2129
			_d = &vring->va[vring->swtail].tx.legacy;
2130

2131
			*d = *_d;
2132

2133 2134 2135 2136
			dmalen = le16_to_cpu(d->dma.length);
			trace_wil6210_tx_done(ringid, vring->swtail, dmalen,
					      d->dma.error);
			wil_dbg_txrx(wil,
V
Vladimir Kondratiev 已提交
2137 2138 2139 2140
				     "TxC[%2d][%3d] : %d bytes, status 0x%02x err 0x%02x\n",
				     ringid, vring->swtail, dmalen,
				     d->dma.status, d->dma.error);
			wil_hex_dump_txrx("TxCD ", DUMP_PREFIX_NONE, 32, 4,
2141
					  (const void *)d, sizeof(*d), false);
2142

2143 2144 2145
			wil->txrx_ops.tx_desc_unmap(dev,
						    (union wil_tx_desc *)d,
						    ctx);
2146 2147

			if (skb) {
2148
				if (likely(d->dma.error == 0)) {
2149 2150
					ndev->stats.tx_packets++;
					ndev->stats.tx_bytes += skb->len;
2151 2152 2153 2154
					if (stats) {
						stats->tx_packets++;
						stats->tx_bytes += skb->len;
					}
2155 2156
				} else {
					ndev->stats.tx_errors++;
2157 2158
					if (stats)
						stats->tx_errors++;
2159
				}
2160
				wil_consume_skb(skb, d->dma.error == 0);
2161 2162
			}
			memset(ctx, 0, sizeof(*ctx));
2163
			/* Make sure the ctx is zeroed before updating the tail
2164
			 * to prevent a case where wil_tx_ring will see
2165 2166 2167 2168
			 * this descriptor as used and handle it before ctx zero
			 * is completed.
			 */
			wmb();
2169 2170 2171 2172 2173
			/* There is no need to touch HW descriptor:
			 * - ststus bit TX_DMA_STATUS_DU is set by design,
			 *   so hardware will not try to process this desc.,
			 * - rest of descriptor will be initialized on Tx.
			 */
2174
			vring->swtail = wil_ring_next_tail(vring);
2175
			done++;
2176 2177
		}
	}
2178

2179
	/* performance monitoring */
2180 2181
	used_new = wil_ring_used_tx(vring);
	if (wil_val_in_range(wil->ring_idle_trsh,
2182 2183 2184
			     used_new, used_before_complete)) {
		wil_dbg_txrx(wil, "Ring[%2d] idle %d -> %d\n",
			     ringid, used_before_complete, used_new);
2185 2186
		txdata->last_idle = get_cycles();
	}
2187

D
Dedy Lansky 已提交
2188 2189
	/* shall we wake net queues? */
	if (done)
2190
		wil_update_net_queues(wil, vif, vring, false);
V
Vladimir Kondratiev 已提交
2191 2192

	return done;
2193
}
2194 2195 2196 2197 2198 2199 2200 2201

static inline int wil_tx_init(struct wil6210_priv *wil)
{
	return 0;
}

static inline void wil_tx_fini(struct wil6210_priv *wil) {}

2202 2203
static void wil_get_reorder_params(struct wil6210_priv *wil,
				   struct sk_buff *skb, int *tid, int *cid,
2204
				   int *mid, u16 *seq, int *mcast, int *retry)
2205 2206 2207 2208 2209 2210 2211 2212
{
	struct vring_rx_desc *d = wil_skb_rxdesc(skb);

	*tid = wil_rxdesc_tid(d);
	*cid = wil_rxdesc_cid(d);
	*mid = wil_rxdesc_mid(d);
	*seq = wil_rxdesc_seq(d);
	*mcast = wil_rxdesc_mcast(d);
2213
	*retry = wil_rxdesc_retry(d);
2214 2215
}

2216 2217 2218 2219 2220
void wil_init_txrx_ops_legacy_dma(struct wil6210_priv *wil)
{
	wil->txrx_ops.configure_interrupt_moderation =
		wil_configure_interrupt_moderation;
	/* TX ops */
2221 2222 2223
	wil->txrx_ops.tx_desc_map = wil_tx_desc_map;
	wil->txrx_ops.tx_desc_unmap = wil_txdesc_unmap;
	wil->txrx_ops.tx_ring_tso =  __wil_tx_vring_tso;
2224 2225 2226 2227 2228 2229 2230
	wil->txrx_ops.ring_init_tx = wil_vring_init_tx;
	wil->txrx_ops.ring_fini_tx = wil_vring_free;
	wil->txrx_ops.ring_init_bcast = wil_vring_init_bcast;
	wil->txrx_ops.tx_init = wil_tx_init;
	wil->txrx_ops.tx_fini = wil_tx_fini;
	/* RX ops */
	wil->txrx_ops.rx_init = wil_rx_init;
2231 2232 2233 2234 2235
	wil->txrx_ops.wmi_addba_rx_resp = wmi_addba_rx_resp;
	wil->txrx_ops.get_reorder_params = wil_get_reorder_params;
	wil->txrx_ops.get_netif_rx_params =
		wil_get_netif_rx_params;
	wil->txrx_ops.rx_crypto_check = wil_rx_crypto_check;
2236
	wil->txrx_ops.rx_error_check = wil_rx_error_check;
2237
	wil->txrx_ops.is_rx_idle = wil_is_rx_idle;
2238 2239
	wil->txrx_ops.rx_fini = wil_rx_fini;
}