enh_desc.c 12.3 KB
Newer Older
1 2 3
/*******************************************************************************
  This contains the functions to handle the enhanced descriptors.

4
  Copyright (C) 2007-2014  STMicroelectronics Ltd
5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24

  This program is free software; you can redistribute it and/or modify it
  under the terms and conditions of the GNU General Public License,
  version 2, as published by the Free Software Foundation.

  This program is distributed in the hope it will be useful, but WITHOUT
  ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
  more details.

  You should have received a copy of the GNU General Public License along with
  this program; if not, write to the Free Software Foundation, Inc.,
  51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.

  The full GNU General Public License is included in this distribution in
  the file called "COPYING".

  Author: Giuseppe Cavallaro <peppe.cavallaro@st.com>
*******************************************************************************/

25
#include <linux/stmmac.h>
26
#include "common.h"
27
#include "descs_com.h"
28 29

static int enh_desc_get_tx_status(void *data, struct stmmac_extra_stats *x,
30
				  struct dma_desc *p, void __iomem *ioaddr)
31 32
{
	struct net_device_stats *stats = (struct net_device_stats *)data;
33
	unsigned int tdes0 = p->des0;
34 35 36 37 38 39 40 41 42
	int ret = tx_done;

	/* Get tx owner first */
	if (unlikely(tdes0 & ETDES0_OWN))
		return tx_dma_own;

	/* Verify tx error by looking at the last segment. */
	if (likely(!(tdes0 & ETDES0_LAST_SEGMENT)))
		return tx_not_ls;
43

44 45
	if (unlikely(tdes0 & ETDES0_ERROR_SUMMARY)) {
		if (unlikely(tdes0 & ETDES0_JABBER_TIMEOUT))
46 47
			x->tx_jabber++;

48
		if (unlikely(tdes0 & ETDES0_FRAME_FLUSHED)) {
49
			x->tx_frame_flushed++;
50
			dwmac_dma_flush_tx_fifo(ioaddr);
51 52
		}

53
		if (unlikely(tdes0 & ETDES0_LOSS_CARRIER)) {
54 55 56
			x->tx_losscarrier++;
			stats->tx_carrier_errors++;
		}
57
		if (unlikely(tdes0 & ETDES0_NO_CARRIER)) {
58 59 60
			x->tx_carrier++;
			stats->tx_carrier_errors++;
		}
61 62 63 64
		if (unlikely((tdes0 & ETDES0_LATE_COLLISION) ||
			     (tdes0 & ETDES0_EXCESSIVE_COLLISIONS)))
			stats->collisions +=
				(tdes0 & ETDES0_COLLISION_COUNT_MASK) >> 3;
65

66
		if (unlikely(tdes0 & ETDES0_EXCESSIVE_DEFERRAL))
67 68
			x->tx_deferred++;

69
		if (unlikely(tdes0 & ETDES0_UNDERFLOW_ERROR)) {
70
			dwmac_dma_flush_tx_fifo(ioaddr);
71 72 73
			x->tx_underflow++;
		}

74
		if (unlikely(tdes0 & ETDES0_IP_HEADER_ERROR))
75 76
			x->tx_ip_header_error++;

77
		if (unlikely(tdes0 & ETDES0_PAYLOAD_ERROR)) {
78
			x->tx_payload_error++;
79
			dwmac_dma_flush_tx_fifo(ioaddr);
80 81
		}

82
		ret = tx_err;
83 84
	}

85
	if (unlikely(tdes0 & ETDES0_DEFERRED))
86
		x->tx_deferred++;
87

88
#ifdef STMMAC_VLAN_TAG_USED
89
	if (tdes0 & ETDES0_VLAN_FRAME)
90 91 92 93 94 95 96 97
		x->tx_vlan++;
#endif

	return ret;
}

static int enh_desc_get_tx_len(struct dma_desc *p)
{
98
	return (p->des1 & ETDES1_BUFFER1_SIZE_MASK);
99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116
}

static int enh_desc_coe_rdes0(int ipc_err, int type, int payload_err)
{
	int ret = good_frame;
	u32 status = (type << 2 | ipc_err << 1 | payload_err) & 0x7;

	/* bits 5 7 0 | Frame status
	 * ----------------------------------------------------------
	 *      0 0 0 | IEEE 802.3 Type frame (length < 1536 octects)
	 *      1 0 0 | IPv4/6 No CSUM errorS.
	 *      1 0 1 | IPv4/6 CSUM PAYLOAD error
	 *      1 1 0 | IPv4/6 CSUM IP HR error
	 *      1 1 1 | IPv4/6 IP PAYLOAD AND HEADER errorS
	 *      0 0 1 | IPv4/6 unsupported IP PAYLOAD
	 *      0 1 1 | COE bypassed.. no IPv4/6 frame
	 *      0 1 0 | Reserved.
	 */
117
	if (status == 0x0)
118
		ret = llc_snap;
119
	else if (status == 0x4)
120
		ret = good_frame;
121
	else if (status == 0x5)
122
		ret = csum_none;
123
	else if (status == 0x6)
124
		ret = csum_none;
125
	else if (status == 0x7)
126
		ret = csum_none;
127
	else if (status == 0x1)
128
		ret = discard_frame;
129
	else if (status == 0x3)
130 131 132 133
		ret = discard_frame;
	return ret;
}

134 135 136
static void enh_desc_get_ext_status(void *data, struct stmmac_extra_stats *x,
				    struct dma_extended_desc *p)
{
137 138 139 140 141 142 143
	unsigned int rdes0 = p->basic.des0;
	unsigned int rdes4 = p->des4;

	if (unlikely(rdes0 & ERDES0_RX_MAC_ADDR)) {
		int message_type = (rdes4 & ERDES4_MSG_TYPE_MASK) >> 8;

		if (rdes4 & ERDES4_IP_HDR_ERR)
144
			x->ip_hdr_err++;
145
		if (rdes4 & ERDES4_IP_PAYLOAD_ERR)
146
			x->ip_payload_err++;
147
		if (rdes4 & ERDES4_IP_CSUM_BYPASSED)
148
			x->ip_csum_bypassed++;
149
		if (rdes4 & ERDES4_IPV4_PKT_RCVD)
150
			x->ipv4_pkt_rcvd++;
151
		if (rdes4 & ERDES4_IPV6_PKT_RCVD)
152
			x->ipv6_pkt_rcvd++;
153 154 155 156 157

		if (message_type == RDES_EXT_NO_PTP)
			x->no_ptp_rx_msg_type_ext++;
		else if (message_type == RDES_EXT_SYNC)
			x->ptp_rx_msg_type_sync++;
158
		else if (message_type == RDES_EXT_FOLLOW_UP)
159
			x->ptp_rx_msg_type_follow_up++;
160
		else if (message_type == RDES_EXT_DELAY_REQ)
161
			x->ptp_rx_msg_type_delay_req++;
162
		else if (message_type == RDES_EXT_DELAY_RESP)
163
			x->ptp_rx_msg_type_delay_resp++;
164
		else if (message_type == RDES_EXT_PDELAY_REQ)
165
			x->ptp_rx_msg_type_pdelay_req++;
166
		else if (message_type == RDES_EXT_PDELAY_RESP)
167
			x->ptp_rx_msg_type_pdelay_resp++;
168
		else if (message_type == RDES_EXT_PDELAY_FOLLOW_UP)
169 170 171 172 173 174 175 176
			x->ptp_rx_msg_type_pdelay_follow_up++;
		else if (message_type == RDES_PTP_ANNOUNCE)
			x->ptp_rx_msg_type_announce++;
		else if (message_type == RDES_PTP_MANAGEMENT)
			x->ptp_rx_msg_type_management++;
		else if (message_type == RDES_PTP_PKT_RESERVED_TYPE)
			x->ptp_rx_msg_pkt_reserved_type++;

177
		if (rdes4 & ERDES4_PTP_FRAME_TYPE)
178
			x->ptp_frame_type++;
179
		if (rdes4 & ERDES4_PTP_VER)
180
			x->ptp_ver++;
181
		if (rdes4 & ERDES4_TIMESTAMP_DROPPED)
182
			x->timestamp_dropped++;
183
		if (rdes4 & ERDES4_AV_PKT_RCVD)
184
			x->av_pkt_rcvd++;
185
		if (rdes4 & ERDES4_AV_TAGGED_PKT_RCVD)
186
			x->av_tagged_pkt_rcvd++;
187
		if ((rdes4 & ERDES4_VLAN_TAG_PRI_VAL_MASK) >> 18)
188
			x->vlan_tag_priority_val++;
189
		if (rdes4 & ERDES4_L3_FILTER_MATCH)
190
			x->l3_filter_match++;
191
		if (rdes4 & ERDES4_L4_FILTER_MATCH)
192
			x->l4_filter_match++;
193
		if ((rdes4 & ERDES4_L3_L4_FILT_NO_MATCH_MASK) >> 26)
194 195 196 197
			x->l3_l4_filter_no_match++;
	}
}

198 199 200 201
static int enh_desc_get_rx_status(void *data, struct stmmac_extra_stats *x,
				  struct dma_desc *p)
{
	struct net_device_stats *stats = (struct net_device_stats *)data;
202 203
	unsigned int rdes0 = p->des0;
	int ret = good_frame;
204

205 206 207
	if (unlikely(rdes0 & RDES0_OWN))
		return dma_own;

208 209
	if (unlikely(rdes0 & RDES0_ERROR_SUMMARY)) {
		if (unlikely(rdes0 & RDES0_DESCRIPTOR_ERROR)) {
210 211 212
			x->rx_desc++;
			stats->rx_length_errors++;
		}
213
		if (unlikely(rdes0 & RDES0_OVERFLOW_ERROR))
214 215
			x->rx_gmac_overflow++;

216
		if (unlikely(rdes0 & RDES0_IPC_CSUM_ERROR))
217
			pr_err("\tIPC Csum Error/Giant frame\n");
218

219
		if (unlikely(rdes0 & RDES0_COLLISION))
220
			stats->collisions++;
221
		if (unlikely(rdes0 & RDES0_RECEIVE_WATCHDOG))
222
			x->rx_watchdog++;
223

224
		if (unlikely(rdes0 & RDES0_MII_ERROR))	/* GMII */
225
			x->rx_mii++;
226

227
		if (unlikely(rdes0 & RDES0_CRC_ERROR)) {
228 229 230 231 232 233 234 235 236 237
			x->rx_crc++;
			stats->rx_crc_errors++;
		}
		ret = discard_frame;
	}

	/* After a payload csum error, the ES bit is set.
	 * It doesn't match with the information reported into the databook.
	 * At any rate, we need to understand if the CSUM hw computation is ok
	 * and report this info to the upper layers. */
238 239 240
	ret = enh_desc_coe_rdes0(!!(rdes0 & RDES0_IPC_CSUM_ERROR),
				 !!(rdes0 & RDES0_FRAME_TYPE),
				 !!(rdes0 & ERDES0_RX_MAC_ADDR));
241

242
	if (unlikely(rdes0 & RDES0_DRIBBLING))
243
		x->dribbling_bit++;
244

245
	if (unlikely(rdes0 & RDES0_SA_FILTER_FAIL)) {
246 247 248
		x->sa_rx_filter_fail++;
		ret = discard_frame;
	}
249
	if (unlikely(rdes0 & RDES0_DA_FILTER_FAIL)) {
250 251 252
		x->da_rx_filter_fail++;
		ret = discard_frame;
	}
253
	if (unlikely(rdes0 & RDES0_LENGTH_ERROR)) {
254 255 256 257
		x->rx_length++;
		ret = discard_frame;
	}
#ifdef STMMAC_VLAN_TAG_USED
258
	if (rdes0 & RDES0_VLAN_TAG)
259 260
		x->rx_vlan++;
#endif
261

262 263 264
	return ret;
}

265 266
static void enh_desc_init_rx_desc(struct dma_desc *p, int disable_rx_ic,
				  int mode, int end)
267
{
268 269
	p->des0 |= RDES0_OWN;
	p->des1 |= ((BUF_SIZE_8KiB - 1) & ERDES1_BUFFER1_SIZE_MASK);
270

271
	if (mode == STMMAC_CHAIN_MODE)
272
		ehn_desc_rx_set_on_chain(p);
273 274
	else
		ehn_desc_rx_set_on_ring(p, end);
275

276
	if (disable_rx_ic)
277
		p->des1 |= ERDES1_DISABLE_IC;
278 279
}

280
static void enh_desc_init_tx_desc(struct dma_desc *p, int mode, int end)
281
{
282
	p->des0 &= ~ETDES0_OWN;
283
	if (mode == STMMAC_CHAIN_MODE)
284
		enh_desc_end_tx_desc_on_chain(p);
285
	else
286
		enh_desc_end_tx_desc_on_ring(p, end);
287 288 289 290
}

static int enh_desc_get_tx_owner(struct dma_desc *p)
{
291
	return (p->des0 & ETDES0_OWN) >> 31;
292 293 294 295
}

static void enh_desc_set_tx_owner(struct dma_desc *p)
{
296
	p->des0 |= ETDES0_OWN;
297 298 299 300
}

static void enh_desc_set_rx_owner(struct dma_desc *p)
{
301
	p->des0 |= RDES0_OWN;
302 303 304 305
}

static int enh_desc_get_tx_ls(struct dma_desc *p)
{
306
	return (p->des0 & ETDES0_LAST_SEGMENT) >> 29;
307 308
}

309
static void enh_desc_release_tx_desc(struct dma_desc *p, int mode)
310
{
311
	int ter = (p->des0 & ETDES0_END_RING) >> 21;
312

313
	memset(p, 0, offsetof(struct dma_desc, des2));
314
	if (mode == STMMAC_CHAIN_MODE)
315
		enh_desc_end_tx_desc_on_chain(p);
316 317
	else
		enh_desc_end_tx_desc_on_ring(p, ter);
318 319 320
}

static void enh_desc_prepare_tx_desc(struct dma_desc *p, int is_fs, int len,
321
				     bool csum_flag, int mode, bool tx_own,
322
				     bool ls)
323
{
324 325
	unsigned int tdes0 = p->des0;

326 327 328 329 330
	if (mode == STMMAC_CHAIN_MODE)
		enh_set_tx_desc_len_on_chain(p, len);
	else
		enh_set_tx_desc_len_on_ring(p, len);

331 332 333 334 335 336 337 338 339 340
	if (is_fs)
		tdes0 |= ETDES0_FIRST_SEGMENT;
	else
		tdes0 &= ~ETDES0_FIRST_SEGMENT;

	if (likely(csum_flag))
		tdes0 |= (TX_CIC_FULL << ETDES0_CHECKSUM_INSERTION_SHIFT);
	else
		tdes0 &= ~(TX_CIC_FULL << ETDES0_CHECKSUM_INSERTION_SHIFT);

341 342 343 344
	if (ls)
		tdes0 |= ETDES0_LAST_SEGMENT;

	/* Finally set the OWN bit. Later the DMA will start! */
345 346 347 348 349 350 351 352 353 354
	if (tx_own)
		tdes0 |= ETDES0_OWN;

	if (is_fs & tx_own)
		/* When the own bit, for the first frame, has to be set, all
		 * descriptors for the same frame has to be set before, to
		 * avoid race condition.
		 */
		wmb();

355
	p->des0 = tdes0;
356 357
}

358
static void enh_desc_set_tx_ic(struct dma_desc *p)
359
{
360
	p->des0 |= ETDES0_INTERRUPT;
361 362
}

363
static int enh_desc_get_rx_frame_len(struct dma_desc *p, int rx_coe_type)
364
{
365
	unsigned int csum = 0;
366 367 368 369
	/* The type-1 checksum offload engines append the checksum at
	 * the end of frame and the two bytes of checksum are added in
	 * the length.
	 * Adjust for that in the framelen for type-1 checksum offload
370 371
	 * engines.
	 */
372
	if (rx_coe_type == STMMAC_RX_COE_TYPE1)
373 374 375 376
		csum = 2;

	return (((p->des0 & RDES0_FRAME_LEN_MASK) >> RDES0_FRAME_LEN_SHIFT) -
		csum);
377 378
}

379 380
static void enh_desc_enable_tx_timestamp(struct dma_desc *p)
{
381
	p->des0 |= ETDES0_TIME_STAMP_ENABLE;
382 383 384 385
}

static int enh_desc_get_tx_timestamp_status(struct dma_desc *p)
{
386
	return (p->des0 & ETDES0_TIME_STAMP_STATUS) >> 17;
387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410
}

static u64 enh_desc_get_timestamp(void *desc, u32 ats)
{
	u64 ns;

	if (ats) {
		struct dma_extended_desc *p = (struct dma_extended_desc *)desc;
		ns = p->des6;
		/* convert high/sec time stamp value to nanosecond */
		ns += p->des7 * 1000000000ULL;
	} else {
		struct dma_desc *p = (struct dma_desc *)desc;
		ns = p->des2;
		ns += p->des3 * 1000000000ULL;
	}

	return ns;
}

static int enh_desc_get_rx_timestamp_status(void *desc, u32 ats)
{
	if (ats) {
		struct dma_extended_desc *p = (struct dma_extended_desc *)desc;
411
		return (p->basic.des0 & RDES0_IPC_CSUM_ERROR) >> 7;
412 413 414 415 416 417 418 419 420 421
	} else {
		struct dma_desc *p = (struct dma_desc *)desc;
		if ((p->des2 == 0xffffffff) && (p->des3 == 0xffffffff))
			/* timestamp is corrupted, hence don't store it */
			return 0;
		else
			return 1;
	}
}

422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441
static void enh_desc_display_ring(void *head, unsigned int size, bool rx)
{
	struct dma_extended_desc *ep = (struct dma_extended_desc *)head;
	int i;

	pr_info("Extended %s descriptor ring:\n", rx ? "RX" : "TX");

	for (i = 0; i < size; i++) {
		u64 x;

		x = *(u64 *)ep;
		pr_info("%d [0x%x]: 0x%x 0x%x 0x%x 0x%x\n",
			i, (unsigned int)virt_to_phys(ep),
			(unsigned int)x, (unsigned int)(x >> 32),
			ep->basic.des2, ep->basic.des3);
		ep++;
	}
	pr_info("\n");
}

442
const struct stmmac_desc_ops enh_desc_ops = {
443 444 445 446 447 448 449 450
	.tx_status = enh_desc_get_tx_status,
	.rx_status = enh_desc_get_rx_status,
	.get_tx_len = enh_desc_get_tx_len,
	.init_rx_desc = enh_desc_init_rx_desc,
	.init_tx_desc = enh_desc_init_tx_desc,
	.get_tx_owner = enh_desc_get_tx_owner,
	.release_tx_desc = enh_desc_release_tx_desc,
	.prepare_tx_desc = enh_desc_prepare_tx_desc,
451
	.set_tx_ic = enh_desc_set_tx_ic,
452 453 454 455
	.get_tx_ls = enh_desc_get_tx_ls,
	.set_tx_owner = enh_desc_set_tx_owner,
	.set_rx_owner = enh_desc_set_rx_owner,
	.get_rx_frame_len = enh_desc_get_rx_frame_len,
456
	.rx_extended_status = enh_desc_get_ext_status,
457 458 459 460
	.enable_tx_timestamp = enh_desc_enable_tx_timestamp,
	.get_tx_timestamp_status = enh_desc_get_tx_timestamp_status,
	.get_timestamp = enh_desc_get_timestamp,
	.get_rx_timestamp_status = enh_desc_get_rx_timestamp_status,
461
	.display_ring = enh_desc_display_ring,
462
};