qed.h 19.6 KB
Newer Older
1
/* QLogic qed NIC Driver
M
Mintz, Yuval 已提交
2
 * Copyright (c) 2015-2017  QLogic Corporation
3
 *
M
Mintz, Yuval 已提交
4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
 * This software is available to you under a choice of one of two
 * licenses.  You may choose to be licensed under the terms of the GNU
 * General Public License (GPL) Version 2, available from the file
 * COPYING in the main directory of this source tree, or the
 * OpenIB.org BSD license below:
 *
 *     Redistribution and use in source and binary forms, with or
 *     without modification, are permitted provided that the following
 *     conditions are met:
 *
 *      - Redistributions of source code must retain the above
 *        copyright notice, this list of conditions and the following
 *        disclaimer.
 *
 *      - Redistributions in binary form must reproduce the above
 *        copyright notice, this list of conditions and the following
 *        disclaimer in the documentation and /or other materials
 *        provided with the distribution.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 * SOFTWARE.
31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49
 */

#ifndef _QED_H
#define _QED_H

#include <linux/types.h>
#include <linux/io.h>
#include <linux/delay.h>
#include <linux/firmware.h>
#include <linux/interrupt.h>
#include <linux/list.h>
#include <linux/mutex.h>
#include <linux/pci.h>
#include <linux/slab.h>
#include <linux/string.h>
#include <linux/workqueue.h>
#include <linux/zlib.h>
#include <linux/hashtable.h>
#include <linux/qed/qed_if.h>
50
#include "qed_debug.h"
51 52
#include "qed_hsi.h"

Y
Yuval Mintz 已提交
53
extern const struct qed_common_ops qed_common_ops_pass;
54 55 56 57 58 59 60 61 62 63 64 65 66

#define QED_MAJOR_VERSION               8
#define QED_MINOR_VERSION               10
#define QED_REVISION_VERSION            10
#define QED_ENGINEERING_VERSION 21

#define QED_VERSION						 \
	((QED_MAJOR_VERSION << 24) | (QED_MINOR_VERSION << 16) | \
	 (QED_REVISION_VERSION << 8) | QED_ENGINEERING_VERSION)

#define STORM_FW_VERSION				       \
	((FW_MAJOR_VERSION << 24) | (FW_MINOR_VERSION << 16) | \
	 (FW_REVISION_VERSION << 8) | FW_ENGINEERING_VERSION)
67 68 69 70 71

#define MAX_HWFNS_PER_DEVICE    (4)
#define NAME_SIZE 16
#define VER_SIZE 16

72 73
#define QED_WFQ_UNIT	100

R
Ram Amrani 已提交
74 75 76
#define QED_WID_SIZE            (1024)
#define QED_PF_DEMS_SIZE        (4)

77 78 79 80 81 82 83 84
/* cau states */
enum qed_coalescing_mode {
	QED_COAL_MODE_DISABLE,
	QED_COAL_MODE_ENABLE
};

struct qed_eth_cb_ops;
struct qed_dev_info;
85 86
union qed_mcp_protocol_stats;
enum qed_mcp_protocol_type;
87 88

/* helpers */
89 90 91 92 93 94 95 96 97
#define QED_MFW_GET_FIELD(name, field) \
	(((name) & (field ## _MASK)) >> (field ## _SHIFT))

#define QED_MFW_SET_FIELD(name, field, value)				       \
	do {								       \
		(name)	&= ~((field ## _MASK) << (field ## _SHIFT));	       \
		(name)	|= (((value) << (field ## _SHIFT)) & (field ## _MASK));\
	} while (0)

98
static inline u32 qed_db_addr(u32 cid, u32 DEMS)
R
Ram Amrani 已提交
99 100 101 102 103 104 105 106
{
	u32 db_addr = FIELD_VALUE(DB_LEGACY_ADDR_DEMS, DEMS) |
		      (cid * QED_PF_DEMS_SIZE);

	return db_addr;
}

static inline u32 qed_db_addr_vf(u32 cid, u32 DEMS)
107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130
{
	u32 db_addr = FIELD_VALUE(DB_LEGACY_ADDR_DEMS, DEMS) |
		      FIELD_VALUE(DB_LEGACY_ADDR_ICID, cid);

	return db_addr;
}

#define ALIGNED_TYPE_SIZE(type_name, p_hwfn)				     \
	((sizeof(type_name) + (u32)(1 << (p_hwfn->cdev->cache_shift)) - 1) & \
	 ~((1 << (p_hwfn->cdev->cache_shift)) - 1))

#define for_each_hwfn(cdev, i)  for (i = 0; i < cdev->num_hwfns; i++)

#define D_TRINE(val, cond1, cond2, true1, true2, def) \
	(val == (cond1) ? true1 :		      \
	 (val == (cond2) ? true2 : def))

/* forward */
struct qed_ptt_pool;
struct qed_spq;
struct qed_sb_info;
struct qed_sb_attn_info;
struct qed_cxt_mngr;
struct qed_sb_sp_info;
Y
Yuval Mintz 已提交
131
struct qed_ll2_info;
132 133 134
struct qed_mcp_info;

struct qed_rt_data {
Y
Yuval Mintz 已提交
135 136
	u32	*init_val;
	bool	*b_valid;
137 138
};

139 140 141 142 143 144 145 146 147 148 149 150 151
enum qed_tunn_mode {
	QED_MODE_L2GENEVE_TUNN,
	QED_MODE_IPGENEVE_TUNN,
	QED_MODE_L2GRE_TUNN,
	QED_MODE_IPGRE_TUNN,
	QED_MODE_VXLAN_TUNN,
};

enum qed_tunn_clss {
	QED_TUNN_CLSS_MAC_VLAN,
	QED_TUNN_CLSS_MAC_VNI,
	QED_TUNN_CLSS_INNER_MAC_VLAN,
	QED_TUNN_CLSS_INNER_MAC_VNI,
152
	QED_TUNN_CLSS_MAC_VLAN_DUAL_STAGE,
153 154 155
	MAX_QED_TUNN_CLSS,
};

156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180
struct qed_tunn_update_type {
	bool b_update_mode;
	bool b_mode_enabled;
	enum qed_tunn_clss tun_cls;
};

struct qed_tunn_update_udp_port {
	bool b_update_port;
	u16 port;
};

struct qed_tunnel_info {
	struct qed_tunn_update_type vxlan;
	struct qed_tunn_update_type l2_geneve;
	struct qed_tunn_update_type ip_geneve;
	struct qed_tunn_update_type l2_gre;
	struct qed_tunn_update_type ip_gre;

	struct qed_tunn_update_udp_port vxlan_port;
	struct qed_tunn_update_udp_port geneve_port;

	bool b_update_rx_cls;
	bool b_update_tx_cls;
};

181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209
struct qed_tunn_start_params {
	unsigned long	tunn_mode;
	u16		vxlan_udp_port;
	u16		geneve_udp_port;
	u8		update_vxlan_udp_port;
	u8		update_geneve_udp_port;
	u8		tunn_clss_vxlan;
	u8		tunn_clss_l2geneve;
	u8		tunn_clss_ipgeneve;
	u8		tunn_clss_l2gre;
	u8		tunn_clss_ipgre;
};

struct qed_tunn_update_params {
	unsigned long	tunn_mode_update_mask;
	unsigned long	tunn_mode;
	u16		vxlan_udp_port;
	u16		geneve_udp_port;
	u8		update_rx_pf_clss;
	u8		update_tx_pf_clss;
	u8		update_vxlan_udp_port;
	u8		update_geneve_udp_port;
	u8		tunn_clss_vxlan;
	u8		tunn_clss_l2geneve;
	u8		tunn_clss_ipgeneve;
	u8		tunn_clss_l2gre;
	u8		tunn_clss_ipgre;
};

210 211 212 213 214 215
/* The PCI personality is not quite synonymous to protocol ID:
 * 1. All personalities need CORE connections
 * 2. The Ethernet personality may support also the RoCE protocol
 */
enum qed_pci_personality {
	QED_PCI_ETH,
216
	QED_PCI_FCOE,
Y
Yuval Mintz 已提交
217 218
	QED_PCI_ISCSI,
	QED_PCI_ETH_ROCE,
219 220 221 222 223 224 225 226 227 228
	QED_PCI_DEFAULT /* default in shmem */
};

/* All VFs are symmetric, all counters are PF + all VFs */
struct qed_qm_iids {
	u32 cids;
	u32 vf_cids;
	u32 tids;
};

229 230 231 232
/* HW / FW resources, output of features supported below, most information
 * is received from MFW.
 */
enum qed_resources {
233
	QED_SB,
Y
Yuval Mintz 已提交
234
	QED_L2_QUEUE,
235
	QED_VPORT,
Y
Yuval Mintz 已提交
236
	QED_RSS_ENG,
237 238
	QED_PQ,
	QED_RL,
Y
Yuval Mintz 已提交
239 240
	QED_MAC,
	QED_VLAN,
R
Ram Amrani 已提交
241
	QED_RDMA_CNQ_RAM,
242
	QED_ILT,
Y
Yuval Mintz 已提交
243
	QED_LL2_QUEUE,
244
	QED_CMDQS_CQS,
R
Ram Amrani 已提交
245
	QED_RDMA_STATS_QUEUE,
246
	QED_BDQ,
247 248 249
	QED_MAX_RESC,
};

Y
Yuval Mintz 已提交
250 251
enum QED_FEATURE {
	QED_PF_L2_QUE,
Y
Yuval Mintz 已提交
252
	QED_VF,
R
Ram Amrani 已提交
253
	QED_RDMA_CNQ,
254
	QED_ISCSI_CQ,
255
	QED_FCOE_CQ,
256
	QED_VF_L2_QUE,
Y
Yuval Mintz 已提交
257 258 259
	QED_MAX_FEATURES,
};

Y
Yuval Mintz 已提交
260 261 262 263 264 265 266 267 268
enum QED_PORT_MODE {
	QED_PORT_MODE_DE_2X40G,
	QED_PORT_MODE_DE_2X50G,
	QED_PORT_MODE_DE_1X100G,
	QED_PORT_MODE_DE_4X10G_F,
	QED_PORT_MODE_DE_4X10G_E,
	QED_PORT_MODE_DE_4X20G,
	QED_PORT_MODE_DE_1X40G,
	QED_PORT_MODE_DE_2X25G,
269 270 271
	QED_PORT_MODE_DE_1X25G,
	QED_PORT_MODE_DE_4X25G,
	QED_PORT_MODE_DE_2X10G,
Y
Yuval Mintz 已提交
272 273
};

Y
Yuval Mintz 已提交
274 275
enum qed_dev_cap {
	QED_DEV_CAP_ETH,
276
	QED_DEV_CAP_FCOE,
Y
Yuval Mintz 已提交
277 278
	QED_DEV_CAP_ISCSI,
	QED_DEV_CAP_ROCE,
Y
Yuval Mintz 已提交
279 280
};

M
Mintz, Yuval 已提交
281 282 283 284 285
enum qed_wol_support {
	QED_WOL_SUPPORT_NONE,
	QED_WOL_SUPPORT_PME,
};

286 287 288 289 290 291 292
struct qed_hw_info {
	/* PCI personality */
	enum qed_pci_personality	personality;

	/* Resource Allocation scheme results */
	u32				resc_start[QED_MAX_RESC];
	u32				resc_num[QED_MAX_RESC];
Y
Yuval Mintz 已提交
293
	u32				feat_num[QED_MAX_FEATURES];
294 295 296

#define RESC_START(_p_hwfn, resc) ((_p_hwfn)->hw_info.resc_start[resc])
#define RESC_NUM(_p_hwfn, resc) ((_p_hwfn)->hw_info.resc_num[resc])
297 298
#define RESC_END(_p_hwfn, resc) (RESC_START(_p_hwfn, resc) + \
				 RESC_NUM(_p_hwfn, resc))
299 300
#define FEAT_NUM(_p_hwfn, resc) ((_p_hwfn)->hw_info.feat_num[resc])

A
Ariel Elior 已提交
301 302 303 304 305 306 307
	/* Amount of traffic classes HW supports */
	u8 num_hw_tc;

	/* Amount of TCs which should be active according to DCBx or upper
	 * layer driver configuration.
	 */
	u8 num_active_tc;
308 309 310 311 312 313 314 315
	u8				offload_tc;

	u32				concrete_fid;
	u16				opaque_fid;
	u16				ovlan;
	u32				part_num[4];

	unsigned char			hw_mac_addr[ETH_ALEN];
316 317 318 319
	u64				node_wwn;
	u64				port_wwn;

	u16				num_fcoe_conns;
320 321 322 323 324

	struct qed_igu_info		*p_igu_info;

	u32				port_mode;
	u32				hw_mode;
Y
Yuval Mintz 已提交
325
	unsigned long		device_capabilities;
326
	u16				mtu;
M
Mintz, Yuval 已提交
327 328

	enum qed_wol_support b_wol_support;
329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358
};

/* maximun size of read/write commands (HW limit) */
#define DMAE_MAX_RW_SIZE        0x2000

struct qed_dmae_info {
	/* Mutex for synchronizing access to functions */
	struct mutex	mutex;

	u8		channel;

	dma_addr_t	completion_word_phys_addr;

	/* The memory location where the DMAE writes the completion
	 * value when an operation is finished on this context.
	 */
	u32		*p_completion_word;

	dma_addr_t	intermediate_buffer_phys_addr;

	/* An intermediate buffer for DMAE operations that use virtual
	 * addresses - data is DMA'd to/from this buffer and then
	 * memcpy'd to/from the virtual address
	 */
	u32		*p_intermediate_buffer;

	dma_addr_t	dmae_cmd_phys_addr;
	struct dmae_cmd *p_dmae_cmd;
};

359 360 361 362 363 364
struct qed_wfq_data {
	/* when feature is configured for at least 1 vport */
	u32	min_speed;
	bool	configured;
};

365 366 367 368 369 370
struct qed_qm_info {
	struct init_qm_pq_params	*qm_pq_params;
	struct init_qm_vport_params	*qm_vport_params;
	struct init_qm_port_params	*qm_port_params;
	u16				start_pq;
	u8				start_vport;
A
Ariel Elior 已提交
371 372 373 374 375 376 377 378
	u16				 pure_lb_pq;
	u16				offload_pq;
	u16				low_latency_pq;
	u16				pure_ack_pq;
	u16				ooo_pq;
	u16				first_vf_pq;
	u16				first_mcos_pq;
	u16				first_rl_pq;
379 380 381 382
	u16				num_pqs;
	u16				num_vf_pqs;
	u8				num_vports;
	u8				max_phys_tcs_per_port;
A
Ariel Elior 已提交
383
	u8				ooo_tc;
384 385 386 387 388 389
	bool				pf_rl_en;
	bool				pf_wfq_en;
	bool				vport_rl_en;
	bool				vport_wfq_en;
	u8				pf_wfq;
	u32				pf_rl;
390
	struct qed_wfq_data		*wfq_data;
391
	u8 num_pf_rls;
392 393
};

M
Manish Chopra 已提交
394 395 396 397 398 399 400 401 402 403 404 405
struct storm_stats {
	u32     address;
	u32     len;
};

struct qed_storm_stats {
	struct storm_stats mstats;
	struct storm_stats pstats;
	struct storm_stats tstats;
	struct storm_stats ustats;
};

406
struct qed_fw_data {
M
Manish Chopra 已提交
407
	struct fw_ver_info	*fw_ver_info;
408 409 410 411 412 413
	const u8		*modes_tree_buf;
	union init_op		*init_ops;
	const u32		*arr_data;
	u32			init_ops_size;
};

414 415 416 417 418 419
#define DRV_MODULE_VERSION		      \
	__stringify(QED_MAJOR_VERSION) "."    \
	__stringify(QED_MINOR_VERSION) "."    \
	__stringify(QED_REVISION_VERSION) "." \
	__stringify(QED_ENGINEERING_VERSION)

420 421 422 423 424 425 426 427 428 429 430
struct qed_simd_fp_handler {
	void	*token;
	void	(*func)(void *);
};

struct qed_hwfn {
	struct qed_dev			*cdev;
	u8				my_id;          /* ID inside the PF */
#define IS_LEAD_HWFN(edev)              (!((edev)->my_id))
	u8				rel_pf_id;      /* Relative to engine*/
	u8				abs_pf_id;
431 432
#define QED_PATH_ID(_p_hwfn) \
	(QED_IS_K2((_p_hwfn)->cdev) ? 0 : ((_p_hwfn)->abs_pf_id & 1))
433 434 435 436 437 438 439 440 441 442
	u8				port_id;
	bool				b_active;

	u32				dp_module;
	u8				dp_level;
	char				name[NAME_SIZE];

	bool				first_on_engine;
	bool				hw_init_done;

Y
Yuval Mintz 已提交
443
	u8				num_funcs_on_engine;
444
	u8 enabled_func_idx;
Y
Yuval Mintz 已提交
445

446 447 448 449 450 451 452 453 454 455 456 457 458
	/* BAR access */
	void __iomem			*regview;
	void __iomem			*doorbells;
	u64				db_phys_addr;
	unsigned long			db_size;

	/* PTT pool */
	struct qed_ptt_pool		*p_ptt_pool;

	/* HW info */
	struct qed_hw_info		hw_info;

	/* rt_array (for init-tool) */
Y
Yuval Mintz 已提交
459
	struct qed_rt_data		rt_data;
460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480

	/* SPQ */
	struct qed_spq			*p_spq;

	/* EQ */
	struct qed_eq			*p_eq;

	/* Consolidate Q*/
	struct qed_consq		*p_consq;

	/* Slow-Path definitions */
	struct tasklet_struct		*sp_dpc;
	bool				b_sp_dpc_enabled;

	struct qed_ptt			*p_main_ptt;
	struct qed_ptt			*p_dpc_ptt;

	struct qed_sb_sp_info		*p_sp_sb;
	struct qed_sb_attn_info		*p_sb_attn;

	/* Protocol related */
Y
Yuval Mintz 已提交
481 482
	bool				using_ll2;
	struct qed_ll2_info		*p_ll2_info;
483
	struct qed_ooo_info		*p_ooo_info;
R
Ram Amrani 已提交
484
	struct qed_rdma_info		*p_rdma_info;
485
	struct qed_iscsi_info		*p_iscsi_info;
486
	struct qed_fcoe_info		*p_fcoe_info;
487 488
	struct qed_pf_params		pf_params;

489 490 491
	bool b_rdma_enabled_in_prs;
	u32 rdma_prs_search_reg;

492 493 494 495 496 497 498 499
	/* Array of sb_info of all status blocks */
	struct qed_sb_info		*sbs_info[MAX_SB_PER_PF_MIMD];
	u16				num_sbs;

	struct qed_cxt_mngr		*p_cxt_mngr;

	/* Flag indicating whether interrupts are enabled or not*/
	bool				b_int_enabled;
500
	bool				b_int_requested;
501

502 503 504
	/* True if the driver requests for the link */
	bool				b_drv_link_init;

Y
Yuval Mintz 已提交
505
	struct qed_vf_iov		*vf_iov_info;
Y
Yuval Mintz 已提交
506
	struct qed_pf_iov		*pf_iov_info;
507 508
	struct qed_mcp_info		*mcp_info;

509 510
	struct qed_dcbx_info		*p_dcbx_info;

511 512 513 514
	struct qed_dmae_info		dmae_info;

	/* QM init */
	struct qed_qm_info		qm_info;
M
Manish Chopra 已提交
515
	struct qed_storm_stats		storm_stats;
516 517 518 519

	/* Buffer for unzipping firmware data */
	void				*unzip_buf;

520 521
	struct dbg_tools_data		dbg_info;

R
Ram Amrani 已提交
522 523 524 525 526 527 528 529 530 531 532
	/* PWM region specific data */
	u32				dpi_size;
	u32				dpi_count;

	/* This is used to calculate the doorbell address */
	u32 dpi_start_offset;

	/* If one of the following is set then EDPM shouldn't be used */
	u8 dcbx_no_edpm;
	u8 db_bar_no_edpm;

533 534
	struct qed_ptt *p_arfs_ptt;

535 536
	/* p_ptp_ptt is valid for leading HWFN only */
	struct qed_ptt *p_ptp_ptt;
537 538
	struct qed_simd_fp_handler	simd_proto_handler[64];

539 540 541 542 543 544
#ifdef CONFIG_QED_SRIOV
	struct workqueue_struct *iov_wq;
	struct delayed_work iov_task;
	unsigned long iov_task_flags;
#endif

545
	struct z_stream_s		*stream;
R
Ram Amrani 已提交
546
	struct qed_roce_ll2_info	*ll2;
547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570
};

struct pci_params {
	int		pm_cap;

	unsigned long	mem_start;
	unsigned long	mem_end;
	unsigned int	irq;
	u8		pf_num;
};

struct qed_int_param {
	u32	int_mode;
	u8	num_vectors;
	u8	min_msix_cnt; /* for minimal functionality */
};

struct qed_int_params {
	struct qed_int_param	in;
	struct qed_int_param	out;
	struct msix_entry	*msix_table;
	bool			fp_initialized;
	u8			fp_msix_base;
	u8			fp_msix_cnt;
R
Ram Amrani 已提交
571 572
	u8			rdma_msix_base;
	u8			rdma_msix_cnt;
573 574
};

575 576 577 578 579 580 581 582 583 584 585 586 587
struct qed_dbg_feature {
	struct dentry *dentry;
	u8 *dump_buf;
	u32 buf_size;
	u32 dumped_dwords;
};

struct qed_dbg_params {
	struct qed_dbg_feature features[DBG_FEATURE_NUM];
	u8 engine_for_debug;
	bool print_data;
};

588 589 590 591 592
struct qed_dev {
	u32	dp_module;
	u8	dp_level;
	char	name[NAME_SIZE];

593
	enum	qed_dev_type type;
Y
Yuval Mintz 已提交
594 595 596 597 598 599
/* Translate type/revision combo into the proper conditions */
#define QED_IS_BB(dev)  ((dev)->type == QED_DEV_TYPE_BB)
#define QED_IS_BB_A0(dev)       (QED_IS_BB(dev) && \
				 CHIP_REV_IS_A0(dev))
#define QED_IS_BB_B0(dev)       (QED_IS_BB(dev) && \
				 CHIP_REV_IS_B0(dev))
600 601
#define QED_IS_AH(dev)  ((dev)->type == QED_DEV_TYPE_AH)
#define QED_IS_K2(dev)  QED_IS_AH(dev)
Y
Yuval Mintz 已提交
602 603 604 605 606 607

#define QED_GET_TYPE(dev)       (QED_IS_BB_A0(dev) ? CHIP_BB_A0 : \
				 QED_IS_BB_B0(dev) ? CHIP_BB_B0 : CHIP_K2)

	u16	vendor_id;
	u16	device_id;
608 609 610
#define QED_DEV_ID_MASK		0xff00
#define QED_DEV_ID_MASK_BB	0x1600
#define QED_DEV_ID_MASK_AH	0x8000
611 612 613 614 615 616 617 618

	u16	chip_num;
#define CHIP_NUM_MASK                   0xffff
#define CHIP_NUM_SHIFT                  16

	u16	chip_rev;
#define CHIP_REV_MASK                   0xf
#define CHIP_REV_SHIFT                  12
Y
Yuval Mintz 已提交
619 620
#define CHIP_REV_IS_A0(_cdev)   (!(_cdev)->chip_rev)
#define CHIP_REV_IS_B0(_cdev)   ((_cdev)->chip_rev == 1)
621 622 623 624 625 626 627 628 629 630 631 632 633 634

	u16				chip_metal;
#define CHIP_METAL_MASK                 0xff
#define CHIP_METAL_SHIFT                4

	u16				chip_bond_id;
#define CHIP_BOND_ID_MASK               0xf
#define CHIP_BOND_ID_SHIFT              0

	u8				num_engines;
	u8				num_ports_in_engines;
	u8				num_funcs_in_port;

	u8				path_id;
Y
Yuval Mintz 已提交
635 636 637 638
	enum qed_mf_mode		mf_mode;
#define IS_MF_DEFAULT(_p_hwfn)  (((_p_hwfn)->cdev)->mf_mode == QED_MF_DEFAULT)
#define IS_MF_SI(_p_hwfn)       (((_p_hwfn)->cdev)->mf_mode == QED_MF_NPAR)
#define IS_MF_SD(_p_hwfn)       (((_p_hwfn)->cdev)->mf_mode == QED_MF_OVLAN)
639 640 641 642 643 644 645 646 647

	int				pcie_width;
	int				pcie_speed;
	u8				ver_str[VER_SIZE];

	/* Add MF related configuration */
	u8				mcp_rev;
	u8				boot_mode;

M
Mintz, Yuval 已提交
648 649 650
	/* WoL related configurations */
	u8 wol_config;
	u8 wol_mac[ETH_ALEN];
651 652 653

	u32				int_mode;
	enum qed_coalescing_mode	int_coalescing_mode;
654 655
	u16				rx_coalesce_usecs;
	u16				tx_coalesce_usecs;
656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673

	/* Start Bar offset of first hwfn */
	void __iomem			*regview;
	void __iomem			*doorbells;
	u64				db_phys_addr;
	unsigned long			db_size;

	/* PCI */
	u8				cache_shift;

	/* Init */
	const struct iro		*iro_arr;
#define IRO (p_hwfn->cdev->iro_arr)

	/* HW functions */
	u8				num_hwfns;
	struct qed_hwfn			hwfns[MAX_HWFNS_PER_DEVICE];

Y
Yuval Mintz 已提交
674 675 676
	/* SRIOV */
	struct qed_hw_sriov_info *p_iov_info;
#define IS_QED_SRIOV(cdev)              (!!(cdev)->p_iov_info)
677
	struct qed_tunnel_info		tunnel;
Y
Yuval Mintz 已提交
678
	bool				b_is_vf;
679 680 681 682 683 684 685 686 687
	u32				drv_type;
	struct qed_eth_stats		*reset_stats;
	struct qed_fw_data		*fw_data;

	u32				mcp_nvm_resp;

	/* Linux specific here */
	struct  qede_dev		*edev;
	struct  pci_dev			*pdev;
688 689
	u32 flags;
#define QED_FLAG_STORAGE_STARTED	(BIT(0))
690 691 692 693 694 695 696 697
	int				msg_enable;

	struct pci_params		pci_params;

	struct qed_int_params		int_params;

	u8				protocol;
#define IS_QED_ETH_IF(cdev)     ((cdev)->protocol == QED_PROTOCOL_ETH)
698
#define IS_QED_FCOE_IF(cdev)    ((cdev)->protocol == QED_PROTOCOL_FCOE)
699

Y
Yuval Mintz 已提交
700 701 702 703
	/* Callbacks to protocol driver */
	union {
		struct qed_common_cb_ops	*common;
		struct qed_eth_cb_ops		*eth;
704
		struct qed_fcoe_cb_ops		*fcoe;
705
		struct qed_iscsi_cb_ops		*iscsi;
Y
Yuval Mintz 已提交
706 707 708
	} protocol_ops;
	void				*ops_cookie;

709 710
	struct qed_dbg_params		dbg_params;

Y
Yuval Mintz 已提交
711 712 713 714
#ifdef CONFIG_QED_LL2
	struct qed_cb_ll2_info		*ll2;
	u8				ll2_mac_address[ETH_ALEN];
#endif
715
	DECLARE_HASHTABLE(connections, 10);
716
	const struct firmware		*firmware;
R
Ram Amrani 已提交
717 718 719 720

	u32 rdma_max_sge;
	u32 rdma_max_inline;
	u32 rdma_max_srq_sge;
721
	u16 tunn_feature_mask;
722 723
};

724 725 726 727 728 729 730 731 732 733
#define NUM_OF_VFS(dev)         (QED_IS_BB(dev) ? MAX_NUM_VFS_BB \
						: MAX_NUM_VFS_K2)
#define NUM_OF_L2_QUEUES(dev)   (QED_IS_BB(dev) ? MAX_NUM_L2_QUEUES_BB \
						: MAX_NUM_L2_QUEUES_K2)
#define NUM_OF_PORTS(dev)       (QED_IS_BB(dev) ? MAX_NUM_PORTS_BB \
						: MAX_NUM_PORTS_K2)
#define NUM_OF_SBS(dev)         (QED_IS_BB(dev) ? MAX_SB_PER_PATH_BB \
						: MAX_SB_PER_PATH_K2)
#define NUM_OF_ENG_PFS(dev)     (QED_IS_BB(dev) ? MAX_NUM_PFS_BB \
						: MAX_NUM_PFS_K2)
734 735 736 737 738 739 740 741 742 743 744 745

/**
 * @brief qed_concrete_to_sw_fid - get the sw function id from
 *        the concrete value.
 *
 * @param concrete_fid
 *
 * @return inline u8
 */
static inline u8 qed_concrete_to_sw_fid(struct qed_dev *cdev,
					u32 concrete_fid)
{
746
	u8 vfid = GET_FIELD(concrete_fid, PXP_CONCRETE_FID_VFID);
747
	u8 pfid = GET_FIELD(concrete_fid, PXP_CONCRETE_FID_PFID);
748 749 750
	u8 vf_valid = GET_FIELD(concrete_fid,
				PXP_CONCRETE_FID_VFVALID);
	u8 sw_fid;
751

752 753 754 755 756 757
	if (vf_valid)
		sw_fid = vfid + MAX_NUM_PFS;
	else
		sw_fid = pfid;

	return sw_fid;
758 759 760
}

#define PURE_LB_TC 8
761
#define OOO_LB_TC 9
762

Y
Yuval Mintz 已提交
763
int qed_configure_vport_wfq(struct qed_dev *cdev, u16 vp_id, u32 rate);
764 765 766
void qed_configure_vp_wfq_on_link_change(struct qed_dev *cdev,
					 struct qed_ptt *p_ptt,
					 u32 min_pf_rate);
767

Y
Yuval Mintz 已提交
768
void qed_clean_wfq_db(struct qed_hwfn *p_hwfn, struct qed_ptt *p_ptt);
769
int qed_device_num_engines(struct qed_dev *cdev);
770

A
Ariel Elior 已提交
771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789
#define QED_LEADING_HWFN(dev)   (&dev->hwfns[0])

/* Flags for indication of required queues */
#define PQ_FLAGS_RLS    (BIT(0))
#define PQ_FLAGS_MCOS   (BIT(1))
#define PQ_FLAGS_LB     (BIT(2))
#define PQ_FLAGS_OOO    (BIT(3))
#define PQ_FLAGS_ACK    (BIT(4))
#define PQ_FLAGS_OFLD   (BIT(5))
#define PQ_FLAGS_VFS    (BIT(6))
#define PQ_FLAGS_LLT    (BIT(7))

/* physical queue index for cm context intialization */
u16 qed_get_cm_pq_idx(struct qed_hwfn *p_hwfn, u32 pq_flags);
u16 qed_get_cm_pq_idx_mcos(struct qed_hwfn *p_hwfn, u8 tc);
u16 qed_get_cm_pq_idx_vf(struct qed_hwfn *p_hwfn, u16 vf);

#define QED_LEADING_HWFN(dev)   (&dev->hwfns[0])

790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807
/* Other Linux specific common definitions */
#define DP_NAME(cdev) ((cdev)->name)

#define REG_ADDR(cdev, offset)          (void __iomem *)((u8 __iomem *)\
						(cdev->regview) + \
							 (offset))

#define REG_RD(cdev, offset)            readl(REG_ADDR(cdev, offset))
#define REG_WR(cdev, offset, val)       writel((u32)val, REG_ADDR(cdev, offset))
#define REG_WR16(cdev, offset, val)     writew((u16)val, REG_ADDR(cdev, offset))

#define DOORBELL(cdev, db_addr, val)			 \
	writel((u32)val, (void __iomem *)((u8 __iomem *)\
					  (cdev->doorbells) + (db_addr)))

/* Prototypes */
int qed_fill_dev_info(struct qed_dev *cdev,
		      struct qed_dev_info *dev_info);
Y
Yuval Mintz 已提交
808
void qed_link_update(struct qed_hwfn *hwfn);
809 810 811
u32 qed_unzip_data(struct qed_hwfn *p_hwfn,
		   u32 input_len, u8 *input_buf,
		   u32 max_size, u8 *unzip_buf);
812 813 814
void qed_get_protocol_stats(struct qed_dev *cdev,
			    enum qed_mcp_protocol_type type,
			    union qed_mcp_protocol_stats *stats);
815
int qed_slowpath_irq_req(struct qed_hwfn *hwfn);
T
Tomer Tayar 已提交
816
void qed_slowpath_irq_sync(struct qed_hwfn *p_hwfn);
817

818
#endif /* _QED_H */