/* bnx2x_vfpf.c: Broadcom Everest network driver. * * Copyright 2009-2013 Broadcom Corporation * * Unless you and Broadcom execute a separate written software license * agreement governing use of this software, this software is licensed to you * under the terms of the GNU General Public License version 2, available * at http://www.gnu.org/licenses/old-licenses/gpl-2.0.html (the "GPL"). * * Notwithstanding the above, under no circumstances may you combine this * software in any way with any other Broadcom software provided under a * license other than the GPL, without Broadcom's express prior written * consent. * * Maintained by: Eilon Greenstein * Written by: Shmulik Ravid * Ariel Elior */ #include "bnx2x.h" #include "bnx2x_cmn.h" #include /* place a given tlv on the tlv buffer at a given offset */ void bnx2x_add_tlv(struct bnx2x *bp, void *tlvs_list, u16 offset, u16 type, u16 length) { struct channel_tlv *tl = (struct channel_tlv *)(tlvs_list + offset); tl->type = type; tl->length = length; } /* Clear the mailbox and init the header of the first tlv */ void bnx2x_vfpf_prep(struct bnx2x *bp, struct vfpf_first_tlv *first_tlv, u16 type, u16 length) { mutex_lock(&bp->vf2pf_mutex); DP(BNX2X_MSG_IOV, "preparing to send %d tlv over vf pf channel\n", type); /* Clear mailbox */ memset(bp->vf2pf_mbox, 0, sizeof(struct bnx2x_vf_mbx_msg)); /* init type and length */ bnx2x_add_tlv(bp, &first_tlv->tl, 0, type, length); /* init first tlv header */ first_tlv->resp_msg_offset = sizeof(bp->vf2pf_mbox->req); } /* releases the mailbox */ void bnx2x_vfpf_finalize(struct bnx2x *bp, struct vfpf_first_tlv *first_tlv) { DP(BNX2X_MSG_IOV, "done sending [%d] tlv over vf pf channel\n", first_tlv->tl.type); mutex_unlock(&bp->vf2pf_mutex); } /* list the types and lengths of the tlvs on the buffer */ void bnx2x_dp_tlv_list(struct bnx2x *bp, void *tlvs_list) { int i = 1; struct channel_tlv *tlv = (struct channel_tlv *)tlvs_list; while (tlv->type != CHANNEL_TLV_LIST_END) { /* output tlv */ DP(BNX2X_MSG_IOV, "TLV number %d: type %d, length %d\n", i, tlv->type, tlv->length); /* advance to next tlv */ tlvs_list += tlv->length; /* cast general tlv list pointer to channel tlv header*/ tlv = (struct channel_tlv *)tlvs_list; i++; /* break condition for this loop */ if (i > MAX_TLVS_IN_LIST) { WARN(true, "corrupt tlvs"); return; } } /* output last tlv */ DP(BNX2X_MSG_IOV, "TLV number %d: type %d, length %d\n", i, tlv->type, tlv->length); } /* test whether we support a tlv type */ bool bnx2x_tlv_supported(u16 tlvtype) { return CHANNEL_TLV_NONE < tlvtype && tlvtype < CHANNEL_TLV_MAX; } static inline int bnx2x_pfvf_status_codes(int rc) { switch (rc) { case 0: return PFVF_STATUS_SUCCESS; case -ENOMEM: return PFVF_STATUS_NO_RESOURCE; default: return PFVF_STATUS_FAILURE; } } static int bnx2x_send_msg2pf(struct bnx2x *bp, u8 *done, dma_addr_t msg_mapping) { struct cstorm_vf_zone_data __iomem *zone_data = REG_ADDR(bp, PXP_VF_ADDR_CSDM_GLOBAL_START); int tout = 600, interval = 100; /* wait for 60 seconds */ if (*done) { BNX2X_ERR("done was non zero before message to pf was sent\n"); WARN_ON(true); return -EINVAL; } /* Write message address */ writel(U64_LO(msg_mapping), &zone_data->non_trigger.vf_pf_channel.msg_addr_lo); writel(U64_HI(msg_mapping), &zone_data->non_trigger.vf_pf_channel.msg_addr_hi); /* make sure the address is written before FW accesses it */ wmb(); /* Trigger the PF FW */ writeb(1, &zone_data->trigger.vf_pf_channel.addr_valid); /* Wait for PF to complete */ while ((tout >= 0) && (!*done)) { msleep(interval); tout -= 1; /* progress indicator - HV can take its own sweet time in * answering VFs... */ DP_CONT(BNX2X_MSG_IOV, "."); } if (!*done) { BNX2X_ERR("PF response has timed out\n"); return -EAGAIN; } DP(BNX2X_MSG_SP, "Got a response from PF\n"); return 0; } static int bnx2x_get_vf_id(struct bnx2x *bp, u32 *vf_id) { u32 me_reg; int tout = 10, interval = 100; /* Wait for 1 sec */ do { /* pxp traps vf read of doorbells and returns me reg value */ me_reg = readl(bp->doorbells); if (GOOD_ME_REG(me_reg)) break; msleep(interval); BNX2X_ERR("Invalid ME register value: 0x%08x\n. Is pf driver up?", me_reg); } while (tout-- > 0); if (!GOOD_ME_REG(me_reg)) { BNX2X_ERR("Invalid ME register value: 0x%08x\n", me_reg); return -EINVAL; } BNX2X_ERR("valid ME register value: 0x%08x\n", me_reg); *vf_id = (me_reg & ME_REG_VF_NUM_MASK) >> ME_REG_VF_NUM_SHIFT; return 0; } int bnx2x_vfpf_acquire(struct bnx2x *bp, u8 tx_count, u8 rx_count) { int rc = 0, attempts = 0; struct vfpf_acquire_tlv *req = &bp->vf2pf_mbox->req.acquire; struct pfvf_acquire_resp_tlv *resp = &bp->vf2pf_mbox->resp.acquire_resp; u32 vf_id; bool resources_acquired = false; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_ACQUIRE, sizeof(*req)); if (bnx2x_get_vf_id(bp, &vf_id)) { rc = -EAGAIN; goto out; } req->vfdev_info.vf_id = vf_id; req->vfdev_info.vf_os = 0; req->resc_request.num_rxqs = rx_count; req->resc_request.num_txqs = tx_count; req->resc_request.num_sbs = bp->igu_sb_cnt; req->resc_request.num_mac_filters = VF_ACQUIRE_MAC_FILTERS; req->resc_request.num_mc_filters = VF_ACQUIRE_MC_FILTERS; /* pf 2 vf bulletin board address */ req->bulletin_addr = bp->pf2vf_bulletin_mapping; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); while (!resources_acquired) { DP(BNX2X_MSG_SP, "attempting to acquire resources\n"); /* send acquire request */ rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); /* PF timeout */ if (rc) goto out; /* copy acquire response from buffer to bp */ memcpy(&bp->acquire_resp, resp, sizeof(bp->acquire_resp)); attempts++; /* test whether the PF accepted our request. If not, humble the * the request and try again. */ if (bp->acquire_resp.hdr.status == PFVF_STATUS_SUCCESS) { DP(BNX2X_MSG_SP, "resources acquired\n"); resources_acquired = true; } else if (bp->acquire_resp.hdr.status == PFVF_STATUS_NO_RESOURCE && attempts < VF_ACQUIRE_THRESH) { DP(BNX2X_MSG_SP, "PF unwilling to fulfill resource request. Try PF recommended amount\n"); /* humble our request */ req->resc_request.num_txqs = bp->acquire_resp.resc.num_txqs; req->resc_request.num_rxqs = bp->acquire_resp.resc.num_rxqs; req->resc_request.num_sbs = bp->acquire_resp.resc.num_sbs; req->resc_request.num_mac_filters = bp->acquire_resp.resc.num_mac_filters; req->resc_request.num_vlan_filters = bp->acquire_resp.resc.num_vlan_filters; req->resc_request.num_mc_filters = bp->acquire_resp.resc.num_mc_filters; /* Clear response buffer */ memset(&bp->vf2pf_mbox->resp, 0, sizeof(union pfvf_tlvs)); } else { /* PF reports error */ BNX2X_ERR("Failed to get the requested amount of resources: %d. Breaking...\n", bp->acquire_resp.hdr.status); rc = -EAGAIN; goto out; } } /* get HW info */ bp->common.chip_id |= (bp->acquire_resp.pfdev_info.chip_num & 0xffff); bp->link_params.chip_id = bp->common.chip_id; bp->db_size = bp->acquire_resp.pfdev_info.db_size; bp->common.int_block = INT_BLOCK_IGU; bp->common.chip_port_mode = CHIP_2_PORT_MODE; bp->igu_dsb_id = -1; bp->mf_ov = 0; bp->mf_mode = 0; bp->common.flash_size = 0; bp->flags |= NO_WOL_FLAG | NO_ISCSI_OOO_FLAG | NO_ISCSI_FLAG | NO_FCOE_FLAG; bp->igu_sb_cnt = 1; bp->igu_base_sb = bp->acquire_resp.resc.hw_sbs[0].hw_sb_id; strlcpy(bp->fw_ver, bp->acquire_resp.pfdev_info.fw_ver, sizeof(bp->fw_ver)); if (is_valid_ether_addr(bp->acquire_resp.resc.current_mac_addr)) memcpy(bp->dev->dev_addr, bp->acquire_resp.resc.current_mac_addr, ETH_ALEN); out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return rc; } int bnx2x_vfpf_release(struct bnx2x *bp) { struct vfpf_release_tlv *req = &bp->vf2pf_mbox->req.release; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; u32 rc, vf_id; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_RELEASE, sizeof(*req)); if (bnx2x_get_vf_id(bp, &vf_id)) { rc = -EAGAIN; goto out; } req->vf_id = vf_id; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); /* send release request */ rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) /* PF timeout */ goto out; if (resp->hdr.status == PFVF_STATUS_SUCCESS) { /* PF released us */ DP(BNX2X_MSG_SP, "vf released\n"); } else { /* PF reports error */ BNX2X_ERR("PF failed our release request - are we out of sync? response status: %d\n", resp->hdr.status); rc = -EAGAIN; goto out; } out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return rc; } /* Tell PF about SB addresses */ int bnx2x_vfpf_init(struct bnx2x *bp) { struct vfpf_init_tlv *req = &bp->vf2pf_mbox->req.init; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; int rc, i; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_INIT, sizeof(*req)); /* status blocks */ for_each_eth_queue(bp, i) req->sb_addr[i] = (dma_addr_t)bnx2x_fp(bp, i, status_blk_mapping); /* statistics - requests only supports single queue for now */ req->stats_addr = bp->fw_stats_data_mapping + offsetof(struct bnx2x_fw_stats_data, queue_stats); /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) goto out; if (resp->hdr.status != PFVF_STATUS_SUCCESS) { BNX2X_ERR("INIT VF failed: %d. Breaking...\n", resp->hdr.status); rc = -EAGAIN; goto out; } DP(BNX2X_MSG_SP, "INIT VF Succeeded\n"); out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return rc; } /* CLOSE VF - opposite to INIT_VF */ void bnx2x_vfpf_close_vf(struct bnx2x *bp) { struct vfpf_close_tlv *req = &bp->vf2pf_mbox->req.close; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; int i, rc; u32 vf_id; /* If we haven't got a valid VF id, there is no sense to * continue with sending messages */ if (bnx2x_get_vf_id(bp, &vf_id)) goto free_irq; /* Close the queues */ for_each_queue(bp, i) bnx2x_vfpf_teardown_queue(bp, i); /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_CLOSE, sizeof(*req)); req->vf_id = vf_id; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) BNX2X_ERR("Sending CLOSE failed. rc was: %d\n", rc); else if (resp->hdr.status != PFVF_STATUS_SUCCESS) BNX2X_ERR("Sending CLOSE failed: pf response was %d\n", resp->hdr.status); bnx2x_vfpf_finalize(bp, &req->first_tlv); free_irq: /* Disable HW interrupts, NAPI */ bnx2x_netif_stop(bp, 0); /* Delete all NAPI objects */ bnx2x_del_all_napi(bp); /* Release IRQs */ bnx2x_free_irq(bp); } /* ask the pf to open a queue for the vf */ int bnx2x_vfpf_setup_q(struct bnx2x *bp, int fp_idx) { struct vfpf_setup_q_tlv *req = &bp->vf2pf_mbox->req.setup_q; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; struct bnx2x_fastpath *fp = &bp->fp[fp_idx]; u16 tpa_agg_size = 0, flags = 0; int rc; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_SETUP_Q, sizeof(*req)); /* select tpa mode to request */ if (!fp->disable_tpa) { flags |= VFPF_QUEUE_FLG_TPA; flags |= VFPF_QUEUE_FLG_TPA_IPV6; if (fp->mode == TPA_MODE_GRO) flags |= VFPF_QUEUE_FLG_TPA_GRO; tpa_agg_size = TPA_AGG_SIZE; } /* calculate queue flags */ flags |= VFPF_QUEUE_FLG_STATS; flags |= VFPF_QUEUE_FLG_CACHE_ALIGN; flags |= VFPF_QUEUE_FLG_VLAN; DP(NETIF_MSG_IFUP, "vlan removal enabled\n"); /* Common */ req->vf_qid = fp_idx; req->param_valid = VFPF_RXQ_VALID | VFPF_TXQ_VALID; /* Rx */ req->rxq.rcq_addr = fp->rx_comp_mapping; req->rxq.rcq_np_addr = fp->rx_comp_mapping + BCM_PAGE_SIZE; req->rxq.rxq_addr = fp->rx_desc_mapping; req->rxq.sge_addr = fp->rx_sge_mapping; req->rxq.vf_sb = fp_idx; req->rxq.sb_index = HC_INDEX_ETH_RX_CQ_CONS; req->rxq.hc_rate = bp->rx_ticks ? 1000000/bp->rx_ticks : 0; req->rxq.mtu = bp->dev->mtu; req->rxq.buf_sz = fp->rx_buf_size; req->rxq.sge_buf_sz = BCM_PAGE_SIZE * PAGES_PER_SGE; req->rxq.tpa_agg_sz = tpa_agg_size; req->rxq.max_sge_pkt = SGE_PAGE_ALIGN(bp->dev->mtu) >> SGE_PAGE_SHIFT; req->rxq.max_sge_pkt = ((req->rxq.max_sge_pkt + PAGES_PER_SGE - 1) & (~(PAGES_PER_SGE-1))) >> PAGES_PER_SGE_SHIFT; req->rxq.flags = flags; req->rxq.drop_flags = 0; req->rxq.cache_line_log = BNX2X_RX_ALIGN_SHIFT; req->rxq.stat_id = -1; /* No stats at the moment */ /* Tx */ req->txq.txq_addr = fp->txdata_ptr[FIRST_TX_COS_INDEX]->tx_desc_mapping; req->txq.vf_sb = fp_idx; req->txq.sb_index = HC_INDEX_ETH_TX_CQ_CONS_COS0; req->txq.hc_rate = bp->tx_ticks ? 1000000/bp->tx_ticks : 0; req->txq.flags = flags; req->txq.traffic_type = LLFC_TRAFFIC_TYPE_NW; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) BNX2X_ERR("Sending SETUP_Q message for queue[%d] failed!\n", fp_idx); if (resp->hdr.status != PFVF_STATUS_SUCCESS) { BNX2X_ERR("Status of SETUP_Q for queue[%d] is %d\n", fp_idx, resp->hdr.status); rc = -EINVAL; } bnx2x_vfpf_finalize(bp, &req->first_tlv); return rc; } int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx) { struct vfpf_q_op_tlv *req = &bp->vf2pf_mbox->req.q_op; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; int rc; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_TEARDOWN_Q, sizeof(*req)); req->vf_qid = qidx; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) { BNX2X_ERR("Sending TEARDOWN for queue %d failed: %d\n", qidx, rc); goto out; } /* PF failed the transaction */ if (resp->hdr.status != PFVF_STATUS_SUCCESS) { BNX2X_ERR("TEARDOWN for queue %d failed: %d\n", qidx, resp->hdr.status); rc = -EINVAL; } out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return rc; } /* request pf to add a mac for the vf */ int bnx2x_vfpf_set_mac(struct bnx2x *bp) { struct vfpf_set_q_filters_tlv *req = &bp->vf2pf_mbox->req.set_q_filters; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; int rc = 0; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_SET_Q_FILTERS, sizeof(*req)); req->flags = VFPF_SET_Q_FILTERS_MAC_VLAN_CHANGED; req->vf_qid = 0; req->n_mac_vlan_filters = 1; req->filters[0].flags = VFPF_Q_FILTER_DEST_MAC_VALID | VFPF_Q_FILTER_SET_MAC; /* sample bulletin board for new mac */ bnx2x_sample_bulletin(bp); /* copy mac from device to request */ memcpy(req->filters[0].mac, bp->dev->dev_addr, ETH_ALEN); /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); /* send message to pf */ rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) { BNX2X_ERR("failed to send message to pf. rc was %d\n", rc); goto out; } /* failure may mean PF was configured with a new mac for us */ while (resp->hdr.status == PFVF_STATUS_FAILURE) { DP(BNX2X_MSG_IOV, "vfpf SET MAC failed. Check bulletin board for new posts\n"); /* check if bulletin board was updated */ if (bnx2x_sample_bulletin(bp) == PFVF_BULLETIN_UPDATED) { /* copy mac from device to request */ memcpy(req->filters[0].mac, bp->dev->dev_addr, ETH_ALEN); /* send message to pf */ rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); } else { /* no new info in bulletin */ break; } } if (resp->hdr.status != PFVF_STATUS_SUCCESS) { BNX2X_ERR("vfpf SET MAC failed: %d\n", resp->hdr.status); rc = -EINVAL; } out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return 0; } int bnx2x_vfpf_set_mcast(struct net_device *dev) { struct bnx2x *bp = netdev_priv(dev); struct vfpf_set_q_filters_tlv *req = &bp->vf2pf_mbox->req.set_q_filters; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; int rc, i = 0; struct netdev_hw_addr *ha; if (bp->state != BNX2X_STATE_OPEN) { DP(NETIF_MSG_IFUP, "state is %x, returning\n", bp->state); return -EINVAL; } /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_SET_Q_FILTERS, sizeof(*req)); /* Get Rx mode requested */ DP(NETIF_MSG_IFUP, "dev->flags = %x\n", dev->flags); netdev_for_each_mc_addr(ha, dev) { DP(NETIF_MSG_IFUP, "Adding mcast MAC: %pM\n", bnx2x_mc_addr(ha)); memcpy(req->multicast[i], bnx2x_mc_addr(ha), ETH_ALEN); i++; } /* We support four PFVF_MAX_MULTICAST_PER_VF mcast * addresses tops */ if (i >= PFVF_MAX_MULTICAST_PER_VF) { DP(NETIF_MSG_IFUP, "VF supports not more than %d multicast MAC addresses\n", PFVF_MAX_MULTICAST_PER_VF); return -EINVAL; } req->n_multicast = i; req->flags |= VFPF_SET_Q_FILTERS_MULTICAST_CHANGED; req->vf_qid = 0; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) { BNX2X_ERR("Sending a message failed: %d\n", rc); goto out; } if (resp->hdr.status != PFVF_STATUS_SUCCESS) { BNX2X_ERR("Set Rx mode/multicast failed: %d\n", resp->hdr.status); rc = -EINVAL; } out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return 0; } int bnx2x_vfpf_storm_rx_mode(struct bnx2x *bp) { int mode = bp->rx_mode; struct vfpf_set_q_filters_tlv *req = &bp->vf2pf_mbox->req.set_q_filters; struct pfvf_general_resp_tlv *resp = &bp->vf2pf_mbox->resp.general_resp; int rc; /* clear mailbox and prep first tlv */ bnx2x_vfpf_prep(bp, &req->first_tlv, CHANNEL_TLV_SET_Q_FILTERS, sizeof(*req)); DP(NETIF_MSG_IFUP, "Rx mode is %d\n", mode); switch (mode) { case BNX2X_RX_MODE_NONE: /* no Rx */ req->rx_mask = VFPF_RX_MASK_ACCEPT_NONE; break; case BNX2X_RX_MODE_NORMAL: req->rx_mask = VFPF_RX_MASK_ACCEPT_MATCHED_MULTICAST; req->rx_mask |= VFPF_RX_MASK_ACCEPT_MATCHED_UNICAST; req->rx_mask |= VFPF_RX_MASK_ACCEPT_BROADCAST; break; case BNX2X_RX_MODE_ALLMULTI: req->rx_mask = VFPF_RX_MASK_ACCEPT_ALL_MULTICAST; req->rx_mask |= VFPF_RX_MASK_ACCEPT_MATCHED_UNICAST; req->rx_mask |= VFPF_RX_MASK_ACCEPT_BROADCAST; break; case BNX2X_RX_MODE_PROMISC: req->rx_mask = VFPF_RX_MASK_ACCEPT_ALL_UNICAST; req->rx_mask |= VFPF_RX_MASK_ACCEPT_ALL_MULTICAST; req->rx_mask |= VFPF_RX_MASK_ACCEPT_BROADCAST; break; default: BNX2X_ERR("BAD rx mode (%d)\n", mode); rc = -EINVAL; goto out; } req->flags |= VFPF_SET_Q_FILTERS_RX_MASK_CHANGED; req->vf_qid = 0; /* add list termination tlv */ bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); /* output tlvs list */ bnx2x_dp_tlv_list(bp, req); rc = bnx2x_send_msg2pf(bp, &resp->hdr.status, bp->vf2pf_mbox_mapping); if (rc) BNX2X_ERR("Sending a message failed: %d\n", rc); if (resp->hdr.status != PFVF_STATUS_SUCCESS) { BNX2X_ERR("Set Rx mode failed: %d\n", resp->hdr.status); rc = -EINVAL; } out: bnx2x_vfpf_finalize(bp, &req->first_tlv); return rc; } /* General service functions */ static void storm_memset_vf_mbx_ack(struct bnx2x *bp, u16 abs_fid) { u32 addr = BAR_CSTRORM_INTMEM + CSTORM_VF_PF_CHANNEL_STATE_OFFSET(abs_fid); REG_WR8(bp, addr, VF_PF_CHANNEL_STATE_READY); } static void storm_memset_vf_mbx_valid(struct bnx2x *bp, u16 abs_fid) { u32 addr = BAR_CSTRORM_INTMEM + CSTORM_VF_PF_CHANNEL_VALID_OFFSET(abs_fid); REG_WR8(bp, addr, 1); } static inline void bnx2x_set_vf_mbxs_valid(struct bnx2x *bp) { int i; for_each_vf(bp, i) storm_memset_vf_mbx_valid(bp, bnx2x_vf(bp, i, abs_vfid)); } /* enable vf_pf mailbox (aka vf-pf-chanell) */ void bnx2x_vf_enable_mbx(struct bnx2x *bp, u8 abs_vfid) { bnx2x_vf_flr_clnup_epilog(bp, abs_vfid); /* enable the mailbox in the FW */ storm_memset_vf_mbx_ack(bp, abs_vfid); storm_memset_vf_mbx_valid(bp, abs_vfid); /* enable the VF access to the mailbox */ bnx2x_vf_enable_access(bp, abs_vfid); } /* this works only on !E1h */ static int bnx2x_copy32_vf_dmae(struct bnx2x *bp, u8 from_vf, dma_addr_t pf_addr, u8 vfid, u32 vf_addr_hi, u32 vf_addr_lo, u32 len32) { struct dmae_command dmae; if (CHIP_IS_E1x(bp)) { BNX2X_ERR("Chip revision does not support VFs\n"); return DMAE_NOT_RDY; } if (!bp->dmae_ready) { BNX2X_ERR("DMAE is not ready, can not copy\n"); return DMAE_NOT_RDY; } /* set opcode and fixed command fields */ bnx2x_prep_dmae_with_comp(bp, &dmae, DMAE_SRC_PCI, DMAE_DST_PCI); if (from_vf) { dmae.opcode_iov = (vfid << DMAE_COMMAND_SRC_VFID_SHIFT) | (DMAE_SRC_VF << DMAE_COMMAND_SRC_VFPF_SHIFT) | (DMAE_DST_PF << DMAE_COMMAND_DST_VFPF_SHIFT); dmae.opcode |= (DMAE_C_DST << DMAE_COMMAND_C_FUNC_SHIFT); dmae.src_addr_lo = vf_addr_lo; dmae.src_addr_hi = vf_addr_hi; dmae.dst_addr_lo = U64_LO(pf_addr); dmae.dst_addr_hi = U64_HI(pf_addr); } else { dmae.opcode_iov = (vfid << DMAE_COMMAND_DST_VFID_SHIFT) | (DMAE_DST_VF << DMAE_COMMAND_DST_VFPF_SHIFT) | (DMAE_SRC_PF << DMAE_COMMAND_SRC_VFPF_SHIFT); dmae.opcode |= (DMAE_C_SRC << DMAE_COMMAND_C_FUNC_SHIFT); dmae.src_addr_lo = U64_LO(pf_addr); dmae.src_addr_hi = U64_HI(pf_addr); dmae.dst_addr_lo = vf_addr_lo; dmae.dst_addr_hi = vf_addr_hi; } dmae.len = len32; bnx2x_dp_dmae(bp, &dmae, BNX2X_MSG_DMAE); /* issue the command and wait for completion */ return bnx2x_issue_dmae_with_comp(bp, &dmae); } static void bnx2x_vf_mbx_resp(struct bnx2x *bp, struct bnx2x_virtf *vf) { struct bnx2x_vf_mbx *mbx = BP_VF_MBX(bp, vf->index); u64 vf_addr; dma_addr_t pf_addr; u16 length, type; int rc; struct pfvf_general_resp_tlv *resp = &mbx->msg->resp.general_resp; /* prepare response */ type = mbx->first_tlv.tl.type; length = type == CHANNEL_TLV_ACQUIRE ? sizeof(struct pfvf_acquire_resp_tlv) : sizeof(struct pfvf_general_resp_tlv); bnx2x_add_tlv(bp, resp, 0, type, length); resp->hdr.status = bnx2x_pfvf_status_codes(vf->op_rc); bnx2x_add_tlv(bp, resp, length, CHANNEL_TLV_LIST_END, sizeof(struct channel_list_end_tlv)); bnx2x_dp_tlv_list(bp, resp); DP(BNX2X_MSG_IOV, "mailbox vf address hi 0x%x, lo 0x%x, offset 0x%x\n", mbx->vf_addr_hi, mbx->vf_addr_lo, mbx->first_tlv.resp_msg_offset); /* send response */ vf_addr = HILO_U64(mbx->vf_addr_hi, mbx->vf_addr_lo) + mbx->first_tlv.resp_msg_offset; pf_addr = mbx->msg_mapping + offsetof(struct bnx2x_vf_mbx_msg, resp); /* copy the response body, if there is one, before the header, as the vf * is sensitive to the header being written */ if (resp->hdr.tl.length > sizeof(u64)) { length = resp->hdr.tl.length - sizeof(u64); vf_addr += sizeof(u64); pf_addr += sizeof(u64); rc = bnx2x_copy32_vf_dmae(bp, false, pf_addr, vf->abs_vfid, U64_HI(vf_addr), U64_LO(vf_addr), length/4); if (rc) { BNX2X_ERR("Failed to copy response body to VF %d\n", vf->abs_vfid); goto mbx_error; } vf_addr -= sizeof(u64); pf_addr -= sizeof(u64); } /* ack the FW */ storm_memset_vf_mbx_ack(bp, vf->abs_vfid); mmiowb(); /* initiate dmae to send the response */ mbx->flags &= ~VF_MSG_INPROCESS; /* copy the response header including status-done field, * must be last dmae, must be after FW is acked */ rc = bnx2x_copy32_vf_dmae(bp, false, pf_addr, vf->abs_vfid, U64_HI(vf_addr), U64_LO(vf_addr), sizeof(u64)/4); /* unlock channel mutex */ bnx2x_unlock_vf_pf_channel(bp, vf, mbx->first_tlv.tl.type); if (rc) { BNX2X_ERR("Failed to copy response status to VF %d\n", vf->abs_vfid); goto mbx_error; } return; mbx_error: bnx2x_vf_release(bp, vf, false); /* non blocking */ } static void bnx2x_vf_mbx_acquire_resp(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx, int vfop_status) { int i; struct pfvf_acquire_resp_tlv *resp = &mbx->msg->resp.acquire_resp; struct pf_vf_resc *resc = &resp->resc; u8 status = bnx2x_pfvf_status_codes(vfop_status); memset(resp, 0, sizeof(*resp)); /* fill in pfdev info */ resp->pfdev_info.chip_num = bp->common.chip_id; resp->pfdev_info.db_size = (1 << BNX2X_DB_SHIFT); resp->pfdev_info.indices_per_sb = HC_SB_MAX_INDICES_E2; resp->pfdev_info.pf_cap = (PFVF_CAP_RSS | /* PFVF_CAP_DHC |*/ PFVF_CAP_TPA); bnx2x_fill_fw_str(bp, resp->pfdev_info.fw_ver, sizeof(resp->pfdev_info.fw_ver)); if (status == PFVF_STATUS_NO_RESOURCE || status == PFVF_STATUS_SUCCESS) { /* set resources numbers, if status equals NO_RESOURCE these * are max possible numbers */ resc->num_rxqs = vf_rxq_count(vf) ? : bnx2x_vf_max_queue_cnt(bp, vf); resc->num_txqs = vf_txq_count(vf) ? : bnx2x_vf_max_queue_cnt(bp, vf); resc->num_sbs = vf_sb_count(vf); resc->num_mac_filters = vf_mac_rules_cnt(vf); resc->num_vlan_filters = vf_vlan_rules_cnt(vf); resc->num_mc_filters = 0; if (status == PFVF_STATUS_SUCCESS) { /* fill in the allocated resources */ struct pf_vf_bulletin_content *bulletin = BP_VF_BULLETIN(bp, vf->index); for_each_vfq(vf, i) resc->hw_qid[i] = vfq_qzone_id(vf, vfq_get(vf, i)); for_each_vf_sb(vf, i) { resc->hw_sbs[i].hw_sb_id = vf_igu_sb(vf, i); resc->hw_sbs[i].sb_qid = vf_hc_qzone(vf, i); } /* if a mac has been set for this vf, supply it */ if (bulletin->valid_bitmap & 1 << MAC_ADDR_VALID) { memcpy(resc->current_mac_addr, bulletin->mac, ETH_ALEN); } } } DP(BNX2X_MSG_IOV, "VF[%d] ACQUIRE_RESPONSE: pfdev_info- chip_num=0x%x, db_size=%d, idx_per_sb=%d, pf_cap=0x%x\n" "resources- n_rxq-%d, n_txq-%d, n_sbs-%d, n_macs-%d, n_vlans-%d, n_mcs-%d, fw_ver: '%s'\n", vf->abs_vfid, resp->pfdev_info.chip_num, resp->pfdev_info.db_size, resp->pfdev_info.indices_per_sb, resp->pfdev_info.pf_cap, resc->num_rxqs, resc->num_txqs, resc->num_sbs, resc->num_mac_filters, resc->num_vlan_filters, resc->num_mc_filters, resp->pfdev_info.fw_ver); DP_CONT(BNX2X_MSG_IOV, "hw_qids- [ "); for (i = 0; i < vf_rxq_count(vf); i++) DP_CONT(BNX2X_MSG_IOV, "%d ", resc->hw_qid[i]); DP_CONT(BNX2X_MSG_IOV, "], sb_info- [ "); for (i = 0; i < vf_sb_count(vf); i++) DP_CONT(BNX2X_MSG_IOV, "%d:%d ", resc->hw_sbs[i].hw_sb_id, resc->hw_sbs[i].sb_qid); DP_CONT(BNX2X_MSG_IOV, "]\n"); /* send the response */ vf->op_rc = vfop_status; bnx2x_vf_mbx_resp(bp, vf); } static void bnx2x_vf_mbx_acquire(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { int rc; struct vfpf_acquire_tlv *acquire = &mbx->msg->req.acquire; /* log vfdef info */ DP(BNX2X_MSG_IOV, "VF[%d] ACQUIRE: vfdev_info- vf_id %d, vf_os %d resources- n_rxq-%d, n_txq-%d, n_sbs-%d, n_macs-%d, n_vlans-%d, n_mcs-%d\n", vf->abs_vfid, acquire->vfdev_info.vf_id, acquire->vfdev_info.vf_os, acquire->resc_request.num_rxqs, acquire->resc_request.num_txqs, acquire->resc_request.num_sbs, acquire->resc_request.num_mac_filters, acquire->resc_request.num_vlan_filters, acquire->resc_request.num_mc_filters); /* acquire the resources */ rc = bnx2x_vf_acquire(bp, vf, &acquire->resc_request); /* store address of vf's bulletin board */ vf->bulletin_map = acquire->bulletin_addr; /* response */ bnx2x_vf_mbx_acquire_resp(bp, vf, mbx, rc); } static void bnx2x_vf_mbx_init_vf(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { struct vfpf_init_tlv *init = &mbx->msg->req.init; /* record ghost addresses from vf message */ vf->spq_map = init->spq_addr; vf->fw_stat_map = init->stats_addr; vf->op_rc = bnx2x_vf_init(bp, vf, (dma_addr_t *)init->sb_addr); /* response */ bnx2x_vf_mbx_resp(bp, vf); } /* convert MBX queue-flags to standard SP queue-flags */ static void bnx2x_vf_mbx_set_q_flags(struct bnx2x *bp, u32 mbx_q_flags, unsigned long *sp_q_flags) { if (mbx_q_flags & VFPF_QUEUE_FLG_TPA) __set_bit(BNX2X_Q_FLG_TPA, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_TPA_IPV6) __set_bit(BNX2X_Q_FLG_TPA_IPV6, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_TPA_GRO) __set_bit(BNX2X_Q_FLG_TPA_GRO, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_STATS) __set_bit(BNX2X_Q_FLG_STATS, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_VLAN) __set_bit(BNX2X_Q_FLG_VLAN, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_COS) __set_bit(BNX2X_Q_FLG_COS, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_HC) __set_bit(BNX2X_Q_FLG_HC, sp_q_flags); if (mbx_q_flags & VFPF_QUEUE_FLG_DHC) __set_bit(BNX2X_Q_FLG_DHC, sp_q_flags); /* outer vlan removal is set according to the PF's multi fuction mode */ if (IS_MF_SD(bp)) __set_bit(BNX2X_Q_FLG_OV, sp_q_flags); } static void bnx2x_vf_mbx_setup_q(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { struct vfpf_setup_q_tlv *setup_q = &mbx->msg->req.setup_q; struct bnx2x_vfop_cmd cmd = { .done = bnx2x_vf_mbx_resp, .block = false, }; /* verify vf_qid */ if (setup_q->vf_qid >= vf_rxq_count(vf)) { BNX2X_ERR("vf_qid %d invalid, max queue count is %d\n", setup_q->vf_qid, vf_rxq_count(vf)); vf->op_rc = -EINVAL; goto response; } /* tx queues must be setup alongside rx queues thus if the rx queue * is not marked as valid there's nothing to do. */ if (setup_q->param_valid & (VFPF_RXQ_VALID|VFPF_TXQ_VALID)) { struct bnx2x_vf_queue *q = vfq_get(vf, setup_q->vf_qid); unsigned long q_type = 0; struct bnx2x_queue_init_params *init_p; struct bnx2x_queue_setup_params *setup_p; /* reinit the VF operation context */ memset(&vf->op_params.qctor, 0 , sizeof(vf->op_params.qctor)); setup_p = &vf->op_params.qctor.prep_qsetup; init_p = &vf->op_params.qctor.qstate.params.init; /* activate immediately */ __set_bit(BNX2X_Q_FLG_ACTIVE, &setup_p->flags); if (setup_q->param_valid & VFPF_TXQ_VALID) { struct bnx2x_txq_setup_params *txq_params = &setup_p->txq_params; __set_bit(BNX2X_Q_TYPE_HAS_TX, &q_type); /* save sb resource index */ q->sb_idx = setup_q->txq.vf_sb; /* tx init */ init_p->tx.hc_rate = setup_q->txq.hc_rate; init_p->tx.sb_cq_index = setup_q->txq.sb_index; bnx2x_vf_mbx_set_q_flags(bp, setup_q->txq.flags, &init_p->tx.flags); /* tx setup - flags */ bnx2x_vf_mbx_set_q_flags(bp, setup_q->txq.flags, &setup_p->flags); /* tx setup - general, nothing */ /* tx setup - tx */ txq_params->dscr_map = setup_q->txq.txq_addr; txq_params->sb_cq_index = setup_q->txq.sb_index; txq_params->traffic_type = setup_q->txq.traffic_type; bnx2x_vfop_qctor_dump_tx(bp, vf, init_p, setup_p, q->index, q->sb_idx); } if (setup_q->param_valid & VFPF_RXQ_VALID) { struct bnx2x_rxq_setup_params *rxq_params = &setup_p->rxq_params; __set_bit(BNX2X_Q_TYPE_HAS_RX, &q_type); /* Note: there is no support for different SBs * for TX and RX */ q->sb_idx = setup_q->rxq.vf_sb; /* rx init */ init_p->rx.hc_rate = setup_q->rxq.hc_rate; init_p->rx.sb_cq_index = setup_q->rxq.sb_index; bnx2x_vf_mbx_set_q_flags(bp, setup_q->rxq.flags, &init_p->rx.flags); /* rx setup - flags */ bnx2x_vf_mbx_set_q_flags(bp, setup_q->rxq.flags, &setup_p->flags); /* rx setup - general */ setup_p->gen_params.mtu = setup_q->rxq.mtu; /* rx setup - rx */ rxq_params->drop_flags = setup_q->rxq.drop_flags; rxq_params->dscr_map = setup_q->rxq.rxq_addr; rxq_params->sge_map = setup_q->rxq.sge_addr; rxq_params->rcq_map = setup_q->rxq.rcq_addr; rxq_params->rcq_np_map = setup_q->rxq.rcq_np_addr; rxq_params->buf_sz = setup_q->rxq.buf_sz; rxq_params->tpa_agg_sz = setup_q->rxq.tpa_agg_sz; rxq_params->max_sges_pkt = setup_q->rxq.max_sge_pkt; rxq_params->sge_buf_sz = setup_q->rxq.sge_buf_sz; rxq_params->cache_line_log = setup_q->rxq.cache_line_log; rxq_params->sb_cq_index = setup_q->rxq.sb_index; bnx2x_vfop_qctor_dump_rx(bp, vf, init_p, setup_p, q->index, q->sb_idx); } /* complete the preparations */ bnx2x_vfop_qctor_prep(bp, vf, q, &vf->op_params.qctor, q_type); vf->op_rc = bnx2x_vfop_qsetup_cmd(bp, vf, &cmd, q->index); if (vf->op_rc) goto response; return; } response: bnx2x_vf_mbx_resp(bp, vf); } enum bnx2x_vfop_filters_state { BNX2X_VFOP_MBX_Q_FILTERS_MACS, BNX2X_VFOP_MBX_Q_FILTERS_VLANS, BNX2X_VFOP_MBX_Q_FILTERS_RXMODE, BNX2X_VFOP_MBX_Q_FILTERS_MCAST, BNX2X_VFOP_MBX_Q_FILTERS_DONE }; static int bnx2x_vf_mbx_macvlan_list(struct bnx2x *bp, struct bnx2x_virtf *vf, struct vfpf_set_q_filters_tlv *tlv, struct bnx2x_vfop_filters **pfl, u32 type_flag) { int i, j; struct bnx2x_vfop_filters *fl = NULL; size_t fsz; fsz = tlv->n_mac_vlan_filters * sizeof(struct bnx2x_vfop_filter) + sizeof(struct bnx2x_vfop_filters); fl = kzalloc(fsz, GFP_KERNEL); if (!fl) return -ENOMEM; INIT_LIST_HEAD(&fl->head); for (i = 0, j = 0; i < tlv->n_mac_vlan_filters; i++) { struct vfpf_q_mac_vlan_filter *msg_filter = &tlv->filters[i]; if ((msg_filter->flags & type_flag) != type_flag) continue; if (type_flag == VFPF_Q_FILTER_DEST_MAC_VALID) { fl->filters[j].mac = msg_filter->mac; fl->filters[j].type = BNX2X_VFOP_FILTER_MAC; } else { fl->filters[j].vid = msg_filter->vlan_tag; fl->filters[j].type = BNX2X_VFOP_FILTER_VLAN; } fl->filters[j].add = (msg_filter->flags & VFPF_Q_FILTER_SET_MAC) ? true : false; list_add_tail(&fl->filters[j++].link, &fl->head); } if (list_empty(&fl->head)) kfree(fl); else *pfl = fl; return 0; } static void bnx2x_vf_mbx_dp_q_filter(struct bnx2x *bp, int msglvl, int idx, struct vfpf_q_mac_vlan_filter *filter) { DP(msglvl, "MAC-VLAN[%d] -- flags=0x%x\n", idx, filter->flags); if (filter->flags & VFPF_Q_FILTER_VLAN_TAG_VALID) DP_CONT(msglvl, ", vlan=%d", filter->vlan_tag); if (filter->flags & VFPF_Q_FILTER_DEST_MAC_VALID) DP_CONT(msglvl, ", MAC=%pM", filter->mac); DP_CONT(msglvl, "\n"); } static void bnx2x_vf_mbx_dp_q_filters(struct bnx2x *bp, int msglvl, struct vfpf_set_q_filters_tlv *filters) { int i; if (filters->flags & VFPF_SET_Q_FILTERS_MAC_VLAN_CHANGED) for (i = 0; i < filters->n_mac_vlan_filters; i++) bnx2x_vf_mbx_dp_q_filter(bp, msglvl, i, &filters->filters[i]); if (filters->flags & VFPF_SET_Q_FILTERS_RX_MASK_CHANGED) DP(msglvl, "RX-MASK=0x%x\n", filters->rx_mask); if (filters->flags & VFPF_SET_Q_FILTERS_MULTICAST_CHANGED) for (i = 0; i < filters->n_multicast; i++) DP(msglvl, "MULTICAST=%pM\n", filters->multicast[i]); } #define VFPF_MAC_FILTER VFPF_Q_FILTER_DEST_MAC_VALID #define VFPF_VLAN_FILTER VFPF_Q_FILTER_VLAN_TAG_VALID static void bnx2x_vfop_mbx_qfilters(struct bnx2x *bp, struct bnx2x_virtf *vf) { int rc; struct vfpf_set_q_filters_tlv *msg = &BP_VF_MBX(bp, vf->index)->msg->req.set_q_filters; struct bnx2x_vfop *vfop = bnx2x_vfop_cur(bp, vf); enum bnx2x_vfop_filters_state state = vfop->state; struct bnx2x_vfop_cmd cmd = { .done = bnx2x_vfop_mbx_qfilters, .block = false, }; DP(BNX2X_MSG_IOV, "STATE: %d\n", state); if (vfop->rc < 0) goto op_err; switch (state) { case BNX2X_VFOP_MBX_Q_FILTERS_MACS: /* next state */ vfop->state = BNX2X_VFOP_MBX_Q_FILTERS_VLANS; /* check for any vlan/mac changes */ if (msg->flags & VFPF_SET_Q_FILTERS_MAC_VLAN_CHANGED) { /* build mac list */ struct bnx2x_vfop_filters *fl = NULL; vfop->rc = bnx2x_vf_mbx_macvlan_list(bp, vf, msg, &fl, VFPF_MAC_FILTER); if (vfop->rc) goto op_err; if (fl) { /* set mac list */ rc = bnx2x_vfop_mac_list_cmd(bp, vf, &cmd, fl, msg->vf_qid, false); if (rc) { vfop->rc = rc; goto op_err; } return; } } /* fall through */ case BNX2X_VFOP_MBX_Q_FILTERS_VLANS: /* next state */ vfop->state = BNX2X_VFOP_MBX_Q_FILTERS_RXMODE; /* check for any vlan/mac changes */ if (msg->flags & VFPF_SET_Q_FILTERS_MAC_VLAN_CHANGED) { /* build vlan list */ struct bnx2x_vfop_filters *fl = NULL; vfop->rc = bnx2x_vf_mbx_macvlan_list(bp, vf, msg, &fl, VFPF_VLAN_FILTER); if (vfop->rc) goto op_err; if (fl) { /* set vlan list */ rc = bnx2x_vfop_vlan_list_cmd(bp, vf, &cmd, fl, msg->vf_qid, false); if (rc) { vfop->rc = rc; goto op_err; } return; } } /* fall through */ case BNX2X_VFOP_MBX_Q_FILTERS_RXMODE: /* next state */ vfop->state = BNX2X_VFOP_MBX_Q_FILTERS_MCAST; if (msg->flags & VFPF_SET_Q_FILTERS_RX_MASK_CHANGED) { unsigned long accept = 0; /* covert VF-PF if mask to bnx2x accept flags */ if (msg->rx_mask & VFPF_RX_MASK_ACCEPT_MATCHED_UNICAST) __set_bit(BNX2X_ACCEPT_UNICAST, &accept); if (msg->rx_mask & VFPF_RX_MASK_ACCEPT_MATCHED_MULTICAST) __set_bit(BNX2X_ACCEPT_MULTICAST, &accept); if (msg->rx_mask & VFPF_RX_MASK_ACCEPT_ALL_UNICAST) __set_bit(BNX2X_ACCEPT_ALL_UNICAST, &accept); if (msg->rx_mask & VFPF_RX_MASK_ACCEPT_ALL_MULTICAST) __set_bit(BNX2X_ACCEPT_ALL_MULTICAST, &accept); if (msg->rx_mask & VFPF_RX_MASK_ACCEPT_BROADCAST) __set_bit(BNX2X_ACCEPT_BROADCAST, &accept); /* A packet arriving the vf's mac should be accepted * with any vlan */ __set_bit(BNX2X_ACCEPT_ANY_VLAN, &accept); /* set rx-mode */ rc = bnx2x_vfop_rxmode_cmd(bp, vf, &cmd, msg->vf_qid, accept); if (rc) { vfop->rc = rc; goto op_err; } return; } /* fall through */ case BNX2X_VFOP_MBX_Q_FILTERS_MCAST: /* next state */ vfop->state = BNX2X_VFOP_MBX_Q_FILTERS_DONE; if (msg->flags & VFPF_SET_Q_FILTERS_MULTICAST_CHANGED) { /* set mcasts */ rc = bnx2x_vfop_mcast_cmd(bp, vf, &cmd, msg->multicast, msg->n_multicast, false); if (rc) { vfop->rc = rc; goto op_err; } return; } /* fall through */ op_done: case BNX2X_VFOP_MBX_Q_FILTERS_DONE: bnx2x_vfop_end(bp, vf, vfop); return; op_err: BNX2X_ERR("QFILTERS[%d:%d] error: rc %d\n", vf->abs_vfid, msg->vf_qid, vfop->rc); goto op_done; default: bnx2x_vfop_default(state); } } static int bnx2x_vfop_mbx_qfilters_cmd(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vfop_cmd *cmd) { struct bnx2x_vfop *vfop = bnx2x_vfop_add(bp, vf); if (vfop) { bnx2x_vfop_opset(BNX2X_VFOP_MBX_Q_FILTERS_MACS, bnx2x_vfop_mbx_qfilters, cmd->done); return bnx2x_vfop_transition(bp, vf, bnx2x_vfop_mbx_qfilters, cmd->block); } return -ENOMEM; } static void bnx2x_vf_mbx_set_q_filters(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { struct vfpf_set_q_filters_tlv *filters = &mbx->msg->req.set_q_filters; struct pf_vf_bulletin_content *bulletin = BP_VF_BULLETIN(bp, vf->index); struct bnx2x_vfop_cmd cmd = { .done = bnx2x_vf_mbx_resp, .block = false, }; /* if a mac was already set for this VF via the set vf mac ndo, we only * accept mac configurations of that mac. Why accept them at all? * because PF may have been unable to configure the mac at the time * since queue was not set up. */ if (bulletin->valid_bitmap & 1 << MAC_ADDR_VALID) { /* once a mac was set by ndo can only accept a single mac... */ if (filters->n_mac_vlan_filters > 1) { BNX2X_ERR("VF[%d] requested the addition of multiple macs after set_vf_mac ndo was called\n", vf->abs_vfid); vf->op_rc = -EPERM; goto response; } /* ...and only the mac set by the ndo */ if (filters->n_mac_vlan_filters == 1 && memcmp(filters->filters->mac, bulletin->mac, ETH_ALEN)) { BNX2X_ERR("VF[%d] requested the addition of a mac address not matching the one configured by set_vf_mac ndo\n", vf->abs_vfid); vf->op_rc = -EPERM; goto response; } } /* verify vf_qid */ if (filters->vf_qid > vf_rxq_count(vf)) goto response; DP(BNX2X_MSG_IOV, "VF[%d] Q_FILTERS: queue[%d]\n", vf->abs_vfid, filters->vf_qid); /* print q_filter message */ bnx2x_vf_mbx_dp_q_filters(bp, BNX2X_MSG_IOV, filters); vf->op_rc = bnx2x_vfop_mbx_qfilters_cmd(bp, vf, &cmd); if (vf->op_rc) goto response; return; response: bnx2x_vf_mbx_resp(bp, vf); } static void bnx2x_vf_mbx_teardown_q(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { int qid = mbx->msg->req.q_op.vf_qid; struct bnx2x_vfop_cmd cmd = { .done = bnx2x_vf_mbx_resp, .block = false, }; DP(BNX2X_MSG_IOV, "VF[%d] Q_TEARDOWN: vf_qid=%d\n", vf->abs_vfid, qid); vf->op_rc = bnx2x_vfop_qdown_cmd(bp, vf, &cmd, qid); if (vf->op_rc) bnx2x_vf_mbx_resp(bp, vf); } static void bnx2x_vf_mbx_close_vf(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { struct bnx2x_vfop_cmd cmd = { .done = bnx2x_vf_mbx_resp, .block = false, }; DP(BNX2X_MSG_IOV, "VF[%d] VF_CLOSE\n", vf->abs_vfid); vf->op_rc = bnx2x_vfop_close_cmd(bp, vf, &cmd); if (vf->op_rc) bnx2x_vf_mbx_resp(bp, vf); } static void bnx2x_vf_mbx_release_vf(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { struct bnx2x_vfop_cmd cmd = { .done = bnx2x_vf_mbx_resp, .block = false, }; DP(BNX2X_MSG_IOV, "VF[%d] VF_RELEASE\n", vf->abs_vfid); vf->op_rc = bnx2x_vfop_release_cmd(bp, vf, &cmd); if (vf->op_rc) bnx2x_vf_mbx_resp(bp, vf); } /* dispatch request */ static void bnx2x_vf_mbx_request(struct bnx2x *bp, struct bnx2x_virtf *vf, struct bnx2x_vf_mbx *mbx) { int i; /* check if tlv type is known */ if (bnx2x_tlv_supported(mbx->first_tlv.tl.type)) { /* Lock the per vf op mutex and note the locker's identity. * The unlock will take place in mbx response. */ bnx2x_lock_vf_pf_channel(bp, vf, mbx->first_tlv.tl.type); /* switch on the opcode */ switch (mbx->first_tlv.tl.type) { case CHANNEL_TLV_ACQUIRE: bnx2x_vf_mbx_acquire(bp, vf, mbx); break; case CHANNEL_TLV_INIT: bnx2x_vf_mbx_init_vf(bp, vf, mbx); break; case CHANNEL_TLV_SETUP_Q: bnx2x_vf_mbx_setup_q(bp, vf, mbx); break; case CHANNEL_TLV_SET_Q_FILTERS: bnx2x_vf_mbx_set_q_filters(bp, vf, mbx); break; case CHANNEL_TLV_TEARDOWN_Q: bnx2x_vf_mbx_teardown_q(bp, vf, mbx); break; case CHANNEL_TLV_CLOSE: bnx2x_vf_mbx_close_vf(bp, vf, mbx); break; case CHANNEL_TLV_RELEASE: bnx2x_vf_mbx_release_vf(bp, vf, mbx); break; } } else { /* unknown TLV - this may belong to a VF driver from the future * - a version written after this PF driver was written, which * supports features unknown as of yet. Too bad since we don't * support them. Or this may be because someone wrote a crappy * VF driver and is sending garbage over the channel. */ BNX2X_ERR("unknown TLV. type %d length %d. first 20 bytes of mailbox buffer:\n", mbx->first_tlv.tl.type, mbx->first_tlv.tl.length); for (i = 0; i < 20; i++) DP_CONT(BNX2X_MSG_IOV, "%x ", mbx->msg->req.tlv_buf_size.tlv_buffer[i]); /* test whether we can respond to the VF (do we have an address * for it?) */ if (vf->state == VF_ACQUIRED) { /* mbx_resp uses the op_rc of the VF */ vf->op_rc = PFVF_STATUS_NOT_SUPPORTED; /* notify the VF that we do not support this request */ bnx2x_vf_mbx_resp(bp, vf); } else { /* can't send a response since this VF is unknown to us * just unlock the channel and be done with. */ bnx2x_unlock_vf_pf_channel(bp, vf, mbx->first_tlv.tl.type); } } } /* handle new vf-pf message */ void bnx2x_vf_mbx(struct bnx2x *bp, struct vf_pf_event_data *vfpf_event) { struct bnx2x_virtf *vf; struct bnx2x_vf_mbx *mbx; u8 vf_idx; int rc; DP(BNX2X_MSG_IOV, "vf pf event received: vfid %d, address_hi %x, address lo %x", vfpf_event->vf_id, vfpf_event->msg_addr_hi, vfpf_event->msg_addr_lo); /* Sanity checks consider removing later */ /* check if the vf_id is valid */ if (vfpf_event->vf_id - BP_VFDB(bp)->sriov.first_vf_in_pf > BNX2X_NR_VIRTFN(bp)) { BNX2X_ERR("Illegal vf_id %d max allowed: %d\n", vfpf_event->vf_id, BNX2X_NR_VIRTFN(bp)); goto mbx_done; } vf_idx = bnx2x_vf_idx_by_abs_fid(bp, vfpf_event->vf_id); mbx = BP_VF_MBX(bp, vf_idx); /* verify an event is not currently being processed - * debug failsafe only */ if (mbx->flags & VF_MSG_INPROCESS) { BNX2X_ERR("Previous message is still being processed, vf_id %d\n", vfpf_event->vf_id); goto mbx_done; } vf = BP_VF(bp, vf_idx); /* save the VF message address */ mbx->vf_addr_hi = vfpf_event->msg_addr_hi; mbx->vf_addr_lo = vfpf_event->msg_addr_lo; DP(BNX2X_MSG_IOV, "mailbox vf address hi 0x%x, lo 0x%x, offset 0x%x\n", mbx->vf_addr_hi, mbx->vf_addr_lo, mbx->first_tlv.resp_msg_offset); /* dmae to get the VF request */ rc = bnx2x_copy32_vf_dmae(bp, true, mbx->msg_mapping, vf->abs_vfid, mbx->vf_addr_hi, mbx->vf_addr_lo, sizeof(union vfpf_tlvs)/4); if (rc) { BNX2X_ERR("Failed to copy request VF %d\n", vf->abs_vfid); goto mbx_error; } /* process the VF message header */ mbx->first_tlv = mbx->msg->req.first_tlv; /* dispatch the request (will prepare the response) */ bnx2x_vf_mbx_request(bp, vf, mbx); goto mbx_done; mbx_error: bnx2x_vf_release(bp, vf, false); /* non blocking */ mbx_done: return; } /* propagate local bulletin board to vf */ int bnx2x_post_vf_bulletin(struct bnx2x *bp, int vf) { struct pf_vf_bulletin_content *bulletin = BP_VF_BULLETIN(bp, vf); dma_addr_t pf_addr = BP_VF_BULLETIN_DMA(bp)->mapping + vf * BULLETIN_CONTENT_SIZE; dma_addr_t vf_addr = bnx2x_vf(bp, vf, bulletin_map); int rc; /* can only update vf after init took place */ if (bnx2x_vf(bp, vf, state) != VF_ENABLED && bnx2x_vf(bp, vf, state) != VF_ACQUIRED) return 0; /* increment bulletin board version and compute crc */ bulletin->version++; bulletin->length = BULLETIN_CONTENT_SIZE; bulletin->crc = bnx2x_crc_vf_bulletin(bp, bulletin); /* propagate bulletin board via dmae to vm memory */ rc = bnx2x_copy32_vf_dmae(bp, false, pf_addr, bnx2x_vf(bp, vf, abs_vfid), U64_HI(vf_addr), U64_LO(vf_addr), bulletin->length / 4); return rc; }