提交 be283eeb 编写于 作者: W Wei Yang 提交者: Michael Ellerman

powerpc/powernv: allocate sparse PE# when using M64 BAR in Single PE mode

When M64 BAR is set to Single PE mode, the PE# assigned to VF could be
sparse.

This patch restructures the code to allocate sparse PE# for VFs when M64
BAR is set to Single PE mode. Also it rename the offset to pe_num_map to
reflect the content is the PE number.
Signed-off-by: NWei Yang <weiyang@linux.vnet.ibm.com>
Reviewed-by: NGavin Shan <gwshan@linux.vnet.ibm.com>
Acked-by: NAlexey Kardashevskiy <aik@ozlabs.ru>
Signed-off-by: NMichael Ellerman <mpe@ellerman.id.au>
上级 dfcc8d45
...@@ -215,7 +215,7 @@ struct pci_dn { ...@@ -215,7 +215,7 @@ struct pci_dn {
#ifdef CONFIG_PCI_IOV #ifdef CONFIG_PCI_IOV
u16 vfs_expanded; /* number of VFs IOV BAR expanded */ u16 vfs_expanded; /* number of VFs IOV BAR expanded */
u16 num_vfs; /* number of VFs enabled*/ u16 num_vfs; /* number of VFs enabled*/
int offset; /* PE# for the first VF PE */ int *pe_num_map; /* PE# for the first VF PE or array */
bool m64_single_mode; /* Use M64 BAR in Single Mode */ bool m64_single_mode; /* Use M64 BAR in Single Mode */
#define IODA_INVALID_M64 (-1) #define IODA_INVALID_M64 (-1)
int (*m64_map)[PCI_SRIOV_NUM_BARS]; int (*m64_map)[PCI_SRIOV_NUM_BARS];
......
...@@ -1285,7 +1285,7 @@ static int pnv_pci_vf_assign_m64(struct pci_dev *pdev, u16 num_vfs) ...@@ -1285,7 +1285,7 @@ static int pnv_pci_vf_assign_m64(struct pci_dev *pdev, u16 num_vfs)
/* Map the M64 here */ /* Map the M64 here */
if (pdn->m64_single_mode) { if (pdn->m64_single_mode) {
pe_num = pdn->offset + j; pe_num = pdn->pe_num_map[j];
rc = opal_pci_map_pe_mmio_window(phb->opal_id, rc = opal_pci_map_pe_mmio_window(phb->opal_id,
pe_num, OPAL_M64_WINDOW_TYPE, pe_num, OPAL_M64_WINDOW_TYPE,
pdn->m64_map[j][i], 0); pdn->m64_map[j][i], 0);
...@@ -1389,7 +1389,7 @@ void pnv_pci_sriov_disable(struct pci_dev *pdev) ...@@ -1389,7 +1389,7 @@ void pnv_pci_sriov_disable(struct pci_dev *pdev)
struct pnv_phb *phb; struct pnv_phb *phb;
struct pci_dn *pdn; struct pci_dn *pdn;
struct pci_sriov *iov; struct pci_sriov *iov;
u16 num_vfs; u16 num_vfs, i;
bus = pdev->bus; bus = pdev->bus;
hose = pci_bus_to_host(bus); hose = pci_bus_to_host(bus);
...@@ -1403,14 +1403,21 @@ void pnv_pci_sriov_disable(struct pci_dev *pdev) ...@@ -1403,14 +1403,21 @@ void pnv_pci_sriov_disable(struct pci_dev *pdev)
if (phb->type == PNV_PHB_IODA2) { if (phb->type == PNV_PHB_IODA2) {
if (!pdn->m64_single_mode) if (!pdn->m64_single_mode)
pnv_pci_vf_resource_shift(pdev, -pdn->offset); pnv_pci_vf_resource_shift(pdev, -*pdn->pe_num_map);
/* Release M64 windows */ /* Release M64 windows */
pnv_pci_vf_release_m64(pdev, num_vfs); pnv_pci_vf_release_m64(pdev, num_vfs);
/* Release PE numbers */ /* Release PE numbers */
bitmap_clear(phb->ioda.pe_alloc, pdn->offset, num_vfs); if (pdn->m64_single_mode) {
pdn->offset = 0; for (i = 0; i < num_vfs; i++) {
if (pdn->pe_num_map[i] != IODA_INVALID_PE)
pnv_ioda_free_pe(phb, pdn->pe_num_map[i]);
}
} else
bitmap_clear(phb->ioda.pe_alloc, *pdn->pe_num_map, num_vfs);
/* Releasing pe_num_map */
kfree(pdn->pe_num_map);
} }
} }
...@@ -1436,7 +1443,10 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs) ...@@ -1436,7 +1443,10 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs)
/* Reserve PE for each VF */ /* Reserve PE for each VF */
for (vf_index = 0; vf_index < num_vfs; vf_index++) { for (vf_index = 0; vf_index < num_vfs; vf_index++) {
pe_num = pdn->offset + vf_index; if (pdn->m64_single_mode)
pe_num = pdn->pe_num_map[vf_index];
else
pe_num = *pdn->pe_num_map + vf_index;
pe = &phb->ioda.pe_array[pe_num]; pe = &phb->ioda.pe_array[pe_num];
pe->pe_number = pe_num; pe->pe_number = pe_num;
...@@ -1478,6 +1488,7 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -1478,6 +1488,7 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
struct pnv_phb *phb; struct pnv_phb *phb;
struct pci_dn *pdn; struct pci_dn *pdn;
int ret; int ret;
u16 i;
bus = pdev->bus; bus = pdev->bus;
hose = pci_bus_to_host(bus); hose = pci_bus_to_host(bus);
...@@ -1500,20 +1511,44 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -1500,20 +1511,44 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
return -EBUSY; return -EBUSY;
} }
/* Allocating pe_num_map */
if (pdn->m64_single_mode)
pdn->pe_num_map = kmalloc(sizeof(*pdn->pe_num_map) * num_vfs,
GFP_KERNEL);
else
pdn->pe_num_map = kmalloc(sizeof(*pdn->pe_num_map), GFP_KERNEL);
if (!pdn->pe_num_map)
return -ENOMEM;
if (pdn->m64_single_mode)
for (i = 0; i < num_vfs; i++)
pdn->pe_num_map[i] = IODA_INVALID_PE;
/* Calculate available PE for required VFs */ /* Calculate available PE for required VFs */
if (pdn->m64_single_mode) {
for (i = 0; i < num_vfs; i++) {
pdn->pe_num_map[i] = pnv_ioda_alloc_pe(phb);
if (pdn->pe_num_map[i] == IODA_INVALID_PE) {
ret = -EBUSY;
goto m64_failed;
}
}
} else {
mutex_lock(&phb->ioda.pe_alloc_mutex); mutex_lock(&phb->ioda.pe_alloc_mutex);
pdn->offset = bitmap_find_next_zero_area( *pdn->pe_num_map = bitmap_find_next_zero_area(
phb->ioda.pe_alloc, phb->ioda.total_pe, phb->ioda.pe_alloc, phb->ioda.total_pe,
0, num_vfs, 0); 0, num_vfs, 0);
if (pdn->offset >= phb->ioda.total_pe) { if (*pdn->pe_num_map >= phb->ioda.total_pe) {
mutex_unlock(&phb->ioda.pe_alloc_mutex); mutex_unlock(&phb->ioda.pe_alloc_mutex);
dev_info(&pdev->dev, "Failed to enable VF%d\n", num_vfs); dev_info(&pdev->dev, "Failed to enable VF%d\n", num_vfs);
pdn->offset = 0; kfree(pdn->pe_num_map);
return -EBUSY; return -EBUSY;
} }
bitmap_set(phb->ioda.pe_alloc, pdn->offset, num_vfs); bitmap_set(phb->ioda.pe_alloc, *pdn->pe_num_map, num_vfs);
pdn->num_vfs = num_vfs;
mutex_unlock(&phb->ioda.pe_alloc_mutex); mutex_unlock(&phb->ioda.pe_alloc_mutex);
}
pdn->num_vfs = num_vfs;
/* Assign M64 window accordingly */ /* Assign M64 window accordingly */
ret = pnv_pci_vf_assign_m64(pdev, num_vfs); ret = pnv_pci_vf_assign_m64(pdev, num_vfs);
...@@ -1528,7 +1563,7 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -1528,7 +1563,7 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
* Otherwise, the PE# for the VF will conflict with others. * Otherwise, the PE# for the VF will conflict with others.
*/ */
if (!pdn->m64_single_mode) { if (!pdn->m64_single_mode) {
ret = pnv_pci_vf_resource_shift(pdev, pdn->offset); ret = pnv_pci_vf_resource_shift(pdev, *pdn->pe_num_map);
if (ret) if (ret)
goto m64_failed; goto m64_failed;
} }
...@@ -1540,8 +1575,16 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -1540,8 +1575,16 @@ int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
return 0; return 0;
m64_failed: m64_failed:
bitmap_clear(phb->ioda.pe_alloc, pdn->offset, num_vfs); if (pdn->m64_single_mode) {
pdn->offset = 0; for (i = 0; i < num_vfs; i++) {
if (pdn->pe_num_map[i] != IODA_INVALID_PE)
pnv_ioda_free_pe(phb, pdn->pe_num_map[i]);
}
} else
bitmap_clear(phb->ioda.pe_alloc, *pdn->pe_num_map, num_vfs);
/* Releasing pe_num_map */
kfree(pdn->pe_num_map);
return ret; return ret;
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册