Commit d29a2488 authored by Oliver O'Halloran's avatar Oliver O'Halloran Committed by Michael Ellerman

powerpc/powernv/sriov: Drop iov->pe_num_map[]

Currently the iov->pe_num_map[] does one of two things depending on
whether single PE mode is being used or not. When it is, this contains an
array which maps a vf_index to the corresponding PE number. When single PE
mode is not being used this contains a scalar which is the base PE for the
set of enabled VFs (for for VFn is base + n).

The array was necessary because when calling pnv_ioda_alloc_pe() there is
no guarantee that the allocated PEs would be contigious. We can now
allocate contigious blocks of PEs so this is no longer an issue. This
allows us to drop the if (single_mode) {} .. else {} block scattered
through the SR-IOV code which is a nice clean up.

This also fixes a bug in pnv_pci_sriov_disable() which is the non-atomic
bitmap_clear() to manipulate the PE allocation map. Other users of the map
assume it will be accessed with atomic ops.
Signed-off-by: default avatarOliver O'Halloran <oohall@gmail.com>
Reviewed-by: default avatarAlexey Kardashevskiy <aik@ozlabs.ru>
Signed-off-by: default avatarMichael Ellerman <mpe@ellerman.id.au>
Link: https://lore.kernel.org/r/20200722065715.1432738-11-oohall@gmail.com
parent a4bc676e
...@@ -452,11 +452,13 @@ static int pnv_pci_vf_assign_m64(struct pci_dev *pdev, u16 num_vfs) ...@@ -452,11 +452,13 @@ static int pnv_pci_vf_assign_m64(struct pci_dev *pdev, u16 num_vfs)
if (iov->m64_single_mode) { if (iov->m64_single_mode) {
int pe_num = iov->vf_pe_arr[j].pe_number;
size = pci_iov_resource_size(pdev, size = pci_iov_resource_size(pdev,
PCI_IOV_RESOURCES + i); PCI_IOV_RESOURCES + i);
start = res->start + size * j; start = res->start + size * j;
rc = pnv_ioda_map_m64_single(phb, win, rc = pnv_ioda_map_m64_single(phb, win,
iov->pe_num_map[j], pe_num,
start, start,
size); size);
} else { } else {
...@@ -595,38 +597,24 @@ static int pnv_pci_vf_resource_shift(struct pci_dev *dev, int offset) ...@@ -595,38 +597,24 @@ static int pnv_pci_vf_resource_shift(struct pci_dev *dev, int offset)
static void pnv_pci_sriov_disable(struct pci_dev *pdev) static void pnv_pci_sriov_disable(struct pci_dev *pdev)
{ {
u16 num_vfs, base_pe;
struct pnv_phb *phb; struct pnv_phb *phb;
struct pnv_ioda_pe *pe;
struct pnv_iov_data *iov; struct pnv_iov_data *iov;
u16 num_vfs, i;
phb = pci_bus_to_pnvhb(pdev->bus); phb = pci_bus_to_pnvhb(pdev->bus);
iov = pnv_iov_get(pdev); iov = pnv_iov_get(pdev);
num_vfs = iov->num_vfs; num_vfs = iov->num_vfs;
base_pe = iov->vf_pe_arr[0].pe_number;
/* Release VF PEs */ /* Release VF PEs */
pnv_ioda_release_vf_PE(pdev); pnv_ioda_release_vf_PE(pdev);
if (phb->type == PNV_PHB_IODA2) { if (phb->type == PNV_PHB_IODA2) {
if (!iov->m64_single_mode) if (!iov->m64_single_mode)
pnv_pci_vf_resource_shift(pdev, -*iov->pe_num_map); pnv_pci_vf_resource_shift(pdev, -base_pe);
/* Release M64 windows */ /* Release M64 windows */
pnv_pci_vf_release_m64(pdev, num_vfs); pnv_pci_vf_release_m64(pdev, num_vfs);
/* Release PE numbers */
if (iov->m64_single_mode) {
for (i = 0; i < num_vfs; i++) {
if (iov->pe_num_map[i] == IODA_INVALID_PE)
continue;
pe = &phb->ioda.pe_array[iov->pe_num_map[i]];
pnv_ioda_free_pe(pe);
}
} else
bitmap_clear(phb->ioda.pe_alloc, *iov->pe_num_map, num_vfs);
/* Releasing pe_num_map */
kfree(iov->pe_num_map);
} }
} }
...@@ -652,13 +640,7 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs) ...@@ -652,13 +640,7 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs)
int vf_bus = pci_iov_virtfn_bus(pdev, vf_index); int vf_bus = pci_iov_virtfn_bus(pdev, vf_index);
struct pci_dn *vf_pdn; struct pci_dn *vf_pdn;
if (iov->m64_single_mode) pe = &iov->vf_pe_arr[vf_index];
pe_num = iov->pe_num_map[vf_index];
else
pe_num = *iov->pe_num_map + vf_index;
pe = &phb->ioda.pe_array[pe_num];
pe->pe_number = pe_num;
pe->phb = phb; pe->phb = phb;
pe->flags = PNV_IODA_PE_VF; pe->flags = PNV_IODA_PE_VF;
pe->pbus = NULL; pe->pbus = NULL;
...@@ -666,6 +648,7 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs) ...@@ -666,6 +648,7 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs)
pe->mve_number = -1; pe->mve_number = -1;
pe->rid = (vf_bus << 8) | vf_devfn; pe->rid = (vf_bus << 8) | vf_devfn;
pe_num = pe->pe_number;
pe_info(pe, "VF %04d:%02d:%02d.%d associated with PE#%x\n", pe_info(pe, "VF %04d:%02d:%02d.%d associated with PE#%x\n",
pci_domain_nr(pdev->bus), pdev->bus->number, pci_domain_nr(pdev->bus), pdev->bus->number,
PCI_SLOT(vf_devfn), PCI_FUNC(vf_devfn), pe_num); PCI_SLOT(vf_devfn), PCI_FUNC(vf_devfn), pe_num);
...@@ -697,9 +680,9 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs) ...@@ -697,9 +680,9 @@ static void pnv_ioda_setup_vf_PE(struct pci_dev *pdev, u16 num_vfs)
static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
{ {
struct pnv_ioda_pe *base_pe;
struct pnv_iov_data *iov; struct pnv_iov_data *iov;
struct pnv_phb *phb; struct pnv_phb *phb;
struct pnv_ioda_pe *pe;
int ret; int ret;
u16 i; u16 i;
...@@ -713,55 +696,14 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -713,55 +696,14 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
return -ENOSPC; return -ENOSPC;
} }
/* /* allocate a contigious block of PEs for our VFs */
* When M64 BARs functions in Single PE mode, the number of VFs base_pe = pnv_ioda_alloc_pe(phb, num_vfs);
* could be enabled must be less than the number of M64 BARs. if (!base_pe) {
*/ pci_err(pdev, "Unable to allocate PEs for %d VFs\n", num_vfs);
if (iov->m64_single_mode && num_vfs > phb->ioda.m64_bar_idx) {
dev_info(&pdev->dev, "Not enough M64 BAR for VFs\n");
return -EBUSY; return -EBUSY;
} }
/* Allocating pe_num_map */ iov->vf_pe_arr = base_pe;
if (iov->m64_single_mode)
iov->pe_num_map = kmalloc_array(num_vfs,
sizeof(*iov->pe_num_map),
GFP_KERNEL);
else
iov->pe_num_map = kmalloc(sizeof(*iov->pe_num_map), GFP_KERNEL);
if (!iov->pe_num_map)
return -ENOMEM;
if (iov->m64_single_mode)
for (i = 0; i < num_vfs; i++)
iov->pe_num_map[i] = IODA_INVALID_PE;
/* Calculate available PE for required VFs */
if (iov->m64_single_mode) {
for (i = 0; i < num_vfs; i++) {
pe = pnv_ioda_alloc_pe(phb);
if (!pe) {
ret = -EBUSY;
goto m64_failed;
}
iov->pe_num_map[i] = pe->pe_number;
}
} else {
mutex_lock(&phb->ioda.pe_alloc_mutex);
*iov->pe_num_map = bitmap_find_next_zero_area(
phb->ioda.pe_alloc, phb->ioda.total_pe_num,
0, num_vfs, 0);
if (*iov->pe_num_map >= phb->ioda.total_pe_num) {
mutex_unlock(&phb->ioda.pe_alloc_mutex);
dev_info(&pdev->dev, "Failed to enable VF%d\n", num_vfs);
kfree(iov->pe_num_map);
return -EBUSY;
}
bitmap_set(phb->ioda.pe_alloc, *iov->pe_num_map, num_vfs);
mutex_unlock(&phb->ioda.pe_alloc_mutex);
}
iov->num_vfs = num_vfs; iov->num_vfs = num_vfs;
/* Assign M64 window accordingly */ /* Assign M64 window accordingly */
...@@ -777,9 +719,10 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -777,9 +719,10 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
* Otherwise, the PE# for the VF will conflict with others. * Otherwise, the PE# for the VF will conflict with others.
*/ */
if (!iov->m64_single_mode) { if (!iov->m64_single_mode) {
ret = pnv_pci_vf_resource_shift(pdev, *iov->pe_num_map); ret = pnv_pci_vf_resource_shift(pdev,
base_pe->pe_number);
if (ret) if (ret)
goto m64_failed; goto shift_failed;
} }
} }
...@@ -788,20 +731,12 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) ...@@ -788,20 +731,12 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs)
return 0; return 0;
m64_failed: shift_failed:
if (iov->m64_single_mode) { pnv_pci_vf_release_m64(pdev, num_vfs);
for (i = 0; i < num_vfs; i++) {
if (iov->pe_num_map[i] == IODA_INVALID_PE)
continue;
pe = &phb->ioda.pe_array[iov->pe_num_map[i]];
pnv_ioda_free_pe(pe);
}
} else
bitmap_clear(phb->ioda.pe_alloc, *iov->pe_num_map, num_vfs);
/* Releasing pe_num_map */ m64_failed:
kfree(iov->pe_num_map); for (i = 0; i < num_vfs; i++)
pnv_ioda_free_pe(&iov->vf_pe_arr[i]);
return ret; return ret;
} }
......
...@@ -239,7 +239,12 @@ struct pnv_iov_data { ...@@ -239,7 +239,12 @@ struct pnv_iov_data {
/* number of VFs enabled */ /* number of VFs enabled */
u16 num_vfs; u16 num_vfs;
unsigned int *pe_num_map; /* PE# for the first VF PE or array */
/*
* Pointer to the IODA PE state of each VF. Note that this is a pointer
* into the PHB's PE array (phb->ioda.pe_array).
*/
struct pnv_ioda_pe *vf_pe_arr;
/* Did we map the VF BARs with single-PE IODA BARs? */ /* Did we map the VF BARs with single-PE IODA BARs? */
bool m64_single_mode; bool m64_single_mode;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment