|
@@ -551,15 +551,14 @@ error_attrs:
|
|
|
}
|
|
|
|
|
|
static struct msi_desc *
|
|
|
-msi_setup_entry(struct pci_dev *dev, int nvec, bool affinity)
|
|
|
+msi_setup_entry(struct pci_dev *dev, int nvec, const struct irq_affinity *affd)
|
|
|
{
|
|
|
- static const struct irq_affinity default_affd;
|
|
|
struct cpumask *masks = NULL;
|
|
|
struct msi_desc *entry;
|
|
|
u16 control;
|
|
|
|
|
|
- if (affinity) {
|
|
|
- masks = irq_create_affinity_masks(nvec, &default_affd);
|
|
|
+ if (affd) {
|
|
|
+ masks = irq_create_affinity_masks(nvec, affd);
|
|
|
if (!masks)
|
|
|
pr_err("Unable to allocate affinity masks, ignoring\n");
|
|
|
}
|
|
@@ -619,7 +618,8 @@ static int msi_verify_entries(struct pci_dev *dev)
|
|
|
* an error, and a positive return value indicates the number of interrupts
|
|
|
* which could have been allocated.
|
|
|
*/
|
|
|
-static int msi_capability_init(struct pci_dev *dev, int nvec, bool affinity)
|
|
|
+static int msi_capability_init(struct pci_dev *dev, int nvec,
|
|
|
+ const struct irq_affinity *affd)
|
|
|
{
|
|
|
struct msi_desc *entry;
|
|
|
int ret;
|
|
@@ -627,7 +627,7 @@ static int msi_capability_init(struct pci_dev *dev, int nvec, bool affinity)
|
|
|
|
|
|
pci_msi_set_enable(dev, 0); /* Disable MSI during set up */
|
|
|
|
|
|
- entry = msi_setup_entry(dev, nvec, affinity);
|
|
|
+ entry = msi_setup_entry(dev, nvec, affd);
|
|
|
if (!entry)
|
|
|
return -ENOMEM;
|
|
|
|
|
@@ -691,15 +691,14 @@ static void __iomem *msix_map_region(struct pci_dev *dev, unsigned nr_entries)
|
|
|
|
|
|
static int msix_setup_entries(struct pci_dev *dev, void __iomem *base,
|
|
|
struct msix_entry *entries, int nvec,
|
|
|
- bool affinity)
|
|
|
+ const struct irq_affinity *affd)
|
|
|
{
|
|
|
- static const struct irq_affinity default_affd;
|
|
|
struct cpumask *curmsk, *masks = NULL;
|
|
|
struct msi_desc *entry;
|
|
|
int ret, i;
|
|
|
|
|
|
- if (affinity) {
|
|
|
- masks = irq_create_affinity_masks(nvec, &default_affd);
|
|
|
+ if (affd) {
|
|
|
+ masks = irq_create_affinity_masks(nvec, affd);
|
|
|
if (!masks)
|
|
|
pr_err("Unable to allocate affinity masks, ignoring\n");
|
|
|
}
|
|
@@ -755,14 +754,14 @@ static void msix_program_entries(struct pci_dev *dev,
|
|
|
* @dev: pointer to the pci_dev data structure of MSI-X device function
|
|
|
* @entries: pointer to an array of struct msix_entry entries
|
|
|
* @nvec: number of @entries
|
|
|
- * @affinity: flag to indicate cpu irq affinity mask should be set
|
|
|
+ * @affd: Optional pointer to enable automatic affinity assignement
|
|
|
*
|
|
|
* Setup the MSI-X capability structure of device function with a
|
|
|
* single MSI-X irq. A return of zero indicates the successful setup of
|
|
|
* requested MSI-X entries with allocated irqs or non-zero for otherwise.
|
|
|
**/
|
|
|
static int msix_capability_init(struct pci_dev *dev, struct msix_entry *entries,
|
|
|
- int nvec, bool affinity)
|
|
|
+ int nvec, const struct irq_affinity *affd)
|
|
|
{
|
|
|
int ret;
|
|
|
u16 control;
|
|
@@ -777,7 +776,7 @@ static int msix_capability_init(struct pci_dev *dev, struct msix_entry *entries,
|
|
|
if (!base)
|
|
|
return -ENOMEM;
|
|
|
|
|
|
- ret = msix_setup_entries(dev, base, entries, nvec, affinity);
|
|
|
+ ret = msix_setup_entries(dev, base, entries, nvec, affd);
|
|
|
if (ret)
|
|
|
return ret;
|
|
|
|
|
@@ -958,7 +957,7 @@ int pci_msix_vec_count(struct pci_dev *dev)
|
|
|
EXPORT_SYMBOL(pci_msix_vec_count);
|
|
|
|
|
|
static int __pci_enable_msix(struct pci_dev *dev, struct msix_entry *entries,
|
|
|
- int nvec, bool affinity)
|
|
|
+ int nvec, const struct irq_affinity *affd)
|
|
|
{
|
|
|
int nr_entries;
|
|
|
int i, j;
|
|
@@ -990,7 +989,7 @@ static int __pci_enable_msix(struct pci_dev *dev, struct msix_entry *entries,
|
|
|
dev_info(&dev->dev, "can't enable MSI-X (MSI IRQ already assigned)\n");
|
|
|
return -EINVAL;
|
|
|
}
|
|
|
- return msix_capability_init(dev, entries, nvec, affinity);
|
|
|
+ return msix_capability_init(dev, entries, nvec, affd);
|
|
|
}
|
|
|
|
|
|
/**
|
|
@@ -1010,7 +1009,7 @@ static int __pci_enable_msix(struct pci_dev *dev, struct msix_entry *entries,
|
|
|
**/
|
|
|
int pci_enable_msix(struct pci_dev *dev, struct msix_entry *entries, int nvec)
|
|
|
{
|
|
|
- return __pci_enable_msix(dev, entries, nvec, false);
|
|
|
+ return __pci_enable_msix(dev, entries, nvec, NULL);
|
|
|
}
|
|
|
EXPORT_SYMBOL(pci_enable_msix);
|
|
|
|
|
@@ -1061,10 +1060,8 @@ int pci_msi_enabled(void)
|
|
|
EXPORT_SYMBOL(pci_msi_enabled);
|
|
|
|
|
|
static int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec,
|
|
|
- unsigned int flags)
|
|
|
+ const struct irq_affinity *affd)
|
|
|
{
|
|
|
- static const struct irq_affinity default_affd;
|
|
|
- bool affinity = flags & PCI_IRQ_AFFINITY;
|
|
|
int nvec;
|
|
|
int rc;
|
|
|
|
|
@@ -1093,13 +1090,13 @@ static int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec,
|
|
|
nvec = maxvec;
|
|
|
|
|
|
for (;;) {
|
|
|
- if (affinity) {
|
|
|
- nvec = irq_calc_affinity_vectors(nvec, &default_affd);
|
|
|
+ if (affd) {
|
|
|
+ nvec = irq_calc_affinity_vectors(nvec, affd);
|
|
|
if (nvec < minvec)
|
|
|
return -ENOSPC;
|
|
|
}
|
|
|
|
|
|
- rc = msi_capability_init(dev, nvec, affinity);
|
|
|
+ rc = msi_capability_init(dev, nvec, affd);
|
|
|
if (rc == 0)
|
|
|
return nvec;
|
|
|
|
|
@@ -1126,29 +1123,27 @@ static int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec,
|
|
|
**/
|
|
|
int pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec)
|
|
|
{
|
|
|
- return __pci_enable_msi_range(dev, minvec, maxvec, 0);
|
|
|
+ return __pci_enable_msi_range(dev, minvec, maxvec, NULL);
|
|
|
}
|
|
|
EXPORT_SYMBOL(pci_enable_msi_range);
|
|
|
|
|
|
static int __pci_enable_msix_range(struct pci_dev *dev,
|
|
|
- struct msix_entry *entries, int minvec, int maxvec,
|
|
|
- unsigned int flags)
|
|
|
+ struct msix_entry *entries, int minvec,
|
|
|
+ int maxvec, const struct irq_affinity *affd)
|
|
|
{
|
|
|
- static const struct irq_affinity default_affd;
|
|
|
- bool affinity = flags & PCI_IRQ_AFFINITY;
|
|
|
int rc, nvec = maxvec;
|
|
|
|
|
|
if (maxvec < minvec)
|
|
|
return -ERANGE;
|
|
|
|
|
|
for (;;) {
|
|
|
- if (affinity) {
|
|
|
- nvec = irq_calc_affinity_vectors(nvec, &default_affd);
|
|
|
+ if (affd) {
|
|
|
+ nvec = irq_calc_affinity_vectors(nvec, affd);
|
|
|
if (nvec < minvec)
|
|
|
return -ENOSPC;
|
|
|
}
|
|
|
|
|
|
- rc = __pci_enable_msix(dev, entries, nvec, affinity);
|
|
|
+ rc = __pci_enable_msix(dev, entries, nvec, affd);
|
|
|
if (rc == 0)
|
|
|
return nvec;
|
|
|
|
|
@@ -1179,7 +1174,7 @@ static int __pci_enable_msix_range(struct pci_dev *dev,
|
|
|
int pci_enable_msix_range(struct pci_dev *dev, struct msix_entry *entries,
|
|
|
int minvec, int maxvec)
|
|
|
{
|
|
|
- return __pci_enable_msix_range(dev, entries, minvec, maxvec, 0);
|
|
|
+ return __pci_enable_msix_range(dev, entries, minvec, maxvec, NULL);
|
|
|
}
|
|
|
EXPORT_SYMBOL(pci_enable_msix_range);
|
|
|
|
|
@@ -1203,17 +1198,22 @@ EXPORT_SYMBOL(pci_enable_msix_range);
|
|
|
int pci_alloc_irq_vectors(struct pci_dev *dev, unsigned int min_vecs,
|
|
|
unsigned int max_vecs, unsigned int flags)
|
|
|
{
|
|
|
+ static const struct irq_affinity msi_default_affd;
|
|
|
+ const struct irq_affinity *affd = NULL;
|
|
|
int vecs = -ENOSPC;
|
|
|
|
|
|
+ if (flags & PCI_IRQ_AFFINITY)
|
|
|
+ affd = &msi_default_affd;
|
|
|
+
|
|
|
if (flags & PCI_IRQ_MSIX) {
|
|
|
vecs = __pci_enable_msix_range(dev, NULL, min_vecs, max_vecs,
|
|
|
- flags);
|
|
|
+ affd);
|
|
|
if (vecs > 0)
|
|
|
return vecs;
|
|
|
}
|
|
|
|
|
|
if (flags & PCI_IRQ_MSI) {
|
|
|
- vecs = __pci_enable_msi_range(dev, min_vecs, max_vecs, flags);
|
|
|
+ vecs = __pci_enable_msi_range(dev, min_vecs, max_vecs, affd);
|
|
|
if (vecs > 0)
|
|
|
return vecs;
|
|
|
}
|