|
@@ -2011,27 +2011,62 @@ static unsigned int pnv_pci_ioda_pe_dma_weight(struct pnv_ioda_pe *pe)
|
|
|
}
|
|
|
|
|
|
static void pnv_pci_ioda1_setup_dma_pe(struct pnv_phb *phb,
|
|
|
- struct pnv_ioda_pe *pe,
|
|
|
- unsigned int base,
|
|
|
- unsigned int segs)
|
|
|
+ struct pnv_ioda_pe *pe)
|
|
|
{
|
|
|
|
|
|
struct page *tce_mem = NULL;
|
|
|
struct iommu_table *tbl;
|
|
|
- unsigned int tce32_segsz, i;
|
|
|
+ unsigned int weight, total_weight = 0;
|
|
|
+ unsigned int tce32_segsz, base, segs, avail, i;
|
|
|
int64_t rc;
|
|
|
void *addr;
|
|
|
|
|
|
/* XXX FIXME: Handle 64-bit only DMA devices */
|
|
|
/* XXX FIXME: Provide 64-bit DMA facilities & non-4K TCE tables etc.. */
|
|
|
/* XXX FIXME: Allocate multi-level tables on PHB3 */
|
|
|
+ weight = pnv_pci_ioda_pe_dma_weight(pe);
|
|
|
+ if (!weight)
|
|
|
+ return;
|
|
|
+
|
|
|
+ pci_walk_bus(phb->hose->bus, pnv_pci_ioda_dev_dma_weight,
|
|
|
+ &total_weight);
|
|
|
+ segs = (weight * phb->ioda.dma32_count) / total_weight;
|
|
|
+ if (!segs)
|
|
|
+ segs = 1;
|
|
|
|
|
|
+ /*
|
|
|
+ * Allocate contiguous DMA32 segments. We begin with the expected
|
|
|
+ * number of segments. With one more attempt, the number of DMA32
|
|
|
+ * segments to be allocated is decreased by one until one segment
|
|
|
+ * is allocated successfully.
|
|
|
+ */
|
|
|
+ do {
|
|
|
+ for (base = 0; base <= phb->ioda.dma32_count - segs; base++) {
|
|
|
+ for (avail = 0, i = base; i < base + segs; i++) {
|
|
|
+ if (phb->ioda.dma32_segmap[i] ==
|
|
|
+ IODA_INVALID_PE)
|
|
|
+ avail++;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (avail == segs)
|
|
|
+ goto found;
|
|
|
+ }
|
|
|
+ } while (--segs);
|
|
|
+
|
|
|
+ if (!segs) {
|
|
|
+ pe_warn(pe, "No available DMA32 segments\n");
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+found:
|
|
|
tbl = pnv_pci_table_alloc(phb->hose->node);
|
|
|
iommu_register_group(&pe->table_group, phb->hose->global_number,
|
|
|
pe->pe_number);
|
|
|
pnv_pci_link_table_and_group(phb->hose->node, 0, tbl, &pe->table_group);
|
|
|
|
|
|
/* Grab a 32-bit TCE table */
|
|
|
+ pe_info(pe, "DMA weight %d (%d), assigned (%d) %d DMA32 segments\n",
|
|
|
+ weight, total_weight, base, segs);
|
|
|
pe_info(pe, " Setting up 32-bit TCE table at %08x..%08x\n",
|
|
|
base * PNV_IODA1_DMA32_SEGSIZE,
|
|
|
(base + segs) * PNV_IODA1_DMA32_SEGSIZE - 1);
|
|
@@ -2068,6 +2103,10 @@ static void pnv_pci_ioda1_setup_dma_pe(struct pnv_phb *phb,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
+ /* Setup DMA32 segment mapping */
|
|
|
+ for (i = base; i < base + segs; i++)
|
|
|
+ phb->ioda.dma32_segmap[i] = pe->pe_number;
|
|
|
+
|
|
|
/* Setup linux iommu table */
|
|
|
pnv_pci_setup_iommu_table(tbl, addr, tce32_segsz * segs,
|
|
|
base * PNV_IODA1_DMA32_SEGSIZE,
|
|
@@ -2542,73 +2581,34 @@ static void pnv_pci_ioda2_setup_dma_pe(struct pnv_phb *phb,
|
|
|
static void pnv_ioda_setup_dma(struct pnv_phb *phb)
|
|
|
{
|
|
|
struct pci_controller *hose = phb->hose;
|
|
|
- unsigned int weight, total_weight, dma_pe_count;
|
|
|
- unsigned int residual, remaining, segs, base;
|
|
|
struct pnv_ioda_pe *pe;
|
|
|
-
|
|
|
- total_weight = 0;
|
|
|
- pci_walk_bus(phb->hose->bus, pnv_pci_ioda_dev_dma_weight,
|
|
|
- &total_weight);
|
|
|
-
|
|
|
- dma_pe_count = 0;
|
|
|
- list_for_each_entry(pe, &phb->ioda.pe_list, list) {
|
|
|
- weight = pnv_pci_ioda_pe_dma_weight(pe);
|
|
|
- if (weight > 0)
|
|
|
- dma_pe_count++;
|
|
|
- }
|
|
|
+ unsigned int weight;
|
|
|
|
|
|
/* If we have more PE# than segments available, hand out one
|
|
|
* per PE until we run out and let the rest fail. If not,
|
|
|
* then we assign at least one segment per PE, plus more based
|
|
|
* on the amount of devices under that PE
|
|
|
*/
|
|
|
- if (dma_pe_count > phb->ioda.tce32_count)
|
|
|
- residual = 0;
|
|
|
- else
|
|
|
- residual = phb->ioda.tce32_count - dma_pe_count;
|
|
|
-
|
|
|
- pr_info("PCI: Domain %04x has %ld available 32-bit DMA segments\n",
|
|
|
- hose->global_number, phb->ioda.tce32_count);
|
|
|
- pr_info("PCI: %d PE# for a total weight of %d\n",
|
|
|
- dma_pe_count, total_weight);
|
|
|
+ pr_info("PCI: Domain %04x has %d available 32-bit DMA segments\n",
|
|
|
+ hose->global_number, phb->ioda.dma32_count);
|
|
|
|
|
|
pnv_pci_ioda_setup_opal_tce_kill(phb);
|
|
|
|
|
|
- /* Walk our PE list and configure their DMA segments, hand them
|
|
|
- * out one base segment plus any residual segments based on
|
|
|
- * weight
|
|
|
- */
|
|
|
- remaining = phb->ioda.tce32_count;
|
|
|
- base = 0;
|
|
|
+ /* Walk our PE list and configure their DMA segments */
|
|
|
list_for_each_entry(pe, &phb->ioda.pe_list, list) {
|
|
|
weight = pnv_pci_ioda_pe_dma_weight(pe);
|
|
|
if (!weight)
|
|
|
continue;
|
|
|
|
|
|
- if (!remaining) {
|
|
|
- pe_warn(pe, "No DMA32 resources available\n");
|
|
|
- continue;
|
|
|
- }
|
|
|
- segs = 1;
|
|
|
- if (residual) {
|
|
|
- segs += ((weight * residual) + (total_weight / 2)) /
|
|
|
- total_weight;
|
|
|
- if (segs > remaining)
|
|
|
- segs = remaining;
|
|
|
- }
|
|
|
-
|
|
|
/*
|
|
|
* For IODA2 compliant PHB3, we needn't care about the weight.
|
|
|
* The all available 32-bits DMA space will be assigned to
|
|
|
* the specific PE.
|
|
|
*/
|
|
|
if (phb->type == PNV_PHB_IODA1) {
|
|
|
- pe_info(pe, "DMA weight %d, assigned %d DMA32 segments\n",
|
|
|
- weight, segs);
|
|
|
- pnv_pci_ioda1_setup_dma_pe(phb, pe, base, segs);
|
|
|
+ pnv_pci_ioda1_setup_dma_pe(phb, pe);
|
|
|
} else if (phb->type == PNV_PHB_IODA2) {
|
|
|
pe_info(pe, "Assign DMA32 space\n");
|
|
|
- segs = 0;
|
|
|
pnv_pci_ioda2_setup_dma_pe(phb, pe);
|
|
|
} else if (phb->type == PNV_PHB_NPU) {
|
|
|
/*
|
|
@@ -2618,9 +2618,6 @@ static void pnv_ioda_setup_dma(struct pnv_phb *phb)
|
|
|
* as the PHB3 TVT.
|
|
|
*/
|
|
|
}
|
|
|
-
|
|
|
- remaining -= segs;
|
|
|
- base += segs;
|
|
|
}
|
|
|
}
|
|
|
|
|
@@ -3327,7 +3324,8 @@ static void __init pnv_pci_init_ioda_phb(struct device_node *np,
|
|
|
{
|
|
|
struct pci_controller *hose;
|
|
|
struct pnv_phb *phb;
|
|
|
- unsigned long size, m64map_off, m32map_off, pemap_off, iomap_off = 0;
|
|
|
+ unsigned long size, m64map_off, m32map_off, pemap_off;
|
|
|
+ unsigned long iomap_off = 0, dma32map_off = 0;
|
|
|
const __be64 *prop64;
|
|
|
const __be32 *prop32;
|
|
|
int len;
|
|
@@ -3413,6 +3411,10 @@ static void __init pnv_pci_init_ioda_phb(struct device_node *np,
|
|
|
phb->ioda.io_segsize = phb->ioda.io_size / phb->ioda.total_pe_num;
|
|
|
phb->ioda.io_pci_base = 0; /* XXX calculate this ? */
|
|
|
|
|
|
+ /* Calculate how many 32-bit TCE segments we have */
|
|
|
+ phb->ioda.dma32_count = phb->ioda.m32_pci_base /
|
|
|
+ PNV_IODA1_DMA32_SEGSIZE;
|
|
|
+
|
|
|
/* Allocate aux data & arrays. We don't have IO ports on PHB3 */
|
|
|
size = _ALIGN_UP(phb->ioda.total_pe_num / 8, sizeof(unsigned long));
|
|
|
m64map_off = size;
|
|
@@ -3422,6 +3424,9 @@ static void __init pnv_pci_init_ioda_phb(struct device_node *np,
|
|
|
if (phb->type == PNV_PHB_IODA1) {
|
|
|
iomap_off = size;
|
|
|
size += phb->ioda.total_pe_num * sizeof(phb->ioda.io_segmap[0]);
|
|
|
+ dma32map_off = size;
|
|
|
+ size += phb->ioda.dma32_count *
|
|
|
+ sizeof(phb->ioda.dma32_segmap[0]);
|
|
|
}
|
|
|
pemap_off = size;
|
|
|
size += phb->ioda.total_pe_num * sizeof(struct pnv_ioda_pe);
|
|
@@ -3437,6 +3442,10 @@ static void __init pnv_pci_init_ioda_phb(struct device_node *np,
|
|
|
phb->ioda.io_segmap = aux + iomap_off;
|
|
|
for (segno = 0; segno < phb->ioda.total_pe_num; segno++)
|
|
|
phb->ioda.io_segmap[segno] = IODA_INVALID_PE;
|
|
|
+
|
|
|
+ phb->ioda.dma32_segmap = aux + dma32map_off;
|
|
|
+ for (segno = 0; segno < phb->ioda.dma32_count; segno++)
|
|
|
+ phb->ioda.dma32_segmap[segno] = IODA_INVALID_PE;
|
|
|
}
|
|
|
phb->ioda.pe_array = aux + pemap_off;
|
|
|
set_bit(phb->ioda.reserved_pe_idx, phb->ioda.pe_alloc);
|
|
@@ -3445,7 +3454,7 @@ static void __init pnv_pci_init_ioda_phb(struct device_node *np,
|
|
|
mutex_init(&phb->ioda.pe_list_mutex);
|
|
|
|
|
|
/* Calculate how many 32-bit TCE segments we have */
|
|
|
- phb->ioda.tce32_count = phb->ioda.m32_pci_base /
|
|
|
+ phb->ioda.dma32_count = phb->ioda.m32_pci_base /
|
|
|
PNV_IODA1_DMA32_SEGSIZE;
|
|
|
|
|
|
#if 0 /* We should really do that ... */
|