|
@@ -58,7 +58,7 @@ static int form1_affinity;
|
|
|
|
|
|
#define MAX_DISTANCE_REF_POINTS 4
|
|
#define MAX_DISTANCE_REF_POINTS 4
|
|
static int distance_ref_points_depth;
|
|
static int distance_ref_points_depth;
|
|
-static const unsigned int *distance_ref_points;
|
|
|
|
|
|
+static const __be32 *distance_ref_points;
|
|
static int distance_lookup_table[MAX_NUMNODES][MAX_DISTANCE_REF_POINTS];
|
|
static int distance_lookup_table[MAX_NUMNODES][MAX_DISTANCE_REF_POINTS];
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -179,7 +179,7 @@ static void unmap_cpu_from_node(unsigned long cpu)
|
|
#endif /* CONFIG_HOTPLUG_CPU || CONFIG_PPC_SPLPAR */
|
|
#endif /* CONFIG_HOTPLUG_CPU || CONFIG_PPC_SPLPAR */
|
|
|
|
|
|
/* must hold reference to node during call */
|
|
/* must hold reference to node during call */
|
|
-static const int *of_get_associativity(struct device_node *dev)
|
|
|
|
|
|
+static const __be32 *of_get_associativity(struct device_node *dev)
|
|
{
|
|
{
|
|
return of_get_property(dev, "ibm,associativity", NULL);
|
|
return of_get_property(dev, "ibm,associativity", NULL);
|
|
}
|
|
}
|
|
@@ -189,9 +189,9 @@ static const int *of_get_associativity(struct device_node *dev)
|
|
* it exists (the property exists only in kexec/kdump kernels,
|
|
* it exists (the property exists only in kexec/kdump kernels,
|
|
* added by kexec-tools)
|
|
* added by kexec-tools)
|
|
*/
|
|
*/
|
|
-static const u32 *of_get_usable_memory(struct device_node *memory)
|
|
|
|
|
|
+static const __be32 *of_get_usable_memory(struct device_node *memory)
|
|
{
|
|
{
|
|
- const u32 *prop;
|
|
|
|
|
|
+ const __be32 *prop;
|
|
u32 len;
|
|
u32 len;
|
|
prop = of_get_property(memory, "linux,drconf-usable-memory", &len);
|
|
prop = of_get_property(memory, "linux,drconf-usable-memory", &len);
|
|
if (!prop || len < sizeof(unsigned int))
|
|
if (!prop || len < sizeof(unsigned int))
|
|
@@ -219,7 +219,7 @@ int __node_distance(int a, int b)
|
|
}
|
|
}
|
|
|
|
|
|
static void initialize_distance_lookup_table(int nid,
|
|
static void initialize_distance_lookup_table(int nid,
|
|
- const unsigned int *associativity)
|
|
|
|
|
|
+ const __be32 *associativity)
|
|
{
|
|
{
|
|
int i;
|
|
int i;
|
|
|
|
|
|
@@ -227,29 +227,32 @@ static void initialize_distance_lookup_table(int nid,
|
|
return;
|
|
return;
|
|
|
|
|
|
for (i = 0; i < distance_ref_points_depth; i++) {
|
|
for (i = 0; i < distance_ref_points_depth; i++) {
|
|
- distance_lookup_table[nid][i] =
|
|
|
|
- associativity[distance_ref_points[i]];
|
|
|
|
|
|
+ const __be32 *entry;
|
|
|
|
+
|
|
|
|
+ entry = &associativity[be32_to_cpu(distance_ref_points[i])];
|
|
|
|
+ distance_lookup_table[nid][i] = of_read_number(entry, 1);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/* Returns nid in the range [0..MAX_NUMNODES-1], or -1 if no useful numa
|
|
/* Returns nid in the range [0..MAX_NUMNODES-1], or -1 if no useful numa
|
|
* info is found.
|
|
* info is found.
|
|
*/
|
|
*/
|
|
-static int associativity_to_nid(const unsigned int *associativity)
|
|
|
|
|
|
+static int associativity_to_nid(const __be32 *associativity)
|
|
{
|
|
{
|
|
int nid = -1;
|
|
int nid = -1;
|
|
|
|
|
|
if (min_common_depth == -1)
|
|
if (min_common_depth == -1)
|
|
goto out;
|
|
goto out;
|
|
|
|
|
|
- if (associativity[0] >= min_common_depth)
|
|
|
|
- nid = associativity[min_common_depth];
|
|
|
|
|
|
+ if (of_read_number(associativity, 1) >= min_common_depth)
|
|
|
|
+ nid = of_read_number(&associativity[min_common_depth], 1);
|
|
|
|
|
|
/* POWER4 LPAR uses 0xffff as invalid node */
|
|
/* POWER4 LPAR uses 0xffff as invalid node */
|
|
if (nid == 0xffff || nid >= MAX_NUMNODES)
|
|
if (nid == 0xffff || nid >= MAX_NUMNODES)
|
|
nid = -1;
|
|
nid = -1;
|
|
|
|
|
|
- if (nid > 0 && associativity[0] >= distance_ref_points_depth)
|
|
|
|
|
|
+ if (nid > 0 &&
|
|
|
|
+ of_read_number(associativity, 1) >= distance_ref_points_depth)
|
|
initialize_distance_lookup_table(nid, associativity);
|
|
initialize_distance_lookup_table(nid, associativity);
|
|
|
|
|
|
out:
|
|
out:
|
|
@@ -262,7 +265,7 @@ out:
|
|
static int of_node_to_nid_single(struct device_node *device)
|
|
static int of_node_to_nid_single(struct device_node *device)
|
|
{
|
|
{
|
|
int nid = -1;
|
|
int nid = -1;
|
|
- const unsigned int *tmp;
|
|
|
|
|
|
+ const __be32 *tmp;
|
|
|
|
|
|
tmp = of_get_associativity(device);
|
|
tmp = of_get_associativity(device);
|
|
if (tmp)
|
|
if (tmp)
|
|
@@ -334,7 +337,7 @@ static int __init find_min_common_depth(void)
|
|
}
|
|
}
|
|
|
|
|
|
if (form1_affinity) {
|
|
if (form1_affinity) {
|
|
- depth = distance_ref_points[0];
|
|
|
|
|
|
+ depth = of_read_number(distance_ref_points, 1);
|
|
} else {
|
|
} else {
|
|
if (distance_ref_points_depth < 2) {
|
|
if (distance_ref_points_depth < 2) {
|
|
printk(KERN_WARNING "NUMA: "
|
|
printk(KERN_WARNING "NUMA: "
|
|
@@ -342,7 +345,7 @@ static int __init find_min_common_depth(void)
|
|
goto err;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
|
|
- depth = distance_ref_points[1];
|
|
|
|
|
|
+ depth = of_read_number(&distance_ref_points[1], 1);
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -376,12 +379,12 @@ static void __init get_n_mem_cells(int *n_addr_cells, int *n_size_cells)
|
|
of_node_put(memory);
|
|
of_node_put(memory);
|
|
}
|
|
}
|
|
|
|
|
|
-static unsigned long read_n_cells(int n, const unsigned int **buf)
|
|
|
|
|
|
+static unsigned long read_n_cells(int n, const __be32 **buf)
|
|
{
|
|
{
|
|
unsigned long result = 0;
|
|
unsigned long result = 0;
|
|
|
|
|
|
while (n--) {
|
|
while (n--) {
|
|
- result = (result << 32) | **buf;
|
|
|
|
|
|
+ result = (result << 32) | of_read_number(*buf, 1);
|
|
(*buf)++;
|
|
(*buf)++;
|
|
}
|
|
}
|
|
return result;
|
|
return result;
|
|
@@ -391,17 +394,17 @@ static unsigned long read_n_cells(int n, const unsigned int **buf)
|
|
* Read the next memblock list entry from the ibm,dynamic-memory property
|
|
* Read the next memblock list entry from the ibm,dynamic-memory property
|
|
* and return the information in the provided of_drconf_cell structure.
|
|
* and return the information in the provided of_drconf_cell structure.
|
|
*/
|
|
*/
|
|
-static void read_drconf_cell(struct of_drconf_cell *drmem, const u32 **cellp)
|
|
|
|
|
|
+static void read_drconf_cell(struct of_drconf_cell *drmem, const __be32 **cellp)
|
|
{
|
|
{
|
|
- const u32 *cp;
|
|
|
|
|
|
+ const __be32 *cp;
|
|
|
|
|
|
drmem->base_addr = read_n_cells(n_mem_addr_cells, cellp);
|
|
drmem->base_addr = read_n_cells(n_mem_addr_cells, cellp);
|
|
|
|
|
|
cp = *cellp;
|
|
cp = *cellp;
|
|
- drmem->drc_index = cp[0];
|
|
|
|
- drmem->reserved = cp[1];
|
|
|
|
- drmem->aa_index = cp[2];
|
|
|
|
- drmem->flags = cp[3];
|
|
|
|
|
|
+ drmem->drc_index = of_read_number(cp, 1);
|
|
|
|
+ drmem->reserved = of_read_number(&cp[1], 1);
|
|
|
|
+ drmem->aa_index = of_read_number(&cp[2], 1);
|
|
|
|
+ drmem->flags = of_read_number(&cp[3], 1);
|
|
|
|
|
|
*cellp = cp + 4;
|
|
*cellp = cp + 4;
|
|
}
|
|
}
|
|
@@ -413,16 +416,16 @@ static void read_drconf_cell(struct of_drconf_cell *drmem, const u32 **cellp)
|
|
* list entries followed by N memblock list entries. Each memblock list entry
|
|
* list entries followed by N memblock list entries. Each memblock list entry
|
|
* contains information as laid out in the of_drconf_cell struct above.
|
|
* contains information as laid out in the of_drconf_cell struct above.
|
|
*/
|
|
*/
|
|
-static int of_get_drconf_memory(struct device_node *memory, const u32 **dm)
|
|
|
|
|
|
+static int of_get_drconf_memory(struct device_node *memory, const __be32 **dm)
|
|
{
|
|
{
|
|
- const u32 *prop;
|
|
|
|
|
|
+ const __be32 *prop;
|
|
u32 len, entries;
|
|
u32 len, entries;
|
|
|
|
|
|
prop = of_get_property(memory, "ibm,dynamic-memory", &len);
|
|
prop = of_get_property(memory, "ibm,dynamic-memory", &len);
|
|
if (!prop || len < sizeof(unsigned int))
|
|
if (!prop || len < sizeof(unsigned int))
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
- entries = *prop++;
|
|
|
|
|
|
+ entries = of_read_number(prop++, 1);
|
|
|
|
|
|
/* Now that we know the number of entries, revalidate the size
|
|
/* Now that we know the number of entries, revalidate the size
|
|
* of the property read in to ensure we have everything
|
|
* of the property read in to ensure we have everything
|
|
@@ -440,7 +443,7 @@ static int of_get_drconf_memory(struct device_node *memory, const u32 **dm)
|
|
*/
|
|
*/
|
|
static u64 of_get_lmb_size(struct device_node *memory)
|
|
static u64 of_get_lmb_size(struct device_node *memory)
|
|
{
|
|
{
|
|
- const u32 *prop;
|
|
|
|
|
|
+ const __be32 *prop;
|
|
u32 len;
|
|
u32 len;
|
|
|
|
|
|
prop = of_get_property(memory, "ibm,lmb-size", &len);
|
|
prop = of_get_property(memory, "ibm,lmb-size", &len);
|
|
@@ -453,7 +456,7 @@ static u64 of_get_lmb_size(struct device_node *memory)
|
|
struct assoc_arrays {
|
|
struct assoc_arrays {
|
|
u32 n_arrays;
|
|
u32 n_arrays;
|
|
u32 array_sz;
|
|
u32 array_sz;
|
|
- const u32 *arrays;
|
|
|
|
|
|
+ const __be32 *arrays;
|
|
};
|
|
};
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -469,15 +472,15 @@ struct assoc_arrays {
|
|
static int of_get_assoc_arrays(struct device_node *memory,
|
|
static int of_get_assoc_arrays(struct device_node *memory,
|
|
struct assoc_arrays *aa)
|
|
struct assoc_arrays *aa)
|
|
{
|
|
{
|
|
- const u32 *prop;
|
|
|
|
|
|
+ const __be32 *prop;
|
|
u32 len;
|
|
u32 len;
|
|
|
|
|
|
prop = of_get_property(memory, "ibm,associativity-lookup-arrays", &len);
|
|
prop = of_get_property(memory, "ibm,associativity-lookup-arrays", &len);
|
|
if (!prop || len < 2 * sizeof(unsigned int))
|
|
if (!prop || len < 2 * sizeof(unsigned int))
|
|
return -1;
|
|
return -1;
|
|
|
|
|
|
- aa->n_arrays = *prop++;
|
|
|
|
- aa->array_sz = *prop++;
|
|
|
|
|
|
+ aa->n_arrays = of_read_number(prop++, 1);
|
|
|
|
+ aa->array_sz = of_read_number(prop++, 1);
|
|
|
|
|
|
/* Now that we know the number of arrays and size of each array,
|
|
/* Now that we know the number of arrays and size of each array,
|
|
* revalidate the size of the property read in.
|
|
* revalidate the size of the property read in.
|
|
@@ -504,7 +507,7 @@ static int of_drconf_to_nid_single(struct of_drconf_cell *drmem,
|
|
!(drmem->flags & DRCONF_MEM_AI_INVALID) &&
|
|
!(drmem->flags & DRCONF_MEM_AI_INVALID) &&
|
|
drmem->aa_index < aa->n_arrays) {
|
|
drmem->aa_index < aa->n_arrays) {
|
|
index = drmem->aa_index * aa->array_sz + min_common_depth - 1;
|
|
index = drmem->aa_index * aa->array_sz + min_common_depth - 1;
|
|
- nid = aa->arrays[index];
|
|
|
|
|
|
+ nid = of_read_number(&aa->arrays[index], 1);
|
|
|
|
|
|
if (nid == 0xffff || nid >= MAX_NUMNODES)
|
|
if (nid == 0xffff || nid >= MAX_NUMNODES)
|
|
nid = default_nid;
|
|
nid = default_nid;
|
|
@@ -595,7 +598,7 @@ static unsigned long __init numa_enforce_memory_limit(unsigned long start,
|
|
* Reads the counter for a given entry in
|
|
* Reads the counter for a given entry in
|
|
* linux,drconf-usable-memory property
|
|
* linux,drconf-usable-memory property
|
|
*/
|
|
*/
|
|
-static inline int __init read_usm_ranges(const u32 **usm)
|
|
|
|
|
|
+static inline int __init read_usm_ranges(const __be32 **usm)
|
|
{
|
|
{
|
|
/*
|
|
/*
|
|
* For each lmb in ibm,dynamic-memory a corresponding
|
|
* For each lmb in ibm,dynamic-memory a corresponding
|
|
@@ -612,7 +615,7 @@ static inline int __init read_usm_ranges(const u32 **usm)
|
|
*/
|
|
*/
|
|
static void __init parse_drconf_memory(struct device_node *memory)
|
|
static void __init parse_drconf_memory(struct device_node *memory)
|
|
{
|
|
{
|
|
- const u32 *uninitialized_var(dm), *usm;
|
|
|
|
|
|
+ const __be32 *uninitialized_var(dm), *usm;
|
|
unsigned int n, rc, ranges, is_kexec_kdump = 0;
|
|
unsigned int n, rc, ranges, is_kexec_kdump = 0;
|
|
unsigned long lmb_size, base, size, sz;
|
|
unsigned long lmb_size, base, size, sz;
|
|
int nid;
|
|
int nid;
|
|
@@ -721,7 +724,7 @@ static int __init parse_numa_properties(void)
|
|
unsigned long size;
|
|
unsigned long size;
|
|
int nid;
|
|
int nid;
|
|
int ranges;
|
|
int ranges;
|
|
- const unsigned int *memcell_buf;
|
|
|
|
|
|
+ const __be32 *memcell_buf;
|
|
unsigned int len;
|
|
unsigned int len;
|
|
|
|
|
|
memcell_buf = of_get_property(memory,
|
|
memcell_buf = of_get_property(memory,
|
|
@@ -1106,7 +1109,7 @@ early_param("numa", early_numa);
|
|
static int hot_add_drconf_scn_to_nid(struct device_node *memory,
|
|
static int hot_add_drconf_scn_to_nid(struct device_node *memory,
|
|
unsigned long scn_addr)
|
|
unsigned long scn_addr)
|
|
{
|
|
{
|
|
- const u32 *dm;
|
|
|
|
|
|
+ const __be32 *dm;
|
|
unsigned int drconf_cell_cnt, rc;
|
|
unsigned int drconf_cell_cnt, rc;
|
|
unsigned long lmb_size;
|
|
unsigned long lmb_size;
|
|
struct assoc_arrays aa;
|
|
struct assoc_arrays aa;
|
|
@@ -1159,7 +1162,7 @@ int hot_add_node_scn_to_nid(unsigned long scn_addr)
|
|
for_each_node_by_type(memory, "memory") {
|
|
for_each_node_by_type(memory, "memory") {
|
|
unsigned long start, size;
|
|
unsigned long start, size;
|
|
int ranges;
|
|
int ranges;
|
|
- const unsigned int *memcell_buf;
|
|
|
|
|
|
+ const __be32 *memcell_buf;
|
|
unsigned int len;
|
|
unsigned int len;
|
|
|
|
|
|
memcell_buf = of_get_property(memory, "reg", &len);
|
|
memcell_buf = of_get_property(memory, "reg", &len);
|
|
@@ -1232,7 +1235,7 @@ static u64 hot_add_drconf_memory_max(void)
|
|
struct device_node *memory = NULL;
|
|
struct device_node *memory = NULL;
|
|
unsigned int drconf_cell_cnt = 0;
|
|
unsigned int drconf_cell_cnt = 0;
|
|
u64 lmb_size = 0;
|
|
u64 lmb_size = 0;
|
|
- const u32 *dm = 0;
|
|
|
|
|
|
+ const __be32 *dm = 0;
|
|
|
|
|
|
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
|
|
memory = of_find_node_by_path("/ibm,dynamic-reconfiguration-memory");
|
|
if (memory) {
|
|
if (memory) {
|
|
@@ -1337,40 +1340,41 @@ static int update_cpu_associativity_changes_mask(void)
|
|
* Convert the associativity domain numbers returned from the hypervisor
|
|
* Convert the associativity domain numbers returned from the hypervisor
|
|
* to the sequence they would appear in the ibm,associativity property.
|
|
* to the sequence they would appear in the ibm,associativity property.
|
|
*/
|
|
*/
|
|
-static int vphn_unpack_associativity(const long *packed, unsigned int *unpacked)
|
|
|
|
|
|
+static int vphn_unpack_associativity(const long *packed, __be32 *unpacked)
|
|
{
|
|
{
|
|
int i, nr_assoc_doms = 0;
|
|
int i, nr_assoc_doms = 0;
|
|
- const u16 *field = (const u16*) packed;
|
|
|
|
|
|
+ const __be16 *field = (const __be16 *) packed;
|
|
|
|
|
|
#define VPHN_FIELD_UNUSED (0xffff)
|
|
#define VPHN_FIELD_UNUSED (0xffff)
|
|
#define VPHN_FIELD_MSB (0x8000)
|
|
#define VPHN_FIELD_MSB (0x8000)
|
|
#define VPHN_FIELD_MASK (~VPHN_FIELD_MSB)
|
|
#define VPHN_FIELD_MASK (~VPHN_FIELD_MSB)
|
|
|
|
|
|
for (i = 1; i < VPHN_ASSOC_BUFSIZE; i++) {
|
|
for (i = 1; i < VPHN_ASSOC_BUFSIZE; i++) {
|
|
- if (*field == VPHN_FIELD_UNUSED) {
|
|
|
|
|
|
+ if (be16_to_cpup(field) == VPHN_FIELD_UNUSED) {
|
|
/* All significant fields processed, and remaining
|
|
/* All significant fields processed, and remaining
|
|
* fields contain the reserved value of all 1's.
|
|
* fields contain the reserved value of all 1's.
|
|
* Just store them.
|
|
* Just store them.
|
|
*/
|
|
*/
|
|
- unpacked[i] = *((u32*)field);
|
|
|
|
|
|
+ unpacked[i] = *((__be32 *)field);
|
|
field += 2;
|
|
field += 2;
|
|
- } else if (*field & VPHN_FIELD_MSB) {
|
|
|
|
|
|
+ } else if (be16_to_cpup(field) & VPHN_FIELD_MSB) {
|
|
/* Data is in the lower 15 bits of this field */
|
|
/* Data is in the lower 15 bits of this field */
|
|
- unpacked[i] = *field & VPHN_FIELD_MASK;
|
|
|
|
|
|
+ unpacked[i] = cpu_to_be32(
|
|
|
|
+ be16_to_cpup(field) & VPHN_FIELD_MASK);
|
|
field++;
|
|
field++;
|
|
nr_assoc_doms++;
|
|
nr_assoc_doms++;
|
|
} else {
|
|
} else {
|
|
/* Data is in the lower 15 bits of this field
|
|
/* Data is in the lower 15 bits of this field
|
|
* concatenated with the next 16 bit field
|
|
* concatenated with the next 16 bit field
|
|
*/
|
|
*/
|
|
- unpacked[i] = *((u32*)field);
|
|
|
|
|
|
+ unpacked[i] = *((__be32 *)field);
|
|
field += 2;
|
|
field += 2;
|
|
nr_assoc_doms++;
|
|
nr_assoc_doms++;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/* The first cell contains the length of the property */
|
|
/* The first cell contains the length of the property */
|
|
- unpacked[0] = nr_assoc_doms;
|
|
|
|
|
|
+ unpacked[0] = cpu_to_be32(nr_assoc_doms);
|
|
|
|
|
|
return nr_assoc_doms;
|
|
return nr_assoc_doms;
|
|
}
|
|
}
|
|
@@ -1379,7 +1383,7 @@ static int vphn_unpack_associativity(const long *packed, unsigned int *unpacked)
|
|
* Retrieve the new associativity information for a virtual processor's
|
|
* Retrieve the new associativity information for a virtual processor's
|
|
* home node.
|
|
* home node.
|
|
*/
|
|
*/
|
|
-static long hcall_vphn(unsigned long cpu, unsigned int *associativity)
|
|
|
|
|
|
+static long hcall_vphn(unsigned long cpu, __be32 *associativity)
|
|
{
|
|
{
|
|
long rc;
|
|
long rc;
|
|
long retbuf[PLPAR_HCALL9_BUFSIZE] = {0};
|
|
long retbuf[PLPAR_HCALL9_BUFSIZE] = {0};
|
|
@@ -1393,7 +1397,7 @@ static long hcall_vphn(unsigned long cpu, unsigned int *associativity)
|
|
}
|
|
}
|
|
|
|
|
|
static long vphn_get_associativity(unsigned long cpu,
|
|
static long vphn_get_associativity(unsigned long cpu,
|
|
- unsigned int *associativity)
|
|
|
|
|
|
+ __be32 *associativity)
|
|
{
|
|
{
|
|
long rc;
|
|
long rc;
|
|
|
|
|
|
@@ -1450,7 +1454,7 @@ int arch_update_cpu_topology(void)
|
|
{
|
|
{
|
|
unsigned int cpu, sibling, changed = 0;
|
|
unsigned int cpu, sibling, changed = 0;
|
|
struct topology_update_data *updates, *ud;
|
|
struct topology_update_data *updates, *ud;
|
|
- unsigned int associativity[VPHN_ASSOC_BUFSIZE] = {0};
|
|
|
|
|
|
+ __be32 associativity[VPHN_ASSOC_BUFSIZE] = {0};
|
|
cpumask_t updated_cpus;
|
|
cpumask_t updated_cpus;
|
|
struct device *dev;
|
|
struct device *dev;
|
|
int weight, new_nid, i = 0;
|
|
int weight, new_nid, i = 0;
|