#include <asm/apic.h>
#include <asm/amd_nb.h>
-static struct bootnode __initdata nodes[8];
static unsigned char __initdata nodeids[8];
static __init int find_northbridge(void)
pr_info("Node %d MemBase %016lx Limit %016lx\n",
nodeid, base, limit);
- nodes[nodeid].start = base;
- nodes[nodeid].end = limit;
+ numa_nodes[nodeid].start = base;
+ numa_nodes[nodeid].end = limit;
prevbase = base;
int i;
for_each_node_mask(i, mem_nodes_parsed) {
- physnodes[i].start = nodes[i].start;
- physnodes[i].end = nodes[i].end;
+ physnodes[i].start = numa_nodes[i].start;
+ physnodes[i].end = numa_nodes[i].end;
}
}
int i;
for (i = 0; i < 8; i++)
- if (addr >= nodes[i].start && addr < nodes[i].end) {
+ if (addr >= numa_nodes[i].start && addr < numa_nodes[i].end) {
ret = i;
break;
}
{
int i;
- memnode_shift = compute_hash_shift(nodes, 8, NULL);
+ memnode_shift = compute_hash_shift(numa_nodes, 8, NULL);
if (memnode_shift < 0) {
pr_err("No NUMA node hash function found. Contact maintainer\n");
return -1;
/* use the coreid bits from early_identify_cpu */
for_each_node_mask(i, node_possible_map)
memblock_x86_register_active_regions(i,
- nodes[i].start >> PAGE_SHIFT,
- nodes[i].end >> PAGE_SHIFT);
+ numa_nodes[i].start >> PAGE_SHIFT,
+ numa_nodes[i].end >> PAGE_SHIFT);
init_memory_mapping_high();
for_each_node_mask(i, node_possible_map)
- setup_node_bootmem(i, nodes[i].start, nodes[i].end);
+ setup_node_bootmem(i, numa_nodes[i].start, numa_nodes[i].end);
numa_init_array();
return 0;
static unsigned long __initdata nodemap_addr;
static unsigned long __initdata nodemap_size;
+struct bootnode numa_nodes[MAX_NUMNODES] __initdata;
+
/*
* Given a shift value, try to populate memnodemap[]
* Returns :
return NULL;
}
+static __init void cutoff_node(int i, unsigned long start, unsigned long end)
+{
+ struct bootnode *nd = &numa_nodes[i];
+
+ if (nd->start < start) {
+ nd->start = start;
+ if (nd->end < nd->start)
+ nd->start = nd->end;
+ }
+ if (nd->end > end) {
+ nd->end = end;
+ if (nd->start > nd->end)
+ nd->start = nd->end;
+ }
+}
+
/* Initialize bootmem allocator for a node */
void __init
setup_node_bootmem(int nodeid, unsigned long start, unsigned long end)
nodes_clear(mem_nodes_parsed);
nodes_clear(node_possible_map);
nodes_clear(node_online_map);
+ memset(numa_nodes, 0, sizeof(numa_nodes));
if (numa_init[i]() < 0)
continue;
+
+ /* clean up the node list */
+ for (j = 0; j < MAX_NUMNODES; j++)
+ cutoff_node(j, 0, max_pfn << PAGE_SHIFT);
+
#ifdef CONFIG_NUMA_EMU
setup_physnodes(0, max_pfn << PAGE_SHIFT, i == 0, i == 1);
if (cmdline && !numa_emulation(0, max_pfn, i == 0, i == 1))
static struct acpi_table_slit *acpi_slit;
-static struct bootnode nodes[MAX_NUMNODES] __initdata;
static struct bootnode nodes_add[MAX_NUMNODES];
static int num_node_memblks __initdata;
return -1;
}
-static __init void cutoff_node(int i, unsigned long start, unsigned long end)
-{
- struct bootnode *nd = &nodes[i];
-
- if (nd->start < start) {
- nd->start = start;
- if (nd->end < nd->start)
- nd->start = nd->end;
- }
- if (nd->end > end) {
- nd->end = end;
- if (nd->start > nd->end)
- nd->start = nd->end;
- }
-}
-
static __init void bad_srat(void)
{
int i;
printk(KERN_ERR "SRAT: SRAT not used.\n");
acpi_numa = -1;
for (i = 0; i < MAX_NUMNODES; i++) {
- nodes[i].start = nodes[i].end = 0;
+ numa_nodes[i].start = numa_nodes[i].end = 0;
nodes_add[i].start = nodes_add[i].end = 0;
}
remove_all_active_ranges();
if (i == node) {
printk(KERN_WARNING
"SRAT: Warning: PXM %d (%lx-%lx) overlaps with itself (%Lx-%Lx)\n",
- pxm, start, end, nodes[i].start, nodes[i].end);
+ pxm, start, end, numa_nodes[i].start, numa_nodes[i].end);
} else if (i >= 0) {
printk(KERN_ERR
"SRAT: PXM %d (%lx-%lx) overlaps with PXM %d (%Lx-%Lx)\n",
pxm, start, end, node_to_pxm(i),
- nodes[i].start, nodes[i].end);
+ numa_nodes[i].start, numa_nodes[i].end);
bad_srat();
return;
}
start, end);
if (!(ma->flags & ACPI_SRAT_MEM_HOT_PLUGGABLE)) {
- nd = &nodes[node];
+ nd = &numa_nodes[node];
if (!node_test_and_set(node, mem_nodes_parsed)) {
nd->start = start;
nd->end = end;
int i;
for_each_node_mask(i, mem_nodes_parsed) {
- cutoff_node(i, start, end);
- physnodes[i].start = nodes[i].start;
- physnodes[i].end = nodes[i].end;
+ physnodes[i].start = numa_nodes[i].start;
+ physnodes[i].end = numa_nodes[i].end;
}
}
#endif /* CONFIG_NUMA_EMU */
if (acpi_numa <= 0)
return -1;
- /* First clean up the node list */
- for (i = 0; i < MAX_NUMNODES; i++)
- cutoff_node(i, 0, max_pfn << PAGE_SHIFT);
-
/*
* Join together blocks on the same node, holes between
* which don't overlap with memory on other nodes.
/* for out of order entries in SRAT */
sort_node_map();
- if (!nodes_cover_memory(nodes)) {
+ if (!nodes_cover_memory(numa_nodes)) {
bad_srat();
return -1;
}
/* Finally register nodes */
for_each_node_mask(i, node_possible_map)
- setup_node_bootmem(i, nodes[i].start, nodes[i].end);
+ setup_node_bootmem(i, numa_nodes[i].start, numa_nodes[i].end);
/* Try again in case setup_node_bootmem missed one due
to missing bootmem */
for_each_node_mask(i, node_possible_map)
if (!node_online(i))
- setup_node_bootmem(i, nodes[i].start, nodes[i].end);
+ setup_node_bootmem(i, numa_nodes[i].start,
+ numa_nodes[i].end);
for (i = 0; i < nr_cpu_ids; i++) {
int node = early_cpu_to_node(i);
* the sake of simplicity, we only use a real node's starting
* address to determine which emulated node it appears on.
*/
- if (addr >= nodes[i].start && addr < nodes[i].end) {
+ if (addr >= numa_nodes[i].start && addr < numa_nodes[i].end) {
ret = i;
break;
}