x86-64, NUMA: Factor out memblk handling into numa_{add|register}_memblk()
Factor out memblk handling from srat_64.c into two functions in numa_64.c. This patch doesn't introduce any behavior change. The next patch will make all init methods use these functions. - v2: Fixed build failure on 32bit due to misplaced NR_NODE_MEMBLKS. Reported by Ingo. Signed-off-by: Tejun Heo <tj@kernel.org> Cc: Yinghai Lu <yinghai@kernel.org> Cc: Brian Gerst <brgerst@gmail.com> Cc: Cyrill Gorcunov <gorcunov@gmail.com> Cc: Shaohui Zheng <shaohui.zheng@intel.com> Cc: David Rientjes <rientjes@google.com> Cc: Ingo Molnar <mingo@elte.hu> Cc: H. Peter Anvin <hpa@linux.intel.com>
This commit is contained in:
@ -187,7 +187,6 @@ struct bootnode;
|
|||||||
extern int acpi_numa;
|
extern int acpi_numa;
|
||||||
extern int x86_acpi_numa_init(void);
|
extern int x86_acpi_numa_init(void);
|
||||||
extern int acpi_scan_nodes(void);
|
extern int acpi_scan_nodes(void);
|
||||||
#define NR_NODE_MEMBLKS (MAX_NUMNODES*2)
|
|
||||||
|
|
||||||
#ifdef CONFIG_NUMA_EMU
|
#ifdef CONFIG_NUMA_EMU
|
||||||
extern void acpi_fake_nodes(const struct bootnode *fake_nodes,
|
extern void acpi_fake_nodes(const struct bootnode *fake_nodes,
|
||||||
|
@ -5,6 +5,9 @@
|
|||||||
#include <asm/apicdef.h>
|
#include <asm/apicdef.h>
|
||||||
|
|
||||||
#ifdef CONFIG_NUMA
|
#ifdef CONFIG_NUMA
|
||||||
|
|
||||||
|
#define NR_NODE_MEMBLKS (MAX_NUMNODES*2)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* __apicid_to_node[] stores the raw mapping between physical apicid and
|
* __apicid_to_node[] stores the raw mapping between physical apicid and
|
||||||
* node and is used to initialize cpu_to_node mapping.
|
* node and is used to initialize cpu_to_node mapping.
|
||||||
|
@ -32,6 +32,8 @@ extern nodemask_t mem_nodes_parsed __initdata;
|
|||||||
extern struct bootnode numa_nodes[MAX_NUMNODES] __initdata;
|
extern struct bootnode numa_nodes[MAX_NUMNODES] __initdata;
|
||||||
|
|
||||||
extern int __cpuinit numa_cpu_node(int cpu);
|
extern int __cpuinit numa_cpu_node(int cpu);
|
||||||
|
extern int __init numa_add_memblk(int nodeid, u64 start, u64 end);
|
||||||
|
extern int __init numa_register_memblks(void);
|
||||||
|
|
||||||
#ifdef CONFIG_NUMA_EMU
|
#ifdef CONFIG_NUMA_EMU
|
||||||
#define FAKE_NODE_MIN_SIZE ((u64)32 << 20)
|
#define FAKE_NODE_MIN_SIZE ((u64)32 << 20)
|
||||||
|
@ -33,6 +33,10 @@ struct memnode memnode;
|
|||||||
static unsigned long __initdata nodemap_addr;
|
static unsigned long __initdata nodemap_addr;
|
||||||
static unsigned long __initdata nodemap_size;
|
static unsigned long __initdata nodemap_size;
|
||||||
|
|
||||||
|
static int num_node_memblks __initdata;
|
||||||
|
static struct bootnode node_memblk_range[NR_NODE_MEMBLKS] __initdata;
|
||||||
|
static int memblk_nodeid[NR_NODE_MEMBLKS] __initdata;
|
||||||
|
|
||||||
struct bootnode numa_nodes[MAX_NUMNODES] __initdata;
|
struct bootnode numa_nodes[MAX_NUMNODES] __initdata;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -184,6 +188,43 @@ static void * __init early_node_mem(int nodeid, unsigned long start,
|
|||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static __init int conflicting_memblks(unsigned long start, unsigned long end)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
for (i = 0; i < num_node_memblks; i++) {
|
||||||
|
struct bootnode *nd = &node_memblk_range[i];
|
||||||
|
if (nd->start == nd->end)
|
||||||
|
continue;
|
||||||
|
if (nd->end > start && nd->start < end)
|
||||||
|
return memblk_nodeid[i];
|
||||||
|
if (nd->end == end && nd->start == start)
|
||||||
|
return memblk_nodeid[i];
|
||||||
|
}
|
||||||
|
return -1;
|
||||||
|
}
|
||||||
|
|
||||||
|
int __init numa_add_memblk(int nid, u64 start, u64 end)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
|
||||||
|
i = conflicting_memblks(start, end);
|
||||||
|
if (i == nid) {
|
||||||
|
printk(KERN_WARNING "NUMA: Warning: node %d (%Lx-%Lx) overlaps with itself (%Lx-%Lx)\n",
|
||||||
|
nid, start, end, numa_nodes[i].start, numa_nodes[i].end);
|
||||||
|
} else if (i >= 0) {
|
||||||
|
printk(KERN_ERR "NUMA: node %d (%Lx-%Lx) overlaps with node %d (%Lx-%Lx)\n",
|
||||||
|
nid, start, end, i,
|
||||||
|
numa_nodes[i].start, numa_nodes[i].end);
|
||||||
|
return -EINVAL;
|
||||||
|
}
|
||||||
|
|
||||||
|
node_memblk_range[num_node_memblks].start = start;
|
||||||
|
node_memblk_range[num_node_memblks].end = end;
|
||||||
|
memblk_nodeid[num_node_memblks] = nid;
|
||||||
|
num_node_memblks++;
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
static __init void cutoff_node(int i, unsigned long start, unsigned long end)
|
static __init void cutoff_node(int i, unsigned long start, unsigned long end)
|
||||||
{
|
{
|
||||||
struct bootnode *nd = &numa_nodes[i];
|
struct bootnode *nd = &numa_nodes[i];
|
||||||
@ -246,6 +287,71 @@ setup_node_bootmem(int nodeid, unsigned long start, unsigned long end)
|
|||||||
node_set_online(nodeid);
|
node_set_online(nodeid);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int __init numa_register_memblks(void)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Join together blocks on the same node, holes between
|
||||||
|
* which don't overlap with memory on other nodes.
|
||||||
|
*/
|
||||||
|
for (i = 0; i < num_node_memblks; ++i) {
|
||||||
|
int j, k;
|
||||||
|
|
||||||
|
for (j = i + 1; j < num_node_memblks; ++j) {
|
||||||
|
unsigned long start, end;
|
||||||
|
|
||||||
|
if (memblk_nodeid[i] != memblk_nodeid[j])
|
||||||
|
continue;
|
||||||
|
start = min(node_memblk_range[i].end,
|
||||||
|
node_memblk_range[j].end);
|
||||||
|
end = max(node_memblk_range[i].start,
|
||||||
|
node_memblk_range[j].start);
|
||||||
|
for (k = 0; k < num_node_memblks; ++k) {
|
||||||
|
if (memblk_nodeid[i] == memblk_nodeid[k])
|
||||||
|
continue;
|
||||||
|
if (start < node_memblk_range[k].end &&
|
||||||
|
end > node_memblk_range[k].start)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
if (k < num_node_memblks)
|
||||||
|
continue;
|
||||||
|
start = min(node_memblk_range[i].start,
|
||||||
|
node_memblk_range[j].start);
|
||||||
|
end = max(node_memblk_range[i].end,
|
||||||
|
node_memblk_range[j].end);
|
||||||
|
printk(KERN_INFO "NUMA: Node %d [%Lx,%Lx) + [%Lx,%Lx) -> [%lx,%lx)\n",
|
||||||
|
memblk_nodeid[i],
|
||||||
|
node_memblk_range[i].start,
|
||||||
|
node_memblk_range[i].end,
|
||||||
|
node_memblk_range[j].start,
|
||||||
|
node_memblk_range[j].end,
|
||||||
|
start, end);
|
||||||
|
node_memblk_range[i].start = start;
|
||||||
|
node_memblk_range[i].end = end;
|
||||||
|
k = --num_node_memblks - j;
|
||||||
|
memmove(memblk_nodeid + j, memblk_nodeid + j+1,
|
||||||
|
k * sizeof(*memblk_nodeid));
|
||||||
|
memmove(node_memblk_range + j, node_memblk_range + j+1,
|
||||||
|
k * sizeof(*node_memblk_range));
|
||||||
|
--j;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
memnode_shift = compute_hash_shift(node_memblk_range, num_node_memblks,
|
||||||
|
memblk_nodeid);
|
||||||
|
if (memnode_shift < 0) {
|
||||||
|
printk(KERN_ERR "NUMA: No NUMA node hash function found. Contact maintainer\n");
|
||||||
|
return -EINVAL;
|
||||||
|
}
|
||||||
|
|
||||||
|
for (i = 0; i < num_node_memblks; i++)
|
||||||
|
memblock_x86_register_active_regions(memblk_nodeid[i],
|
||||||
|
node_memblk_range[i].start >> PAGE_SHIFT,
|
||||||
|
node_memblk_range[i].end >> PAGE_SHIFT);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_NUMA_EMU
|
#ifdef CONFIG_NUMA_EMU
|
||||||
/* Numa emulation */
|
/* Numa emulation */
|
||||||
static struct bootnode nodes[MAX_NUMNODES] __initdata;
|
static struct bootnode nodes[MAX_NUMNODES] __initdata;
|
||||||
@ -653,6 +759,9 @@ void __init initmem_init(void)
|
|||||||
nodes_clear(mem_nodes_parsed);
|
nodes_clear(mem_nodes_parsed);
|
||||||
nodes_clear(node_possible_map);
|
nodes_clear(node_possible_map);
|
||||||
nodes_clear(node_online_map);
|
nodes_clear(node_online_map);
|
||||||
|
num_node_memblks = 0;
|
||||||
|
memset(node_memblk_range, 0, sizeof(node_memblk_range));
|
||||||
|
memset(memblk_nodeid, 0, sizeof(memblk_nodeid));
|
||||||
memset(numa_nodes, 0, sizeof(numa_nodes));
|
memset(numa_nodes, 0, sizeof(numa_nodes));
|
||||||
|
|
||||||
if (numa_init[i]() < 0)
|
if (numa_init[i]() < 0)
|
||||||
|
@ -30,30 +30,11 @@ static struct acpi_table_slit *acpi_slit;
|
|||||||
|
|
||||||
static struct bootnode nodes_add[MAX_NUMNODES];
|
static struct bootnode nodes_add[MAX_NUMNODES];
|
||||||
|
|
||||||
static int num_node_memblks __initdata;
|
|
||||||
static struct bootnode node_memblk_range[NR_NODE_MEMBLKS] __initdata;
|
|
||||||
static int memblk_nodeid[NR_NODE_MEMBLKS] __initdata;
|
|
||||||
|
|
||||||
static __init int setup_node(int pxm)
|
static __init int setup_node(int pxm)
|
||||||
{
|
{
|
||||||
return acpi_map_pxm_to_node(pxm);
|
return acpi_map_pxm_to_node(pxm);
|
||||||
}
|
}
|
||||||
|
|
||||||
static __init int conflicting_memblks(unsigned long start, unsigned long end)
|
|
||||||
{
|
|
||||||
int i;
|
|
||||||
for (i = 0; i < num_node_memblks; i++) {
|
|
||||||
struct bootnode *nd = &node_memblk_range[i];
|
|
||||||
if (nd->start == nd->end)
|
|
||||||
continue;
|
|
||||||
if (nd->end > start && nd->start < end)
|
|
||||||
return memblk_nodeid[i];
|
|
||||||
if (nd->end == end && nd->start == start)
|
|
||||||
return memblk_nodeid[i];
|
|
||||||
}
|
|
||||||
return -1;
|
|
||||||
}
|
|
||||||
|
|
||||||
static __init void bad_srat(void)
|
static __init void bad_srat(void)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
@ -233,7 +214,6 @@ acpi_numa_memory_affinity_init(struct acpi_srat_mem_affinity *ma)
|
|||||||
struct bootnode *nd;
|
struct bootnode *nd;
|
||||||
unsigned long start, end;
|
unsigned long start, end;
|
||||||
int node, pxm;
|
int node, pxm;
|
||||||
int i;
|
|
||||||
|
|
||||||
if (srat_disabled())
|
if (srat_disabled())
|
||||||
return;
|
return;
|
||||||
@ -255,16 +235,8 @@ acpi_numa_memory_affinity_init(struct acpi_srat_mem_affinity *ma)
|
|||||||
bad_srat();
|
bad_srat();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
i = conflicting_memblks(start, end);
|
|
||||||
if (i == node) {
|
if (numa_add_memblk(node, start, end) < 0) {
|
||||||
printk(KERN_WARNING
|
|
||||||
"SRAT: Warning: PXM %d (%lx-%lx) overlaps with itself (%Lx-%Lx)\n",
|
|
||||||
pxm, start, end, numa_nodes[i].start, numa_nodes[i].end);
|
|
||||||
} else if (i >= 0) {
|
|
||||||
printk(KERN_ERR
|
|
||||||
"SRAT: PXM %d (%lx-%lx) overlaps with PXM %d (%Lx-%Lx)\n",
|
|
||||||
pxm, start, end, node_to_pxm(i),
|
|
||||||
numa_nodes[i].start, numa_nodes[i].end);
|
|
||||||
bad_srat();
|
bad_srat();
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@ -285,11 +257,6 @@ acpi_numa_memory_affinity_init(struct acpi_srat_mem_affinity *ma)
|
|||||||
}
|
}
|
||||||
} else
|
} else
|
||||||
update_nodes_add(node, start, end);
|
update_nodes_add(node, start, end);
|
||||||
|
|
||||||
node_memblk_range[num_node_memblks].start = start;
|
|
||||||
node_memblk_range[num_node_memblks].end = end;
|
|
||||||
memblk_nodeid[num_node_memblks] = node;
|
|
||||||
num_node_memblks++;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Sanity check to catch more bad SRATs (they are amazingly common).
|
/* Sanity check to catch more bad SRATs (they are amazingly common).
|
||||||
@ -341,68 +308,11 @@ int __init acpi_scan_nodes(void)
|
|||||||
if (acpi_numa <= 0)
|
if (acpi_numa <= 0)
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
/*
|
if (numa_register_memblks() < 0) {
|
||||||
* Join together blocks on the same node, holes between
|
|
||||||
* which don't overlap with memory on other nodes.
|
|
||||||
*/
|
|
||||||
for (i = 0; i < num_node_memblks; ++i) {
|
|
||||||
int j, k;
|
|
||||||
|
|
||||||
for (j = i + 1; j < num_node_memblks; ++j) {
|
|
||||||
unsigned long start, end;
|
|
||||||
|
|
||||||
if (memblk_nodeid[i] != memblk_nodeid[j])
|
|
||||||
continue;
|
|
||||||
start = min(node_memblk_range[i].end,
|
|
||||||
node_memblk_range[j].end);
|
|
||||||
end = max(node_memblk_range[i].start,
|
|
||||||
node_memblk_range[j].start);
|
|
||||||
for (k = 0; k < num_node_memblks; ++k) {
|
|
||||||
if (memblk_nodeid[i] == memblk_nodeid[k])
|
|
||||||
continue;
|
|
||||||
if (start < node_memblk_range[k].end &&
|
|
||||||
end > node_memblk_range[k].start)
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
if (k < num_node_memblks)
|
|
||||||
continue;
|
|
||||||
start = min(node_memblk_range[i].start,
|
|
||||||
node_memblk_range[j].start);
|
|
||||||
end = max(node_memblk_range[i].end,
|
|
||||||
node_memblk_range[j].end);
|
|
||||||
printk(KERN_INFO "SRAT: Node %d "
|
|
||||||
"[%Lx,%Lx) + [%Lx,%Lx) -> [%lx,%lx)\n",
|
|
||||||
memblk_nodeid[i],
|
|
||||||
node_memblk_range[i].start,
|
|
||||||
node_memblk_range[i].end,
|
|
||||||
node_memblk_range[j].start,
|
|
||||||
node_memblk_range[j].end,
|
|
||||||
start, end);
|
|
||||||
node_memblk_range[i].start = start;
|
|
||||||
node_memblk_range[i].end = end;
|
|
||||||
k = --num_node_memblks - j;
|
|
||||||
memmove(memblk_nodeid + j, memblk_nodeid + j+1,
|
|
||||||
k * sizeof(*memblk_nodeid));
|
|
||||||
memmove(node_memblk_range + j, node_memblk_range + j+1,
|
|
||||||
k * sizeof(*node_memblk_range));
|
|
||||||
--j;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
memnode_shift = compute_hash_shift(node_memblk_range, num_node_memblks,
|
|
||||||
memblk_nodeid);
|
|
||||||
if (memnode_shift < 0) {
|
|
||||||
printk(KERN_ERR
|
|
||||||
"SRAT: No NUMA node hash function found. Contact maintainer\n");
|
|
||||||
bad_srat();
|
bad_srat();
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
for (i = 0; i < num_node_memblks; i++)
|
|
||||||
memblock_x86_register_active_regions(memblk_nodeid[i],
|
|
||||||
node_memblk_range[i].start >> PAGE_SHIFT,
|
|
||||||
node_memblk_range[i].end >> PAGE_SHIFT);
|
|
||||||
|
|
||||||
/* for out of order entries in SRAT */
|
/* for out of order entries in SRAT */
|
||||||
sort_node_map();
|
sort_node_map();
|
||||||
if (!nodes_cover_memory(numa_nodes)) {
|
if (!nodes_cover_memory(numa_nodes)) {
|
||||||
|
Reference in New Issue
Block a user