Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 834beda1 authored by travis@sgi.com's avatar travis@sgi.com Committed by Ingo Molnar
Browse files

x86: change NR_CPUS arrays in numa_64 fixup



Change the following static arrays sized by NR_CPUS to
per_cpu data variables:

	char cpu_to_node_map[NR_CPUS];

fixup:

  - Split cpu_to_node function into "early" and "late" versions
    so that x86_cpu_to_node_map_early_ptr is not EXPORT'ed and
    the cpu_to_node inline function is more streamlined.

  - This also involves setting up the percpu maps as early as possible.

  - Fix X86_32 NUMA build errors that previous version of this
    patch caused.

V2->V3:
    - add early_cpu_to_node function to keep cpu_to_node efficient
    - move and rename smp_set_apicids() to setup_percpu_maps()
    - call setup_percpu_maps() as early as possible

V1->V2:
    - Removed extraneous casts
    - Fix !NUMA builds with '#ifdef CONFIG_NUMA"

Signed-off-by: default avatarMike Travis <travis@sgi.com>
Reviewed-by: default avatarChristoph Lameter <clameter@sgi.com>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
Signed-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
parent ac72e788
Loading
Loading
Loading
Loading
+39 −2
Original line number Original line Diff line number Diff line
@@ -83,6 +83,40 @@ static int __init nonx32_setup(char *str)
}
}
__setup("noexec32=", nonx32_setup);
__setup("noexec32=", nonx32_setup);


/*
 * Copy data used in early init routines from the initial arrays to the
 * per cpu data areas.  These arrays then become expendable and the *_ptrs
 * are zeroed indicating that the static arrays are gone.
 */
void __init setup_percpu_maps(void)
{
	int cpu;

	for_each_possible_cpu(cpu) {
#ifdef CONFIG_SMP
		if (per_cpu_offset(cpu)) {
#endif
			per_cpu(x86_cpu_to_apicid, cpu) =
						x86_cpu_to_apicid_init[cpu];
#ifdef CONFIG_NUMA
			per_cpu(x86_cpu_to_node_map, cpu) =
						x86_cpu_to_node_map_init[cpu];
#endif
#ifdef CONFIG_SMP
		}
		else
			printk(KERN_NOTICE "per_cpu_offset zero for cpu %d\n",
									cpu);
#endif
	}

	/* indicate the early static arrays are gone */
	x86_cpu_to_apicid_early_ptr = NULL;
#ifdef CONFIG_NUMA
	x86_cpu_to_node_map_early_ptr = NULL;
#endif
}

/*
/*
 * Great future plan:
 * Great future plan:
 * Declare PDA itself and support (irqstack,tss,pgd) as per cpu data.
 * Declare PDA itself and support (irqstack,tss,pgd) as per cpu data.
@@ -104,18 +138,21 @@ void __init setup_per_cpu_areas(void)
	for_each_cpu_mask (i, cpu_possible_map) {
	for_each_cpu_mask (i, cpu_possible_map) {
		char *ptr;
		char *ptr;


		if (!NODE_DATA(cpu_to_node(i))) {
		if (!NODE_DATA(early_cpu_to_node(i))) {
			printk("cpu with no node %d, num_online_nodes %d\n",
			printk("cpu with no node %d, num_online_nodes %d\n",
			       i, num_online_nodes());
			       i, num_online_nodes());
			ptr = alloc_bootmem_pages(size);
			ptr = alloc_bootmem_pages(size);
		} else { 
		} else { 
			ptr = alloc_bootmem_pages_node(NODE_DATA(cpu_to_node(i)), size);
			ptr = alloc_bootmem_pages_node(NODE_DATA(early_cpu_to_node(i)), size);
		}
		}
		if (!ptr)
		if (!ptr)
			panic("Cannot allocate cpu data for CPU %d\n", i);
			panic("Cannot allocate cpu data for CPU %d\n", i);
		cpu_pda(i)->data_offset = ptr - __per_cpu_start;
		cpu_pda(i)->data_offset = ptr - __per_cpu_start;
		memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start);
		memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start);
	}
	}

	/* setup percpu data maps early */
	setup_percpu_maps();
} 
} 


void pda_init(int cpu)
void pda_init(int cpu)
+1 −1
Original line number Original line Diff line number Diff line
@@ -465,7 +465,7 @@ cpumask_t node_to_cpumask_map[MAX_NUMNODES] __read_mostly =
				{ [0 ... MAX_NUMNODES-1] = CPU_MASK_NONE };
				{ [0 ... MAX_NUMNODES-1] = CPU_MASK_NONE };
EXPORT_SYMBOL(node_to_cpumask_map);
EXPORT_SYMBOL(node_to_cpumask_map);
/* which node each logical CPU is on */
/* which node each logical CPU is on */
int cpu_to_node_map[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = 0 };
u8 cpu_to_node_map[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = 0 };
EXPORT_SYMBOL(cpu_to_node_map);
EXPORT_SYMBOL(cpu_to_node_map);


/* set up a mapping between cpu and node. */
/* set up a mapping between cpu and node. */
+0 −34
Original line number Original line Diff line number Diff line
@@ -851,39 +851,6 @@ static int __init smp_sanity_check(unsigned max_cpus)
	return 0;
	return 0;
}
}


/*
 * Copy data used in early init routines from the initial arrays to the
 * per cpu data areas.  These arrays then become expendable and the
 * *_ptrs are zeroed indicating that the static arrays are gone.
 */
void __init smp_set_apicids(void)
{
	int cpu;

	for_each_possible_cpu(cpu) {
		if (per_cpu_offset(cpu)) {
			per_cpu(x86_cpu_to_apicid, cpu) =
						x86_cpu_to_apicid_init[cpu];
#ifdef CONFIG_NUMA
			per_cpu(x86_cpu_to_node_map, cpu) =
						x86_cpu_to_node_map_init[cpu];
#endif
			per_cpu(x86_bios_cpu_apicid, cpu) =
						x86_bios_cpu_apicid_init[cpu];
		}
		else
			printk(KERN_NOTICE "per_cpu_offset zero for cpu %d\n",
									cpu);
	}

	/* indicate the early static arrays are gone */
	x86_cpu_to_apicid_early_ptr = NULL;
#ifdef CONFIG_NUMA
	x86_cpu_to_node_map_early_ptr = NULL;
#endif
	x86_bios_cpu_apicid_early_ptr = NULL;
}

static void __init smp_cpu_index_default(void)
static void __init smp_cpu_index_default(void)
{
{
	int i;
	int i;
@@ -906,7 +873,6 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
	smp_cpu_index_default();
	smp_cpu_index_default();
	current_cpu_data = boot_cpu_data;
	current_cpu_data = boot_cpu_data;
	current_thread_info()->cpu = 0;  /* needed? */
	current_thread_info()->cpu = 0;  /* needed? */
	smp_set_apicids();
	set_cpu_sibling_map(0);
	set_cpu_sibling_map(0);


	if (smp_sanity_check(max_cpus) < 0) {
	if (smp_sanity_check(max_cpus) < 0) {
+3 −2
Original line number Original line Diff line number Diff line
@@ -382,9 +382,10 @@ int __init acpi_scan_nodes(unsigned long start, unsigned long end)
			setup_node_bootmem(i, nodes[i].start, nodes[i].end);
			setup_node_bootmem(i, nodes[i].start, nodes[i].end);


	for (i = 0; i < NR_CPUS; i++) {
	for (i = 0; i < NR_CPUS; i++) {
		if (cpu_to_node(i) == NUMA_NO_NODE)
		int node = cpu_to_node(i);
		if (node == NUMA_NO_NODE)
			continue;
			continue;
		if (!node_isset(cpu_to_node(i), node_possible_map))
		if (!node_isset(node, node_possible_map))
			numa_set_node(i, NUMA_NO_NODE);
			numa_set_node(i, NUMA_NO_NODE);
	}
	}
	numa_init_array();
	numa_init_array();
+23 −0
Original line number Original line Diff line number Diff line
@@ -30,15 +30,29 @@
#include <asm/mpspec.h>
#include <asm/mpspec.h>


/* Mappings between logical cpu number and node number */
/* Mappings between logical cpu number and node number */
#ifdef CONFIG_X86_32
extern u8 cpu_to_node_map[];

#else
DECLARE_PER_CPU(u16, x86_cpu_to_node_map);
DECLARE_PER_CPU(u16, x86_cpu_to_node_map);
extern u16 x86_cpu_to_node_map_init[];
extern u16 x86_cpu_to_node_map_init[];
extern void *x86_cpu_to_node_map_early_ptr;
extern void *x86_cpu_to_node_map_early_ptr;
#endif

extern cpumask_t node_to_cpumask_map[];
extern cpumask_t node_to_cpumask_map[];


#define NUMA_NO_NODE	((u16)(~0))
#define NUMA_NO_NODE	((u16)(~0))


/* Returns the number of the node containing CPU 'cpu' */
/* Returns the number of the node containing CPU 'cpu' */
#ifdef CONFIG_X86_32
#define early_cpu_to_node(cpu)	cpu_to_node(cpu)
static inline int cpu_to_node(int cpu)
static inline int cpu_to_node(int cpu)
{
	return cpu_to_node_map[cpu];
}

#else /* CONFIG_X86_64 */
static inline int early_cpu_to_node(int cpu)
{
{
	u16 *cpu_to_node_map = x86_cpu_to_node_map_early_ptr;
	u16 *cpu_to_node_map = x86_cpu_to_node_map_early_ptr;


@@ -50,6 +64,15 @@ static inline int cpu_to_node(int cpu)
		return NUMA_NO_NODE;
		return NUMA_NO_NODE;
}
}


static inline int cpu_to_node(int cpu)
{
	if(per_cpu_offset(cpu))
		return per_cpu(x86_cpu_to_node_map, cpu);
	else
		return NUMA_NO_NODE;
}
#endif /* CONFIG_X86_64 */

/*
/*
 * Returns the number of the node containing Node 'node'. This
 * Returns the number of the node containing Node 'node'. This
 * architecture is flat, so it is a pretty simple function!
 * architecture is flat, so it is a pretty simple function!