Loading arch/sparc64/kernel/smp.c +2 −5 Original line number Diff line number Diff line Loading @@ -1447,11 +1447,8 @@ void __init setup_per_cpu_areas(void) char *ptr; /* Copy section for each CPU (we discard the original) */ goal = ALIGN(__per_cpu_end - __per_cpu_start, SMP_CACHE_BYTES); #ifdef CONFIG_MODULES if (goal < PERCPU_ENOUGH_ROOM) goal = PERCPU_ENOUGH_ROOM; #endif __per_cpu_shift = 0; for (size = 1UL; size < goal; size <<= 1UL) __per_cpu_shift++; Loading include/asm-sparc64/percpu.h +10 −0 Original line number Diff line number Diff line Loading @@ -5,6 +5,16 @@ #ifdef CONFIG_SMP #ifdef CONFIG_MODULES # define PERCPU_MODULE_RESERVE 8192 #else # define PERCPU_MODULE_RESERVE 0 #endif #define PERCPU_ENOUGH_ROOM \ (ALIGN(__per_cpu_end - __per_cpu_start, SMP_CACHE_BYTES) + \ PERCPU_MODULE_RESERVE) extern void setup_per_cpu_areas(void); extern unsigned long __per_cpu_base; Loading Loading
arch/sparc64/kernel/smp.c +2 −5 Original line number Diff line number Diff line Loading @@ -1447,11 +1447,8 @@ void __init setup_per_cpu_areas(void) char *ptr; /* Copy section for each CPU (we discard the original) */ goal = ALIGN(__per_cpu_end - __per_cpu_start, SMP_CACHE_BYTES); #ifdef CONFIG_MODULES if (goal < PERCPU_ENOUGH_ROOM) goal = PERCPU_ENOUGH_ROOM; #endif __per_cpu_shift = 0; for (size = 1UL; size < goal; size <<= 1UL) __per_cpu_shift++; Loading
include/asm-sparc64/percpu.h +10 −0 Original line number Diff line number Diff line Loading @@ -5,6 +5,16 @@ #ifdef CONFIG_SMP #ifdef CONFIG_MODULES # define PERCPU_MODULE_RESERVE 8192 #else # define PERCPU_MODULE_RESERVE 0 #endif #define PERCPU_ENOUGH_ROOM \ (ALIGN(__per_cpu_end - __per_cpu_start, SMP_CACHE_BYTES) + \ PERCPU_MODULE_RESERVE) extern void setup_per_cpu_areas(void); extern unsigned long __per_cpu_base; Loading