Loading include/linux/lglock.h +32 −4 Original line number Diff line number Diff line Loading @@ -22,6 +22,7 @@ #include <linux/spinlock.h> #include <linux/lockdep.h> #include <linux/percpu.h> #include <linux/cpu.h> /* can make br locks by using local lock for read side, global lock for write */ #define br_lock_init(name) name##_lock_init() Loading Loading @@ -72,9 +73,31 @@ #define DEFINE_LGLOCK(name) \ \ DEFINE_SPINLOCK(name##_cpu_lock); \ cpumask_t name##_cpus __read_mostly; \ DEFINE_PER_CPU(arch_spinlock_t, name##_lock); \ DEFINE_LGLOCK_LOCKDEP(name); \ \ static int \ name##_lg_cpu_callback(struct notifier_block *nb, \ unsigned long action, void *hcpu) \ { \ switch (action & ~CPU_TASKS_FROZEN) { \ case CPU_UP_PREPARE: \ spin_lock(&name##_cpu_lock); \ cpu_set((unsigned long)hcpu, name##_cpus); \ spin_unlock(&name##_cpu_lock); \ break; \ case CPU_UP_CANCELED: case CPU_DEAD: \ spin_lock(&name##_cpu_lock); \ cpu_clear((unsigned long)hcpu, name##_cpus); \ spin_unlock(&name##_cpu_lock); \ } \ return NOTIFY_OK; \ } \ static struct notifier_block name##_lg_cpu_notifier = { \ .notifier_call = name##_lg_cpu_callback, \ }; \ void name##_lock_init(void) { \ int i; \ LOCKDEP_INIT_MAP(&name##_lock_dep_map, #name, &name##_lock_key, 0); \ Loading @@ -83,6 +106,11 @@ lock = &per_cpu(name##_lock, i); \ *lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; \ } \ register_hotcpu_notifier(&name##_lg_cpu_notifier); \ get_online_cpus(); \ for_each_online_cpu(i) \ cpu_set(i, name##_cpus); \ put_online_cpus(); \ } \ EXPORT_SYMBOL(name##_lock_init); \ \ Loading Loading @@ -124,9 +152,9 @@ \ void name##_global_lock_online(void) { \ int i; \ preempt_disable(); \ spin_lock(&name##_cpu_lock); \ rwlock_acquire(&name##_lock_dep_map, 0, 0, _RET_IP_); \ for_each_online_cpu(i) { \ for_each_cpu(i, &name##_cpus) { \ arch_spinlock_t *lock; \ lock = &per_cpu(name##_lock, i); \ arch_spin_lock(lock); \ Loading @@ -137,12 +165,12 @@ void name##_global_unlock_online(void) { \ int i; \ rwlock_release(&name##_lock_dep_map, 1, _RET_IP_); \ for_each_online_cpu(i) { \ for_each_cpu(i, &name##_cpus) { \ arch_spinlock_t *lock; \ lock = &per_cpu(name##_lock, i); \ arch_spin_unlock(lock); \ } \ preempt_enable(); \ spin_unlock(&name##_cpu_lock); \ } \ EXPORT_SYMBOL(name##_global_unlock_online); \ \ Loading Loading
include/linux/lglock.h +32 −4 Original line number Diff line number Diff line Loading @@ -22,6 +22,7 @@ #include <linux/spinlock.h> #include <linux/lockdep.h> #include <linux/percpu.h> #include <linux/cpu.h> /* can make br locks by using local lock for read side, global lock for write */ #define br_lock_init(name) name##_lock_init() Loading Loading @@ -72,9 +73,31 @@ #define DEFINE_LGLOCK(name) \ \ DEFINE_SPINLOCK(name##_cpu_lock); \ cpumask_t name##_cpus __read_mostly; \ DEFINE_PER_CPU(arch_spinlock_t, name##_lock); \ DEFINE_LGLOCK_LOCKDEP(name); \ \ static int \ name##_lg_cpu_callback(struct notifier_block *nb, \ unsigned long action, void *hcpu) \ { \ switch (action & ~CPU_TASKS_FROZEN) { \ case CPU_UP_PREPARE: \ spin_lock(&name##_cpu_lock); \ cpu_set((unsigned long)hcpu, name##_cpus); \ spin_unlock(&name##_cpu_lock); \ break; \ case CPU_UP_CANCELED: case CPU_DEAD: \ spin_lock(&name##_cpu_lock); \ cpu_clear((unsigned long)hcpu, name##_cpus); \ spin_unlock(&name##_cpu_lock); \ } \ return NOTIFY_OK; \ } \ static struct notifier_block name##_lg_cpu_notifier = { \ .notifier_call = name##_lg_cpu_callback, \ }; \ void name##_lock_init(void) { \ int i; \ LOCKDEP_INIT_MAP(&name##_lock_dep_map, #name, &name##_lock_key, 0); \ Loading @@ -83,6 +106,11 @@ lock = &per_cpu(name##_lock, i); \ *lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; \ } \ register_hotcpu_notifier(&name##_lg_cpu_notifier); \ get_online_cpus(); \ for_each_online_cpu(i) \ cpu_set(i, name##_cpus); \ put_online_cpus(); \ } \ EXPORT_SYMBOL(name##_lock_init); \ \ Loading Loading @@ -124,9 +152,9 @@ \ void name##_global_lock_online(void) { \ int i; \ preempt_disable(); \ spin_lock(&name##_cpu_lock); \ rwlock_acquire(&name##_lock_dep_map, 0, 0, _RET_IP_); \ for_each_online_cpu(i) { \ for_each_cpu(i, &name##_cpus) { \ arch_spinlock_t *lock; \ lock = &per_cpu(name##_lock, i); \ arch_spin_lock(lock); \ Loading @@ -137,12 +165,12 @@ void name##_global_unlock_online(void) { \ int i; \ rwlock_release(&name##_lock_dep_map, 1, _RET_IP_); \ for_each_online_cpu(i) { \ for_each_cpu(i, &name##_cpus) { \ arch_spinlock_t *lock; \ lock = &per_cpu(name##_lock, i); \ arch_spin_unlock(lock); \ } \ preempt_enable(); \ spin_unlock(&name##_cpu_lock); \ } \ EXPORT_SYMBOL(name##_global_unlock_online); \ \ Loading