Loading include/asm-cris/arch-v32/atomic.h +5 −5 Original line number Original line Diff line number Diff line #ifndef __ASM_CRIS_ARCH_ATOMIC__ #ifndef __ASM_CRIS_ARCH_ATOMIC__ #define __ASM_CRIS_ARCH_ATOMIC__ #define __ASM_CRIS_ARCH_ATOMIC__ #include <asm/system.h> #include <linux/spinlock_types.h> extern void cris_spin_unlock(void *l, int val); extern void cris_spin_unlock(void *l, int val); extern void cris_spin_lock(void *l); extern void cris_spin_lock(void *l); Loading @@ -18,13 +18,13 @@ extern spinlock_t cris_atomic_locks[]; #define cris_atomic_save(addr, flags) \ #define cris_atomic_save(addr, flags) \ local_irq_save(flags); \ local_irq_save(flags); \ cris_spin_lock((void*)&cris_atomic_locks[HASH_ADDR(addr)].lock); cris_spin_lock((void *)&cris_atomic_locks[HASH_ADDR(addr)].raw_lock.slock); #define cris_atomic_restore(addr, flags) \ #define cris_atomic_restore(addr, flags) \ { \ { \ spinlock_t *lock = (void*)&cris_atomic_locks[HASH_ADDR(addr)]; \ spinlock_t *lock = (void*)&cris_atomic_locks[HASH_ADDR(addr)]; \ __asm__ volatile ("move.d %1,%0" \ __asm__ volatile ("move.d %1,%0" \ : "=m" (lock->lock) \ : "=m" (lock->raw_lock.slock) \ : "r" (1) \ : "r" (1) \ : "memory"); \ : "memory"); \ local_irq_restore(flags); \ local_irq_restore(flags); \ Loading Loading
include/asm-cris/arch-v32/atomic.h +5 −5 Original line number Original line Diff line number Diff line #ifndef __ASM_CRIS_ARCH_ATOMIC__ #ifndef __ASM_CRIS_ARCH_ATOMIC__ #define __ASM_CRIS_ARCH_ATOMIC__ #define __ASM_CRIS_ARCH_ATOMIC__ #include <asm/system.h> #include <linux/spinlock_types.h> extern void cris_spin_unlock(void *l, int val); extern void cris_spin_unlock(void *l, int val); extern void cris_spin_lock(void *l); extern void cris_spin_lock(void *l); Loading @@ -18,13 +18,13 @@ extern spinlock_t cris_atomic_locks[]; #define cris_atomic_save(addr, flags) \ #define cris_atomic_save(addr, flags) \ local_irq_save(flags); \ local_irq_save(flags); \ cris_spin_lock((void*)&cris_atomic_locks[HASH_ADDR(addr)].lock); cris_spin_lock((void *)&cris_atomic_locks[HASH_ADDR(addr)].raw_lock.slock); #define cris_atomic_restore(addr, flags) \ #define cris_atomic_restore(addr, flags) \ { \ { \ spinlock_t *lock = (void*)&cris_atomic_locks[HASH_ADDR(addr)]; \ spinlock_t *lock = (void*)&cris_atomic_locks[HASH_ADDR(addr)]; \ __asm__ volatile ("move.d %1,%0" \ __asm__ volatile ("move.d %1,%0" \ : "=m" (lock->lock) \ : "=m" (lock->raw_lock.slock) \ : "r" (1) \ : "r" (1) \ : "memory"); \ : "memory"); \ local_irq_restore(flags); \ local_irq_restore(flags); \ Loading