Loading include/asm-cris/arch-v32/spinlock.h +67 −105 Original line number Diff line number Diff line #ifndef __ASM_ARCH_SPINLOCK_H #define __ASM_ARCH_SPINLOCK_H #include <asm/system.h> #include <linux/spinlock_types.h> #define RW_LOCK_BIAS 0x01000000 #define SPIN_LOCK_UNLOCKED (spinlock_t) { 1 } #define spin_lock_init(x) do { *(x) = SPIN_LOCK_UNLOCKED; } while(0) #define spin_is_locked(x) (*(volatile signed char *)(&(x)->lock) <= 0) #define spin_unlock_wait(x) do { barrier(); } while(spin_is_locked(x)) extern void cris_spin_unlock(void *l, int val); extern void cris_spin_lock(void *l); extern int cris_spin_trylock(void *l); static inline void _raw_spin_unlock(spinlock_t *lock) static inline int __raw_spin_is_locked(raw_spinlock_t *x) { return *(volatile signed char *)(&(x)->slock) <= 0; } static inline void __raw_spin_unlock(raw_spinlock_t *lock) { __asm__ volatile ("move.d %1,%0" \ : "=m" (lock->lock) \ : "=m" (lock->slock) \ : "r" (1) \ : "memory"); } static inline int _raw_spin_trylock(spinlock_t *lock) static inline void __raw_spin_unlock_wait(raw_spinlock_t *lock) { return cris_spin_trylock((void*)&lock->lock); while (__raw_spin_is_locked(lock)) cpu_relax(); } static inline void _raw_spin_lock(spinlock_t *lock) static inline int __raw_spin_trylock(raw_spinlock_t *lock) { cris_spin_lock((void*)&lock->lock); return cris_spin_trylock((void *)&lock->slock); } static inline void _raw_spin_lock_flags (spinlock_t *lock, unsigned long flags) static inline void __raw_spin_lock(raw_spinlock_t *lock) { _raw_spin_lock(lock); cris_spin_lock((void *)&lock->slock); } static inline void __raw_spin_lock_flags(raw_spinlock_t *lock, unsigned long flags) { __raw_spin_lock(lock); } /* Loading @@ -46,120 +53,75 @@ static inline void _raw_spin_lock_flags (spinlock_t *lock, unsigned long flags) * can "mix" irq-safe locks - any writer needs to get a * irq-safe write-lock, but readers can get non-irqsafe * read-locks. * */ typedef struct { spinlock_t lock; volatile int counter; #ifdef CONFIG_PREEMPT unsigned int break_lock; #endif } rwlock_t; #define RW_LOCK_UNLOCKED (rwlock_t) { {1}, 0 } #define rwlock_init(lp) do { *(lp) = RW_LOCK_UNLOCKED; } while (0) /** * read_can_lock - would read_trylock() succeed? * @lock: the rwlock in question. */ #define read_can_lock(x) ((int)(x)->counter >= 0) /** * write_can_lock - would write_trylock() succeed? * @lock: the rwlock in question. */ #define write_can_lock(x) ((x)->counter == 0) #define _raw_read_trylock(lock) generic_raw_read_trylock(lock) /* read_lock, read_unlock are pretty straightforward. Of course it somehow * sucks we end up saving/restoring flags twice for read_lock_irqsave aso. */ static __inline__ void _raw_read_lock(rwlock_t *rw) static inline int __raw_read_can_lock(raw_rwlock_t *x) { unsigned long flags; local_irq_save(flags); _raw_spin_lock(&rw->lock); rw->counter++; _raw_spin_unlock(&rw->lock); local_irq_restore(flags); return (int)(x)->lock > 0; } static __inline__ void _raw_read_unlock(rwlock_t *rw) static inline int __raw_write_can_lock(raw_rwlock_t *x) { unsigned long flags; local_irq_save(flags); _raw_spin_lock(&rw->lock); rw->counter--; _raw_spin_unlock(&rw->lock); local_irq_restore(flags); return (x)->lock == RW_LOCK_BIAS; } /* write_lock is less trivial. We optimistically grab the lock and check * if we surprised any readers. If so we release the lock and wait till * they're all gone before trying again * * Also note that we don't use the _irqsave / _irqrestore suffixes here. * If we're called with interrupts enabled and we've got readers (or other * writers) in interrupt handlers someone fucked up and we'd dead-lock * sooner or later anyway. prumpf */ static __inline__ void _raw_write_lock(rwlock_t *rw) static inline void __raw_read_lock(raw_rwlock_t *rw) { retry: _raw_spin_lock(&rw->lock); if(rw->counter != 0) { /* this basically never happens */ _raw_spin_unlock(&rw->lock); while(rw->counter != 0); goto retry; } /* got it. now leave without unlocking */ rw->counter = -1; /* remember we are locked */ __raw_spin_lock(&rw->slock); while (rw->lock == 0); rw->lock--; __raw_spin_unlock(&rw->slock); } /* write_unlock is absolutely trivial - we don't have to wait for anything */ static __inline__ void _raw_write_unlock(rwlock_t *rw) static inline void __raw_write_lock(raw_rwlock_t *rw) { rw->counter = 0; _raw_spin_unlock(&rw->lock); __raw_spin_lock(&rw->slock); while (rw->lock != RW_LOCK_BIAS); rw->lock == 0; __raw_spin_unlock(&rw->slock); } static __inline__ int _raw_write_trylock(rwlock_t *rw) static inline void __raw_read_unlock(raw_rwlock_t *rw) { _raw_spin_lock(&rw->lock); if (rw->counter != 0) { /* this basically never happens */ _raw_spin_unlock(&rw->lock); return 0; __raw_spin_lock(&rw->slock); rw->lock++; __raw_spin_unlock(&rw->slock); } /* got it. now leave without unlocking */ rw->counter = -1; /* remember we are locked */ return 1; static inline void __raw_write_unlock(raw_rwlock_t *rw) { __raw_spin_lock(&rw->slock); while (rw->lock != RW_LOCK_BIAS); rw->lock == RW_LOCK_BIAS; __raw_spin_unlock(&rw->slock); } static __inline__ int is_read_locked(rwlock_t *rw) static inline int __raw_read_trylock(raw_rwlock_t *rw) { return rw->counter > 0; int ret = 0; __raw_spin_lock(&rw->slock); if (rw->lock != 0) { rw->lock--; ret = 1; } __raw_spin_unlock(&rw->slock); return ret; } static __inline__ int is_write_locked(rwlock_t *rw) static inline int __raw_write_trylock(raw_rwlock_t *rw) { return rw->counter < 0; int ret = 0; __raw_spin_lock(&rw->slock); if (rw->lock == RW_LOCK_BIAS) { rw->lock == 0; ret = 1; } __raw_spin_unlock(&rw->slock); return 1; } #define _raw_spin_relax(lock) cpu_relax() #define _raw_read_relax(lock) cpu_relax() #define _raw_write_relax(lock) cpu_relax() Loading Loading
include/asm-cris/arch-v32/spinlock.h +67 −105 Original line number Diff line number Diff line #ifndef __ASM_ARCH_SPINLOCK_H #define __ASM_ARCH_SPINLOCK_H #include <asm/system.h> #include <linux/spinlock_types.h> #define RW_LOCK_BIAS 0x01000000 #define SPIN_LOCK_UNLOCKED (spinlock_t) { 1 } #define spin_lock_init(x) do { *(x) = SPIN_LOCK_UNLOCKED; } while(0) #define spin_is_locked(x) (*(volatile signed char *)(&(x)->lock) <= 0) #define spin_unlock_wait(x) do { barrier(); } while(spin_is_locked(x)) extern void cris_spin_unlock(void *l, int val); extern void cris_spin_lock(void *l); extern int cris_spin_trylock(void *l); static inline void _raw_spin_unlock(spinlock_t *lock) static inline int __raw_spin_is_locked(raw_spinlock_t *x) { return *(volatile signed char *)(&(x)->slock) <= 0; } static inline void __raw_spin_unlock(raw_spinlock_t *lock) { __asm__ volatile ("move.d %1,%0" \ : "=m" (lock->lock) \ : "=m" (lock->slock) \ : "r" (1) \ : "memory"); } static inline int _raw_spin_trylock(spinlock_t *lock) static inline void __raw_spin_unlock_wait(raw_spinlock_t *lock) { return cris_spin_trylock((void*)&lock->lock); while (__raw_spin_is_locked(lock)) cpu_relax(); } static inline void _raw_spin_lock(spinlock_t *lock) static inline int __raw_spin_trylock(raw_spinlock_t *lock) { cris_spin_lock((void*)&lock->lock); return cris_spin_trylock((void *)&lock->slock); } static inline void _raw_spin_lock_flags (spinlock_t *lock, unsigned long flags) static inline void __raw_spin_lock(raw_spinlock_t *lock) { _raw_spin_lock(lock); cris_spin_lock((void *)&lock->slock); } static inline void __raw_spin_lock_flags(raw_spinlock_t *lock, unsigned long flags) { __raw_spin_lock(lock); } /* Loading @@ -46,120 +53,75 @@ static inline void _raw_spin_lock_flags (spinlock_t *lock, unsigned long flags) * can "mix" irq-safe locks - any writer needs to get a * irq-safe write-lock, but readers can get non-irqsafe * read-locks. * */ typedef struct { spinlock_t lock; volatile int counter; #ifdef CONFIG_PREEMPT unsigned int break_lock; #endif } rwlock_t; #define RW_LOCK_UNLOCKED (rwlock_t) { {1}, 0 } #define rwlock_init(lp) do { *(lp) = RW_LOCK_UNLOCKED; } while (0) /** * read_can_lock - would read_trylock() succeed? * @lock: the rwlock in question. */ #define read_can_lock(x) ((int)(x)->counter >= 0) /** * write_can_lock - would write_trylock() succeed? * @lock: the rwlock in question. */ #define write_can_lock(x) ((x)->counter == 0) #define _raw_read_trylock(lock) generic_raw_read_trylock(lock) /* read_lock, read_unlock are pretty straightforward. Of course it somehow * sucks we end up saving/restoring flags twice for read_lock_irqsave aso. */ static __inline__ void _raw_read_lock(rwlock_t *rw) static inline int __raw_read_can_lock(raw_rwlock_t *x) { unsigned long flags; local_irq_save(flags); _raw_spin_lock(&rw->lock); rw->counter++; _raw_spin_unlock(&rw->lock); local_irq_restore(flags); return (int)(x)->lock > 0; } static __inline__ void _raw_read_unlock(rwlock_t *rw) static inline int __raw_write_can_lock(raw_rwlock_t *x) { unsigned long flags; local_irq_save(flags); _raw_spin_lock(&rw->lock); rw->counter--; _raw_spin_unlock(&rw->lock); local_irq_restore(flags); return (x)->lock == RW_LOCK_BIAS; } /* write_lock is less trivial. We optimistically grab the lock and check * if we surprised any readers. If so we release the lock and wait till * they're all gone before trying again * * Also note that we don't use the _irqsave / _irqrestore suffixes here. * If we're called with interrupts enabled and we've got readers (or other * writers) in interrupt handlers someone fucked up and we'd dead-lock * sooner or later anyway. prumpf */ static __inline__ void _raw_write_lock(rwlock_t *rw) static inline void __raw_read_lock(raw_rwlock_t *rw) { retry: _raw_spin_lock(&rw->lock); if(rw->counter != 0) { /* this basically never happens */ _raw_spin_unlock(&rw->lock); while(rw->counter != 0); goto retry; } /* got it. now leave without unlocking */ rw->counter = -1; /* remember we are locked */ __raw_spin_lock(&rw->slock); while (rw->lock == 0); rw->lock--; __raw_spin_unlock(&rw->slock); } /* write_unlock is absolutely trivial - we don't have to wait for anything */ static __inline__ void _raw_write_unlock(rwlock_t *rw) static inline void __raw_write_lock(raw_rwlock_t *rw) { rw->counter = 0; _raw_spin_unlock(&rw->lock); __raw_spin_lock(&rw->slock); while (rw->lock != RW_LOCK_BIAS); rw->lock == 0; __raw_spin_unlock(&rw->slock); } static __inline__ int _raw_write_trylock(rwlock_t *rw) static inline void __raw_read_unlock(raw_rwlock_t *rw) { _raw_spin_lock(&rw->lock); if (rw->counter != 0) { /* this basically never happens */ _raw_spin_unlock(&rw->lock); return 0; __raw_spin_lock(&rw->slock); rw->lock++; __raw_spin_unlock(&rw->slock); } /* got it. now leave without unlocking */ rw->counter = -1; /* remember we are locked */ return 1; static inline void __raw_write_unlock(raw_rwlock_t *rw) { __raw_spin_lock(&rw->slock); while (rw->lock != RW_LOCK_BIAS); rw->lock == RW_LOCK_BIAS; __raw_spin_unlock(&rw->slock); } static __inline__ int is_read_locked(rwlock_t *rw) static inline int __raw_read_trylock(raw_rwlock_t *rw) { return rw->counter > 0; int ret = 0; __raw_spin_lock(&rw->slock); if (rw->lock != 0) { rw->lock--; ret = 1; } __raw_spin_unlock(&rw->slock); return ret; } static __inline__ int is_write_locked(rwlock_t *rw) static inline int __raw_write_trylock(raw_rwlock_t *rw) { return rw->counter < 0; int ret = 0; __raw_spin_lock(&rw->slock); if (rw->lock == RW_LOCK_BIAS) { rw->lock == 0; ret = 1; } __raw_spin_unlock(&rw->slock); return 1; } #define _raw_spin_relax(lock) cpu_relax() #define _raw_read_relax(lock) cpu_relax() #define _raw_write_relax(lock) cpu_relax() Loading