Loading include/asm-generic/qrwlock.h +1 −11 Original line number Original line Diff line number Diff line Loading @@ -146,23 +146,13 @@ static inline void queued_read_unlock(struct qrwlock *lock) (void)atomic_sub_return_release(_QR_BIAS, &lock->cnts); (void)atomic_sub_return_release(_QR_BIAS, &lock->cnts); } } /** * __qrwlock_write_byte - retrieve the write byte address of a queue rwlock * @lock : Pointer to queue rwlock structure * Return: the write byte address of a queue rwlock */ static inline u8 *__qrwlock_write_byte(struct qrwlock *lock) { return (u8 *)lock + 3 * IS_BUILTIN(CONFIG_CPU_BIG_ENDIAN); } /** /** * queued_write_unlock - release write lock of a queue rwlock * queued_write_unlock - release write lock of a queue rwlock * @lock : Pointer to queue rwlock structure * @lock : Pointer to queue rwlock structure */ */ static inline void queued_write_unlock(struct qrwlock *lock) static inline void queued_write_unlock(struct qrwlock *lock) { { smp_store_release(__qrwlock_write_byte(lock), 0); smp_store_release(&lock->wmode, 0); } } /* /* Loading include/asm-generic/qrwlock_types.h +13 −2 Original line number Original line Diff line number Diff line Loading @@ -10,12 +10,23 @@ */ */ typedef struct qrwlock { typedef struct qrwlock { union { atomic_t cnts; atomic_t cnts; struct { #ifdef __LITTLE_ENDIAN u8 wmode; /* Writer mode */ u8 rcnts[3]; /* Reader counts */ #else u8 rcnts[3]; /* Reader counts */ u8 wmode; /* Writer mode */ #endif }; }; arch_spinlock_t wait_lock; arch_spinlock_t wait_lock; } arch_rwlock_t; } arch_rwlock_t; #define __ARCH_RW_LOCK_UNLOCKED { \ #define __ARCH_RW_LOCK_UNLOCKED { \ .cnts = ATOMIC_INIT(0), \ { .cnts = ATOMIC_INIT(0), }, \ .wait_lock = __ARCH_SPIN_LOCK_UNLOCKED, \ .wait_lock = __ARCH_SPIN_LOCK_UNLOCKED, \ } } Loading kernel/locking/qrwlock.c +2 −24 Original line number Original line Diff line number Diff line Loading @@ -23,26 +23,6 @@ #include <linux/spinlock.h> #include <linux/spinlock.h> #include <asm/qrwlock.h> #include <asm/qrwlock.h> /* * This internal data structure is used for optimizing access to some of * the subfields within the atomic_t cnts. */ struct __qrwlock { union { atomic_t cnts; struct { #ifdef __LITTLE_ENDIAN u8 wmode; /* Writer mode */ u8 rcnts[3]; /* Reader counts */ #else u8 rcnts[3]; /* Reader counts */ u8 wmode; /* Writer mode */ #endif }; }; arch_spinlock_t lock; }; /** /** * rspin_until_writer_unlock - inc reader count & spin until writer is gone * rspin_until_writer_unlock - inc reader count & spin until writer is gone * @lock : Pointer to queue rwlock structure * @lock : Pointer to queue rwlock structure Loading Loading @@ -125,10 +105,8 @@ void queued_write_lock_slowpath(struct qrwlock *lock) * or wait for a previous writer to go away. * or wait for a previous writer to go away. */ */ for (;;) { for (;;) { struct __qrwlock *l = (struct __qrwlock *)lock; if (!READ_ONCE(lock->wmode) && (cmpxchg_relaxed(&lock->wmode, 0, _QW_WAITING) == 0)) if (!READ_ONCE(l->wmode) && (cmpxchg_relaxed(&l->wmode, 0, _QW_WAITING) == 0)) break; break; cpu_relax(); cpu_relax(); Loading Loading
include/asm-generic/qrwlock.h +1 −11 Original line number Original line Diff line number Diff line Loading @@ -146,23 +146,13 @@ static inline void queued_read_unlock(struct qrwlock *lock) (void)atomic_sub_return_release(_QR_BIAS, &lock->cnts); (void)atomic_sub_return_release(_QR_BIAS, &lock->cnts); } } /** * __qrwlock_write_byte - retrieve the write byte address of a queue rwlock * @lock : Pointer to queue rwlock structure * Return: the write byte address of a queue rwlock */ static inline u8 *__qrwlock_write_byte(struct qrwlock *lock) { return (u8 *)lock + 3 * IS_BUILTIN(CONFIG_CPU_BIG_ENDIAN); } /** /** * queued_write_unlock - release write lock of a queue rwlock * queued_write_unlock - release write lock of a queue rwlock * @lock : Pointer to queue rwlock structure * @lock : Pointer to queue rwlock structure */ */ static inline void queued_write_unlock(struct qrwlock *lock) static inline void queued_write_unlock(struct qrwlock *lock) { { smp_store_release(__qrwlock_write_byte(lock), 0); smp_store_release(&lock->wmode, 0); } } /* /* Loading
include/asm-generic/qrwlock_types.h +13 −2 Original line number Original line Diff line number Diff line Loading @@ -10,12 +10,23 @@ */ */ typedef struct qrwlock { typedef struct qrwlock { union { atomic_t cnts; atomic_t cnts; struct { #ifdef __LITTLE_ENDIAN u8 wmode; /* Writer mode */ u8 rcnts[3]; /* Reader counts */ #else u8 rcnts[3]; /* Reader counts */ u8 wmode; /* Writer mode */ #endif }; }; arch_spinlock_t wait_lock; arch_spinlock_t wait_lock; } arch_rwlock_t; } arch_rwlock_t; #define __ARCH_RW_LOCK_UNLOCKED { \ #define __ARCH_RW_LOCK_UNLOCKED { \ .cnts = ATOMIC_INIT(0), \ { .cnts = ATOMIC_INIT(0), }, \ .wait_lock = __ARCH_SPIN_LOCK_UNLOCKED, \ .wait_lock = __ARCH_SPIN_LOCK_UNLOCKED, \ } } Loading
kernel/locking/qrwlock.c +2 −24 Original line number Original line Diff line number Diff line Loading @@ -23,26 +23,6 @@ #include <linux/spinlock.h> #include <linux/spinlock.h> #include <asm/qrwlock.h> #include <asm/qrwlock.h> /* * This internal data structure is used for optimizing access to some of * the subfields within the atomic_t cnts. */ struct __qrwlock { union { atomic_t cnts; struct { #ifdef __LITTLE_ENDIAN u8 wmode; /* Writer mode */ u8 rcnts[3]; /* Reader counts */ #else u8 rcnts[3]; /* Reader counts */ u8 wmode; /* Writer mode */ #endif }; }; arch_spinlock_t lock; }; /** /** * rspin_until_writer_unlock - inc reader count & spin until writer is gone * rspin_until_writer_unlock - inc reader count & spin until writer is gone * @lock : Pointer to queue rwlock structure * @lock : Pointer to queue rwlock structure Loading Loading @@ -125,10 +105,8 @@ void queued_write_lock_slowpath(struct qrwlock *lock) * or wait for a previous writer to go away. * or wait for a previous writer to go away. */ */ for (;;) { for (;;) { struct __qrwlock *l = (struct __qrwlock *)lock; if (!READ_ONCE(lock->wmode) && (cmpxchg_relaxed(&lock->wmode, 0, _QW_WAITING) == 0)) if (!READ_ONCE(l->wmode) && (cmpxchg_relaxed(&l->wmode, 0, _QW_WAITING) == 0)) break; break; cpu_relax(); cpu_relax(); Loading