summaryrefslogtreecommitdiff
path: root/kernel/locking
diff options
context:
space:
mode:
authorThomas Gleixner <tglx@linutronix.de>2021-08-15 23:28:52 +0200
committerIngo Molnar <mingo@kernel.org>2021-08-17 19:05:03 +0200
commit653a5b0bd9b405db999d5f4bfe08d34691e2c55a (patch)
tree2cfe903f8230c8f0112b10b348326af0456add18 /kernel/locking
parentbdb189148ded4ffa826a1387074c795fda43b3ba (diff)
locking/ww_mutex: Abstract out internal lock accesses
Accessing the internal wait_lock of mutex and rtmutex is slightly different. Provide helper functions for that. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Signed-off-by: Ingo Molnar <mingo@kernel.org> Link: https://lore.kernel.org/r/20210815211304.734635961@linutronix.de
Diffstat (limited to 'kernel/locking')
-rw-r--r--kernel/locking/ww_mutex.h23
1 files changed, 19 insertions, 4 deletions
diff --git a/kernel/locking/ww_mutex.h b/kernel/locking/ww_mutex.h
index 31b075f03660..309f3e4d814a 100644
--- a/kernel/locking/ww_mutex.h
+++ b/kernel/locking/ww_mutex.h
@@ -68,6 +68,21 @@ __ww_mutex_has_waiters(struct mutex *lock)
return atomic_long_read(&lock->owner) & MUTEX_FLAG_WAITERS;
}
+static inline void lock_wait_lock(struct mutex *lock)
+{
+ raw_spin_lock(&lock->wait_lock);
+}
+
+static inline void unlock_wait_lock(struct mutex *lock)
+{
+ raw_spin_unlock(&lock->wait_lock);
+}
+
+static inline void lockdep_assert_wait_lock_held(struct mutex *lock)
+{
+ lockdep_assert_held(&lock->wait_lock);
+}
+
/*
* Wait-Die:
* The newer transactions are killed when:
@@ -174,7 +189,7 @@ static bool __ww_mutex_wound(struct MUTEX *lock,
{
struct task_struct *owner = __ww_mutex_owner(lock);
- lockdep_assert_held(&lock->wait_lock);
+ lockdep_assert_wait_lock_held(lock);
/*
* Possible through __ww_mutex_add_waiter() when we race with
@@ -227,7 +242,7 @@ __ww_mutex_check_waiters(struct MUTEX *lock, struct ww_acquire_ctx *ww_ctx)
{
struct MUTEX_WAITER *cur;
- lockdep_assert_held(&lock->wait_lock);
+ lockdep_assert_wait_lock_held(lock);
for (cur = __ww_waiter_first(lock); cur;
cur = __ww_waiter_next(lock, cur)) {
@@ -275,9 +290,9 @@ ww_mutex_set_context_fastpath(struct ww_mutex *lock, struct ww_acquire_ctx *ctx)
* Uh oh, we raced in fastpath, check if any of the waiters need to
* die or wound us.
*/
- raw_spin_lock(&lock->base.wait_lock);
+ lock_wait_lock(&lock->base);
__ww_mutex_check_waiters(&lock->base, ctx);
- raw_spin_unlock(&lock->base.wait_lock);
+ unlock_wait_lock(&lock->base);
}
static __always_inline int