Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 2e109a28 authored by Tejun Heo's avatar Tejun Heo
Browse files

workqueue: rename workqueue_lock to wq_mayday_lock



With the recent locking updates, the only thing protected by
workqueue_lock is workqueue->maydays list.  Rename workqueue_lock to
wq_mayday_lock.

This patch is pure rename.

Signed-off-by: default avatarTejun Heo <tj@kernel.org>
parent 794b18bc
Loading
Loading
Loading
Loading
+12 −12
Original line number Diff line number Diff line
@@ -125,10 +125,10 @@ enum {
 *
 * PW: pwq_lock protected.
 *
 * W: workqueue_lock protected.
 *
 * FR: wq->flush_mutex and pwq_lock protected for writes.  Sched-RCU
 *     protected for reads.
 *
 * MD: wq_mayday_lock protected.
 */

/* struct worker is defined in workqueue_internal.h */
@@ -194,7 +194,7 @@ struct pool_workqueue {
	int			max_active;	/* L: max active works */
	struct list_head	delayed_works;	/* L: delayed works */
	struct list_head	pwqs_node;	/* FR: node on wq->pwqs */
	struct list_head	mayday_node;	/* W: node on wq->maydays */
	struct list_head	mayday_node;	/* MD: node on wq->maydays */

	/*
	 * Release of unbound pwq is punted to system_wq.  See put_pwq()
@@ -235,7 +235,7 @@ struct workqueue_struct {
	struct list_head	flusher_queue;	/* F: flush waiters */
	struct list_head	flusher_overflow; /* F: flush overflow list */

	struct list_head	maydays;	/* W: pwqs requesting rescue */
	struct list_head	maydays;	/* MD: pwqs requesting rescue */
	struct worker		*rescuer;	/* I: rescue worker */

	int			nr_drainers;	/* WQ: drain in progress */
@@ -254,7 +254,7 @@ static struct kmem_cache *pwq_cache;

static DEFINE_MUTEX(wq_mutex);		/* protects workqueues and pools */
static DEFINE_SPINLOCK(pwq_lock);	/* protects pool_workqueues */
static DEFINE_SPINLOCK(workqueue_lock);
static DEFINE_SPINLOCK(wq_mayday_lock);	/* protects wq->maydays list */

static LIST_HEAD(workqueues);		/* WQ: list of all workqueues */
static bool workqueue_freezing;		/* WQ: have wqs started freezing? */
@@ -1894,7 +1894,7 @@ static void send_mayday(struct work_struct *work)
	struct pool_workqueue *pwq = get_work_pwq(work);
	struct workqueue_struct *wq = pwq->wq;

	lockdep_assert_held(&workqueue_lock);
	lockdep_assert_held(&wq_mayday_lock);

	if (!wq->rescuer)
		return;
@@ -1911,7 +1911,7 @@ static void pool_mayday_timeout(unsigned long __pool)
	struct worker_pool *pool = (void *)__pool;
	struct work_struct *work;

	spin_lock_irq(&workqueue_lock);		/* for wq->maydays */
	spin_lock_irq(&wq_mayday_lock);		/* for wq->maydays */
	spin_lock(&pool->lock);

	if (need_to_create_worker(pool)) {
@@ -1926,7 +1926,7 @@ static void pool_mayday_timeout(unsigned long __pool)
	}

	spin_unlock(&pool->lock);
	spin_unlock_irq(&workqueue_lock);
	spin_unlock_irq(&wq_mayday_lock);

	mod_timer(&pool->mayday_timer, jiffies + MAYDAY_INTERVAL);
}
@@ -2404,7 +2404,7 @@ static int rescuer_thread(void *__rescuer)
	}

	/* see whether any pwq is asking for help */
	spin_lock_irq(&workqueue_lock);
	spin_lock_irq(&wq_mayday_lock);

	while (!list_empty(&wq->maydays)) {
		struct pool_workqueue *pwq = list_first_entry(&wq->maydays,
@@ -2415,7 +2415,7 @@ static int rescuer_thread(void *__rescuer)
		__set_current_state(TASK_RUNNING);
		list_del_init(&pwq->mayday_node);

		spin_unlock_irq(&workqueue_lock);
		spin_unlock_irq(&wq_mayday_lock);

		/* migrate to the target cpu if possible */
		worker_maybe_bind_and_lock(pool);
@@ -2442,10 +2442,10 @@ static int rescuer_thread(void *__rescuer)

		rescuer->pool = NULL;
		spin_unlock(&pool->lock);
		spin_lock(&workqueue_lock);
		spin_lock(&wq_mayday_lock);
	}

	spin_unlock_irq(&workqueue_lock);
	spin_unlock_irq(&wq_mayday_lock);

	/* rescuers should never participate in concurrency management */
	WARN_ON_ONCE(!(rescuer->flags & WORKER_NOT_RUNNING));