Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit f153ad11 authored by Tejun Heo's avatar Tejun Heo
Browse files

Merge branch 'cgroup/for-3.14-fixes' into cgroup/for-3.15



Pull in for-3.14-fixes to receive 532de3fc ("cgroup: update
cgroup_enable_task_cg_lists() to grab siglock") which conflicts with
afeb0f9f ("cgroup: relocate cgroup_enable_task_cg_lists()") and
the following cg_lists updates.  This is likely to cause further
conflicts down the line too, so let's merge it early.

As cgroup_enable_task_cg_lists() is relocated in for-3.15, this merge
causes conflict in the original position.  It's resolved by applying
siglock changes to the updated version in the new location.

Conflicts:
	kernel/cgroup.c

Signed-off-by: default avatarTejun Heo <tj@kernel.org>
parents dc5736ed 532de3fc
Loading
Loading
Loading
Loading
+6 −6
Original line number Original line Diff line number Diff line
@@ -1331,9 +1331,13 @@ static void cgroup_enable_task_cg_lists(void)
		 * We should check if the process is exiting, otherwise
		 * We should check if the process is exiting, otherwise
		 * it will race with cgroup_exit() in that the list
		 * it will race with cgroup_exit() in that the list
		 * entry won't be deleted though the process has exited.
		 * entry won't be deleted though the process has exited.
		 * Do it while holding siglock so that we don't end up
		 * racing against cgroup_exit().
		 */
		 */
		spin_lock_irq(&p->sighand->siglock);
		if (!(p->flags & PF_EXITING))
		if (!(p->flags & PF_EXITING))
			list_add(&p->cg_list, &task_css_set(p)->tasks);
			list_add(&p->cg_list, &task_css_set(p)->tasks);
		spin_unlock_irq(&p->sighand->siglock);


		task_unlock(p);
		task_unlock(p);
	} while_each_thread(g, p);
	} while_each_thread(g, p);
@@ -3968,16 +3972,12 @@ static int __init cgroup_wq_init(void)
	/*
	/*
	 * There isn't much point in executing destruction path in
	 * There isn't much point in executing destruction path in
	 * parallel.  Good chunk is serialized with cgroup_mutex anyway.
	 * parallel.  Good chunk is serialized with cgroup_mutex anyway.
	 *
	 * Use 1 for @max_active.
	 * XXX: Must be ordered to make sure parent is offlined after
	 * children.  The ordering requirement is for memcg where a
	 * parent's offline may wait for a child's leading to deadlock.  In
	 * the long term, this should be fixed from memcg side.
	 *
	 *
	 * We would prefer to do this in cgroup_init() above, but that
	 * We would prefer to do this in cgroup_init() above, but that
	 * is called before init_workqueues(): so leave this until after.
	 * is called before init_workqueues(): so leave this until after.
	 */
	 */
	cgroup_destroy_wq = alloc_ordered_workqueue("cgroup_destroy", 0);
	cgroup_destroy_wq = alloc_workqueue("cgroup_destroy", 0, 1);
	BUG_ON(!cgroup_destroy_wq);
	BUG_ON(!cgroup_destroy_wq);


	/*
	/*