diff options
author | Lai Jiangshan <laijs@cn.fujitsu.com> | 2014-06-03 15:33:27 +0800 |
---|---|---|
committer | Tejun Heo <tj@kernel.org> | 2014-06-19 15:43:33 -0400 |
commit | f05b558d7e0368e0511a098c4beb9654e41246c4 (patch) | |
tree | 46e5a4baeb2fb82ceacfe3e8db263be12b1819ec /kernel/workqueue.c | |
parent | fb009e3a998e48fb33436bf9b563c82801a7329c (diff) | |
download | lwn-f05b558d7e0368e0511a098c4beb9654e41246c4.tar.gz lwn-f05b558d7e0368e0511a098c4beb9654e41246c4.zip |
workqueue: clear POOL_DISASSOCIATED in rebind_workers()
The commit a9ab775bcadf ("workqueue: directly restore CPU affinity of
workers from CPU_ONLINE") moved the pool->lock into rebind_workers()
without also moving "pool->flags &= ~POOL_DISASSOCIATED".
There is nothing wrong with "pool->flags &= ~POOL_DISASSOCIATED" not
being moved together, but there isn't any benefit either. We move it
into rebind_workers() and achieve these benefits:
1) Better readability. POOL_DISASSOCIATED is cleared in
rebind_workers() as expected.
2) When POOL_DISASSOCIATED is cleared, we can ensure that all the
running workers of the pool are on the local CPU (pool->cpu).
tj: Cosmetic updates to the code and description.
Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Signed-off-by: Tejun Heo <tj@kernel.org>
Diffstat (limited to 'kernel/workqueue.c')
-rw-r--r-- | kernel/workqueue.c | 10 |
1 files changed, 3 insertions, 7 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 6203d2900877..129e4c3452ee 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c @@ -4542,6 +4542,7 @@ static void rebind_workers(struct worker_pool *pool) pool->attrs->cpumask) < 0); spin_lock_irq(&pool->lock); + pool->flags &= ~POOL_DISASSOCIATED; for_each_pool_worker(worker, pool) { unsigned int worker_flags = worker->flags; @@ -4643,15 +4644,10 @@ static int workqueue_cpu_up_callback(struct notifier_block *nfb, for_each_pool(pool, pi) { mutex_lock(&pool->attach_mutex); - if (pool->cpu == cpu) { - spin_lock_irq(&pool->lock); - pool->flags &= ~POOL_DISASSOCIATED; - spin_unlock_irq(&pool->lock); - + if (pool->cpu == cpu) rebind_workers(pool); - } else if (pool->cpu < 0) { + else if (pool->cpu < 0) restore_unbound_workers_cpumask(pool, cpu); - } mutex_unlock(&pool->attach_mutex); } |