workqueue: Use cpu_possible_mask instead of cpu_active_mask to break affinity
authorLai Jiangshan <laijs@linux.alibaba.com>
Mon, 11 Jan 2021 15:26:33 +0000 (23:26 +0800)
committerPeter Zijlstra <peterz@infradead.org>
Fri, 22 Jan 2021 14:09:41 +0000 (15:09 +0100)
The scheduler won't break affinity for us any more, and we should
"emulate" the same behavior when the scheduler breaks affinity for
us.  The behavior is "changing the cpumask to cpu_possible_mask".

And there might be some other CPUs online later while the worker is
still running with the pending work items.  The worker should be allowed
to use the later online CPUs as before and process the work items ASAP.
If we use cpu_active_mask here, we can't achieve this goal but
using cpu_possible_mask can.

Fixes: 06249738a41a ("workqueue: Manually break affinity on hotplug")
Signed-off-by: Lai Jiangshan <laijs@linux.alibaba.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Acked-by: Tejun Heo <tj@kernel.org>
Tested-by: Paul E. McKenney <paulmck@kernel.org>
Tested-by: Valentin Schneider <valentin.schneider@arm.com>
Link: https://lkml.kernel.org/r/20210111152638.2417-4-jiangshanlai@gmail.com
kernel/workqueue.c

index 9880b6c0e2721fe5c0758eda58d82c1f36576d19..1646331546eb638e23d8cfef43a52d93b4fe2d28 100644 (file)
@@ -4920,7 +4920,7 @@ static void unbind_workers(int cpu)
                raw_spin_unlock_irq(&pool->lock);
 
                for_each_pool_worker(worker, pool)
-                       WARN_ON_ONCE(set_cpus_allowed_ptr(worker->task, cpu_active_mask) < 0);
+                       WARN_ON_ONCE(set_cpus_allowed_ptr(worker->task, cpu_possible_mask) < 0);
 
                mutex_unlock(&wq_pool_attach_mutex);