diff options
author | Suresh Siddha <suresh.b.siddha@intel.com> | 2007-07-19 21:28:35 +0200 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2007-07-19 21:28:35 +0200 |
commit | 969bb4e4032dac67287951d8f6642a3b5119694e (patch) | |
tree | 2056749332226e750394001ec29d43514c14d0b6 | |
parent | 9439aab8dbc33c2c03c3a19dba267360383ba38c (diff) | |
download | lwn-969bb4e4032dac67287951d8f6642a3b5119694e.tar.gz lwn-969bb4e4032dac67287951d8f6642a3b5119694e.zip |
[PATCH] sched: fix the all pinned logic in load_balance_newidle()
nr_moved is not the correct check for triggering all pinned logic. Fix
the all pinned logic in the case of load_balance_newidle().
Signed-off-by: Suresh Siddha <suresh.b.siddha@intel.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r-- | kernel/sched.c | 6 |
1 files changed, 4 insertions, 2 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index e36d99d1ddb1..a35a92ff38fd 100644 --- a/kernel/sched.c +++ b/kernel/sched.c @@ -2679,6 +2679,7 @@ load_balance_newidle(int this_cpu, struct rq *this_rq, struct sched_domain *sd) unsigned long imbalance; int nr_moved = 0; int sd_idle = 0; + int all_pinned = 0; cpumask_t cpus = CPU_MASK_ALL; /* @@ -2717,10 +2718,11 @@ redo: double_lock_balance(this_rq, busiest); nr_moved = move_tasks(this_rq, this_cpu, busiest, minus_1_or_zero(busiest->nr_running), - imbalance, sd, CPU_NEWLY_IDLE, NULL); + imbalance, sd, CPU_NEWLY_IDLE, + &all_pinned); spin_unlock(&busiest->lock); - if (!nr_moved) { + if (unlikely(all_pinned)) { cpu_clear(cpu_of(busiest), cpus); if (!cpus_empty(cpus)) goto redo; |