sched/numa: Move power adjustment into load_too_imbalanced()
authorRik van Riel <riel@redhat.com>
Mon, 23 Jun 2014 15:46:13 +0000 (11:46 -0400)
committerIngo Molnar <mingo@kernel.org>
Sat, 5 Jul 2014 09:17:34 +0000 (11:17 +0200)
commit28a21745190a0ca613cab817bfe3dc65373158bf
tree258f0b418b980f3cc19ec7762487bc15b35a59e1
parentf0b8a4afd6a8c500161e45065a91738b490bf5ae
sched/numa: Move power adjustment into load_too_imbalanced()

Currently the NUMA code scales the load on each node with the
amount of CPU power available on that node, but it does not
apply any adjustment to the load of the task that is being
moved over.

On systems with SMT/HT, this results in a task being weighed
much more heavily than a CPU core, and a task move that would
even out the load between nodes being disallowed.

The correct thing is to apply the power correction to the
numbers after we have first applied the move of the tasks'
loads to them.

This also allows us to do the power correction with a multiplication,
rather than a division.

Also drop two function arguments for load_too_unbalanced, since it
takes various factors from env already.

Signed-off-by: Rik van Riel <riel@redhat.com>
Cc: chegu_vinod@hp.com
Cc: mgorman@suse.de
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Link: http://lkml.kernel.org/r/1403538378-31571-2-git-send-email-riel@redhat.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
kernel/sched/fair.c