sched/fair: Improve PELT stuff some more
authorPeter Zijlstra <peterz@infradead.org>
Wed, 13 Jul 2016 08:56:25 +0000 (10:56 +0200)
committerIngo Molnar <mingo@kernel.org>
Wed, 10 Aug 2016 12:03:32 +0000 (14:03 +0200)
commit7c3edd2c300b7ef2005a69dc727692ee07434aa5
tree19856ee335a4065233094aeb225b66986738dea4
parenta1fd46565bea62840a24bee7b7c60f65bb12bd21
sched/fair: Improve PELT stuff some more

Vincent noted that the update_tg_load_avg() usage in commit:

  3d30544f0212 ("sched/fair: Apply more PELT fixes")

isn't entirely sufficient. We need to call this function every time
cfs_rq->avg.load changes, this includes when update_cfs_rq_load_avg()
returns true, but {attach,detach}_entity_load_avg() themselves also
change it. This means we need to unconditionally call
update_tg_load_avg().

Also, add more comments.

Reported-by: Vincent Guittot <vincent.guittot@linaro.org>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
kernel/sched/fair.c