diff options
author | Suresh Siddha <suresh.b.siddha@intel.com> | 2011-02-10 09:52:07 +0100 |
---|---|---|
committer | AK <andi@firstfloor.org> | 2011-03-31 11:57:57 -0700 |
commit | 25f24a9de3c47261441bfa3d9cccae3616f42893 (patch) | |
tree | 9a092af29390a97627c85ad0a1938277125fdea5 | |
parent | 824a6dcef20b339735b0f0e89841fcf2fbbe2621 (diff) |
sched: Move sched_avg_update() to update_cpu_load()
Commit: da2b71edd8a7db44fe1746261410a981f3e03632 upstream
Currently sched_avg_update() (which updates rt_avg stats in the rq)
is getting called from scale_rt_power() (in the load balance context)
which doesn't take rq->lock.
Fix it by moving the sched_avg_update() to more appropriate
update_cpu_load() where the CFS load gets updated as well.
Signed-off-by: Suresh Siddha <suresh.b.siddha@intel.com>
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Andi Kleen <ak@linux.intel.com>
LKML-Reference: <1282596171.2694.3.camel@sbsiddha-MOBL3>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Mike Galbraith <efault@gmx.de>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
-rw-r--r-- | kernel/sched.c | 6 | ||||
-rw-r--r-- | kernel/sched_fair.c | 2 |
2 files changed, 6 insertions, 2 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index d4f71b01136..1747071e26f 100644 --- a/kernel/sched.c +++ b/kernel/sched.c @@ -1270,6 +1270,10 @@ static void resched_task(struct task_struct *p) static void sched_rt_avg_update(struct rq *rq, u64 rt_delta) { } + +static void sched_avg_update(struct rq *rq) +{ +} #endif /* CONFIG_SMP */ #if BITS_PER_LONG == 32 @@ -3152,6 +3156,8 @@ static void update_cpu_load(struct rq *this_rq) } calc_load_account_active(this_rq); + + sched_avg_update(this_rq); } #ifdef CONFIG_SMP diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index a878b5332da..44d7bdc800f 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c @@ -2268,8 +2268,6 @@ unsigned long scale_rt_power(int cpu) struct rq *rq = cpu_rq(cpu); u64 total, available; - sched_avg_update(rq); - total = sched_avg_period() + (rq->clock - rq->age_stamp); available = total - rq->rt_avg; |