diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2008-06-27 13:41:12 +0200 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-06-27 14:31:28 +0200 |
commit | c9c294a630e28eec5f2865f028ecfc58d45c0a5a (patch) | |
tree | 62fcfc395e98ff7a4ff6bee3364d0bb86024b171 /kernel/sched_features.h | |
parent | a7be37ac8e1565e00880531f4e2aff421a21c803 (diff) | |
download | linux-c9c294a630e28eec5f2865f028ecfc58d45c0a5a.tar.gz linux-c9c294a630e28eec5f2865f028ecfc58d45c0a5a.tar.bz2 linux-c9c294a630e28eec5f2865f028ecfc58d45c0a5a.zip |
sched: fix calc_delta_asym()
calc_delta_asym() is supposed to do the same as calc_delta_fair() except
linearly shrink the result for negative nice processes - this causes them
to have a smaller preemption threshold so that they are more easily preempted.
The problem is that for task groups se->load.weight is the per cpu share of
the actual task group weight; take that into account.
Also provide a debug switch to disable the asymmetry (which I still don't
like - but it does greatly benefit some workloads)
This would explain the interactivity issues reported against group scheduling.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Cc: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_features.h')
-rw-r--r-- | kernel/sched_features.h | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/kernel/sched_features.h b/kernel/sched_features.h index afa549166d8d..04123af2e678 100644 --- a/kernel/sched_features.h +++ b/kernel/sched_features.h @@ -7,3 +7,4 @@ SCHED_FEAT(CACHE_HOT_BUDDY, 1) SCHED_FEAT(SYNC_WAKEUPS, 1) SCHED_FEAT(HRTICK, 1) SCHED_FEAT(DOUBLE_TICK, 0) +SCHED_FEAT(ASYM_GRAN, 1) |