aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/sched
diff options
context:
space:
mode:
authorQais Yousef <qais.yousef@arm.com>2020-03-05 10:24:50 +0000
committerPeter Zijlstra <peterz@infradead.org>2020-03-20 13:06:20 +0100
commite94f80f6c49020008e6fa0f3d4b806b8595d17d8 (patch)
treeca6a4acf15134ed1f12a178a3a8661d145c3f222 /kernel/sched
parentkthread: Do not preempt current task if it is going to call schedule() (diff)
downloadlinux-dev-e94f80f6c49020008e6fa0f3d4b806b8595d17d8.tar.xz
linux-dev-e94f80f6c49020008e6fa0f3d4b806b8595d17d8.zip
sched/rt: cpupri_find: Trigger a full search as fallback
If we failed to find a fitting CPU, in cpupri_find(), we only fallback to the level we found a hit at. But Steve suggested to fallback to a second full scan instead as this could be a better effort. https://lore.kernel.org/lkml/20200304135404.146c56eb@gandalf.local.home/ We trigger the 2nd search unconditionally since the argument about triggering a full search is that the recorded fall back level might have become empty by then. Which means storing any data about what happened would be meaningless and stale. I had a humble try at timing it and it seemed okay for the small 6 CPUs system I was running on https://lore.kernel.org/lkml/20200305124324.42x6ehjxbnjkklnh@e107158-lin.cambridge.arm.com/ On large system this second full scan could be expensive. But there are no users outside capacity awareness for this fitness function at the moment. Heterogeneous systems tend to be small with 8cores in total. Suggested-by: Steven Rostedt <rostedt@goodmis.org> Signed-off-by: Qais Yousef <qais.yousef@arm.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Reviewed-by: Steven Rostedt (VMware) <rostedt@goodmis.org> Link: https://lkml.kernel.org/r/20200310142219.syxzn5ljpdxqtbgx@e107158-lin.cambridge.arm.com
Diffstat (limited to 'kernel/sched')
-rw-r--r--kernel/sched/cpupri.c29
1 files changed, 6 insertions, 23 deletions
diff --git a/kernel/sched/cpupri.c b/kernel/sched/cpupri.c
index dd3f16d1a04a..0033731a0797 100644
--- a/kernel/sched/cpupri.c
+++ b/kernel/sched/cpupri.c
@@ -122,8 +122,7 @@ int cpupri_find_fitness(struct cpupri *cp, struct task_struct *p,
bool (*fitness_fn)(struct task_struct *p, int cpu))
{
int task_pri = convert_prio(p->prio);
- int best_unfit_idx = -1;
- int idx = 0, cpu;
+ int idx, cpu;
BUG_ON(task_pri >= CPUPRI_NR_PRIORITIES);
@@ -145,31 +144,15 @@ int cpupri_find_fitness(struct cpupri *cp, struct task_struct *p,
* If no CPU at the current priority can fit the task
* continue looking
*/
- if (cpumask_empty(lowest_mask)) {
- /*
- * Store our fallback priority in case we
- * didn't find a fitting CPU
- */
- if (best_unfit_idx == -1)
- best_unfit_idx = idx;
-
+ if (cpumask_empty(lowest_mask))
continue;
- }
return 1;
}
/*
- * If we failed to find a fitting lowest_mask, make sure we fall back
- * to the last known unfitting lowest_mask.
- *
- * Note that the map of the recorded idx might have changed since then,
- * so we must ensure to do the full dance to make sure that level still
- * holds a valid lowest_mask.
- *
- * As per above, the map could have been concurrently emptied while we
- * were busy searching for a fitting lowest_mask at the other priority
- * levels.
+ * If we failed to find a fitting lowest_mask, kick off a new search
+ * but without taking into account any fitness criteria this time.
*
* This rule favours honouring priority over fitting the task in the
* correct CPU (Capacity Awareness being the only user now).
@@ -184,8 +167,8 @@ int cpupri_find_fitness(struct cpupri *cp, struct task_struct *p,
* must do proper RT planning to avoid overloading the system if they
* really care.
*/
- if (best_unfit_idx != -1)
- return __cpupri_find(cp, p, lowest_mask, best_unfit_idx);
+ if (fitness_fn)
+ return cpupri_find(cp, p, lowest_mask);
return 0;
}