mirror of
https://github.com/Fishwaldo/Star64_linux.git
synced 2025-04-27 16:53:57 +00:00
sched/core: Optimize SCHED_SMT
Avoid pointless SCHED_SMT code when running on !SMT hardware. Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Mike Galbraith <efault@gmx.de> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: linux-kernel@vger.kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
parent
10e2f1acd0
commit
1b568f0aab
3 changed files with 43 additions and 7 deletions
|
@ -7412,6 +7412,22 @@ int sched_cpu_dying(unsigned int cpu)
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
#ifdef CONFIG_SCHED_SMT
|
||||||
|
DEFINE_STATIC_KEY_FALSE(sched_smt_present);
|
||||||
|
|
||||||
|
static void sched_init_smt(void)
|
||||||
|
{
|
||||||
|
/*
|
||||||
|
* We've enumerated all CPUs and will assume that if any CPU
|
||||||
|
* has SMT siblings, CPU0 will too.
|
||||||
|
*/
|
||||||
|
if (cpumask_weight(cpu_smt_mask(0)) > 1)
|
||||||
|
static_branch_enable(&sched_smt_present);
|
||||||
|
}
|
||||||
|
#else
|
||||||
|
static inline void sched_init_smt(void) { }
|
||||||
|
#endif
|
||||||
|
|
||||||
void __init sched_init_smp(void)
|
void __init sched_init_smp(void)
|
||||||
{
|
{
|
||||||
cpumask_var_t non_isolated_cpus;
|
cpumask_var_t non_isolated_cpus;
|
||||||
|
@ -7441,6 +7457,9 @@ void __init sched_init_smp(void)
|
||||||
|
|
||||||
init_sched_rt_class();
|
init_sched_rt_class();
|
||||||
init_sched_dl_class();
|
init_sched_dl_class();
|
||||||
|
|
||||||
|
sched_init_smt();
|
||||||
|
|
||||||
sched_smp_initialized = true;
|
sched_smp_initialized = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -5357,7 +5357,7 @@ static inline bool test_idle_cores(int cpu, bool def)
|
||||||
* Since SMT siblings share all cache levels, inspecting this limited remote
|
* Since SMT siblings share all cache levels, inspecting this limited remote
|
||||||
* state should be fairly cheap.
|
* state should be fairly cheap.
|
||||||
*/
|
*/
|
||||||
void update_idle_core(struct rq *rq)
|
void __update_idle_core(struct rq *rq)
|
||||||
{
|
{
|
||||||
int core = cpu_of(rq);
|
int core = cpu_of(rq);
|
||||||
int cpu;
|
int cpu;
|
||||||
|
@ -5389,6 +5389,9 @@ static int select_idle_core(struct task_struct *p, struct sched_domain *sd, int
|
||||||
struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask);
|
struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_idle_mask);
|
||||||
int core, cpu, wrap;
|
int core, cpu, wrap;
|
||||||
|
|
||||||
|
if (!static_branch_likely(&sched_smt_present))
|
||||||
|
return -1;
|
||||||
|
|
||||||
if (!test_idle_cores(target, false))
|
if (!test_idle_cores(target, false))
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
|
@ -5422,6 +5425,9 @@ static int select_idle_smt(struct task_struct *p, struct sched_domain *sd, int t
|
||||||
{
|
{
|
||||||
int cpu;
|
int cpu;
|
||||||
|
|
||||||
|
if (!static_branch_likely(&sched_smt_present))
|
||||||
|
return -1;
|
||||||
|
|
||||||
for_each_cpu(cpu, cpu_smt_mask(target)) {
|
for_each_cpu(cpu, cpu_smt_mask(target)) {
|
||||||
if (!cpumask_test_cpu(cpu, tsk_cpus_allowed(p)))
|
if (!cpumask_test_cpu(cpu, tsk_cpus_allowed(p)))
|
||||||
continue;
|
continue;
|
||||||
|
|
|
@ -36,12 +36,6 @@ extern void cpu_load_update_active(struct rq *this_rq);
|
||||||
static inline void cpu_load_update_active(struct rq *this_rq) { }
|
static inline void cpu_load_update_active(struct rq *this_rq) { }
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_SCHED_SMT
|
|
||||||
extern void update_idle_core(struct rq *rq);
|
|
||||||
#else
|
|
||||||
static inline void update_idle_core(struct rq *rq) { }
|
|
||||||
#endif
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Helpers for converting nanosecond timing to jiffy resolution
|
* Helpers for converting nanosecond timing to jiffy resolution
|
||||||
*/
|
*/
|
||||||
|
@ -730,6 +724,23 @@ static inline int cpu_of(struct rq *rq)
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
#ifdef CONFIG_SCHED_SMT
|
||||||
|
|
||||||
|
extern struct static_key_false sched_smt_present;
|
||||||
|
|
||||||
|
extern void __update_idle_core(struct rq *rq);
|
||||||
|
|
||||||
|
static inline void update_idle_core(struct rq *rq)
|
||||||
|
{
|
||||||
|
if (static_branch_unlikely(&sched_smt_present))
|
||||||
|
__update_idle_core(rq);
|
||||||
|
}
|
||||||
|
|
||||||
|
#else
|
||||||
|
static inline void update_idle_core(struct rq *rq) { }
|
||||||
|
#endif
|
||||||
|
|
||||||
DECLARE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues);
|
DECLARE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues);
|
||||||
|
|
||||||
#define cpu_rq(cpu) (&per_cpu(runqueues, (cpu)))
|
#define cpu_rq(cpu) (&per_cpu(runqueues, (cpu)))
|
||||||
|
|
Loading…
Add table
Reference in a new issue