@@ -6638,6 +6638,8 @@ out:
return ld_moved;
}
+static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask);
+
/*
* idle_balance is called by schedule() if this_cpu is about to become
* idle. Attempts to pull tasks from other CPUs.
@@ -7692,6 +7694,12 @@ void print_cfs_stats(struct seq_file *m, int cpu)
__init void init_sched_fair_class(void)
{
#ifdef CONFIG_SMP
+ unsigned int i;
+ for_each_possible_cpu(i) {
+ zalloc_cpumask_var_node(&per_cpu(local_cpu_mask, i),
+ GFP_KERNEL, cpu_to_node(i));
+ }
+
open_softirq(SCHED_SOFTIRQ, run_rebalance_domains);
#ifdef CONFIG_NO_HZ_COMMON
We need these cpumasks to aid in cosolidated load balancing Signed-off-by: Yuyang Du <yuyang.du@intel.com> --- kernel/sched/fair.c | 8 ++++++++ 1 file changed, 8 insertions(+)