|
91 | 91 | #include <trace/events/sched.h> |
92 | 92 | #include "walt.h" |
93 | 93 |
|
94 | | -static bool have_sched_energy_data(void); |
95 | | - |
96 | 94 | DEFINE_MUTEX(sched_domains_mutex); |
97 | 95 | DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues); |
98 | 96 |
|
@@ -195,10 +193,6 @@ static int sched_feat_set(char *cmp) |
195 | 193 | sysctl_sched_features &= ~(1UL << i); |
196 | 194 | sched_feat_disable(i); |
197 | 195 | } else { |
198 | | - if (i == __SCHED_FEAT_ENERGY_AWARE) |
199 | | - WARN(!have_sched_energy_data(), |
200 | | - "Missing sched energy data\n"); |
201 | | - |
202 | 196 | sysctl_sched_features |= (1UL << i); |
203 | 197 | sched_feat_enable(i); |
204 | 198 | } |
@@ -6655,19 +6649,6 @@ static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) |
6655 | 6649 | atomic_set(&sg->sgc->nr_busy_cpus, sg->group_weight); |
6656 | 6650 | } |
6657 | 6651 |
|
6658 | | -static bool have_sched_energy_data(void) |
6659 | | -{ |
6660 | | - int cpu; |
6661 | | - |
6662 | | - for_each_possible_cpu(cpu) { |
6663 | | - if (!rcu_dereference(per_cpu(sd_scs, cpu)) || |
6664 | | - !rcu_dereference(per_cpu(sd_ea, cpu))) |
6665 | | - return false; |
6666 | | - } |
6667 | | - |
6668 | | - return true; |
6669 | | -} |
6670 | | - |
6671 | 6652 | /* |
6672 | 6653 | * Check that the per-cpu provided sd energy data is consistent for all cpus |
6673 | 6654 | * within the mask. |
@@ -7480,9 +7461,6 @@ static int build_sched_domains(const struct cpumask *cpu_map, |
7480 | 7461 | } |
7481 | 7462 | rcu_read_unlock(); |
7482 | 7463 |
|
7483 | | - WARN(sched_feat(ENERGY_AWARE) && !have_sched_energy_data(), |
7484 | | - "Missing data for energy aware scheduling\n"); |
7485 | | - |
7486 | 7464 | ret = 0; |
7487 | 7465 | error: |
7488 | 7466 | __free_domain_allocs(&d, alloc_state, cpu_map); |
|
0 commit comments