From: Chris Redpath Date: Fri, 24 Mar 2017 17:37:28 +0000 (+0000) Subject: cpufreq/schedutil: use boosted_cpu_util for PELT to match WALT X-Git-Tag: release-20171130_firefly~4^2~100^2~26 X-Git-Url: http://demsky.eecs.uci.edu/git/?a=commitdiff_plain;h=1ffc5888b534e21db4a3560574d437077be300b9;p=firefly-linux-kernel-4.4.55.git cpufreq/schedutil: use boosted_cpu_util for PELT to match WALT When using WALT we always used boosted cpu util for OPP selection. This is the primary purpose for boosted cpu util, but we hadn't changed the PELT utilization check to do the same thing. Fix that here. Change-Id: Id5ffb26eac23b25fe754255221f6d21b8cededfd Signed-off-by: Patrick Bellasi Signed-off-by: Chris Redpath --- diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c index 191ba36a9eeb..75bfbb336722 100644 --- a/kernel/sched/cpufreq_schedutil.c +++ b/kernel/sched/cpufreq_schedutil.c @@ -19,9 +19,7 @@ #include "sched.h" #include "tune.h" -#ifdef CONFIG_SCHED_WALT unsigned long boosted_cpu_util(int cpu); -#endif /* Stub out fast switch routines present on mainline to reduce the backport * overhead. */ @@ -188,6 +186,15 @@ static unsigned int get_next_freq(struct sugov_cpu *sg_cpu, unsigned long util, return cpufreq_driver_resolve_freq(policy, freq); } +static inline bool use_pelt(void) +{ +#ifdef CONFIG_SCHED_WALT + return (!sysctl_sched_use_walt_cpu_util || walt_disabled); +#else + return true; +#endif +} + static void sugov_get_util(unsigned long *util, unsigned long *max, u64 time) { int cpu = smp_processor_id(); @@ -204,11 +211,10 @@ static void sugov_get_util(unsigned long *util, unsigned long *max, u64 time) rt = div64_u64(rq->rt_avg, sched_avg_period() + delta); rt = (rt * max_cap) >> SCHED_CAPACITY_SHIFT; - *util = min(rq->cfs.avg.util_avg + rt, max_cap); -#ifdef CONFIG_SCHED_WALT - if (!walt_disabled && sysctl_sched_use_walt_cpu_util) - *util = boosted_cpu_util(cpu); -#endif + *util = boosted_cpu_util(cpu); + if (likely(use_pelt())) + *util = min((*util + rt), max_cap); + *max = max_cap; }