From 7cf6a7300bb9a88f543061270419427395ab4d2f Mon Sep 17 00:00:00 2001 From: Chris Redpath Date: Fri, 22 Nov 2013 13:19:18 +0000 Subject: sched: hmp: Fix build breakage when not using CONFIG_SCHED_HMP hmp_variable_scale_convert was used without guards in __update_entity_runnable_avg. Guard it. Signed-off-by: Chris Redpath Signed-off-by: Mark Brown --- kernel/sched/fair.c | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index c7d808ee0a3..8a4a02740f0 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -1210,6 +1210,7 @@ static u32 __compute_runnable_contrib(u64 n) return contrib + runnable_avg_yN_sum[n]; } +#ifdef CONFIG_SCHED_HMP #define HMP_VARIABLE_SCALE_SHIFT 16ULL struct hmp_global_attr { struct attribute attr; @@ -1291,6 +1292,7 @@ struct cpufreq_extents { static struct cpufreq_extents freq_scale[CONFIG_NR_CPUS]; #endif /* CONFIG_HMP_FREQUENCY_INVARIANT_SCALE */ +#endif /* CONFIG_SCHED_HMP */ /* We can represent the historical contribution to runnable average as the * coefficients of a geometric series. To do this we sub-divide our runnable @@ -1336,8 +1338,9 @@ static __always_inline int __update_entity_runnable_avg(u64 now, #endif /* CONFIG_HMP_FREQUENCY_INVARIANT_SCALE */ delta = now - sa->last_runnable_update; - +#ifdef CONFIG_SCHED_HMP delta = hmp_variable_scale_convert(delta); +#endif /* * This should only happen when time goes backwards, which it * unfortunately does during sched clock init when we swap over to TSC. -- cgit v1.2.3