|
@@ -146,36 +146,6 @@ static inline bool policy_is_shared(struct cpufreq_policy *policy)
|
|
extern struct kobject *cpufreq_global_kobject;
|
|
extern struct kobject *cpufreq_global_kobject;
|
|
|
|
|
|
#ifdef CONFIG_CPU_FREQ
|
|
#ifdef CONFIG_CPU_FREQ
|
|
-void cpufreq_update_util(u64 time, unsigned long util, unsigned long max);
|
|
|
|
-
|
|
|
|
-/**
|
|
|
|
- * cpufreq_trigger_update - Trigger CPU performance state evaluation if needed.
|
|
|
|
- * @time: Current time.
|
|
|
|
- *
|
|
|
|
- * The way cpufreq is currently arranged requires it to evaluate the CPU
|
|
|
|
- * performance state (frequency/voltage) on a regular basis to prevent it from
|
|
|
|
- * being stuck in a completely inadequate performance level for too long.
|
|
|
|
- * That is not guaranteed to happen if the updates are only triggered from CFS,
|
|
|
|
- * though, because they may not be coming in if RT or deadline tasks are active
|
|
|
|
- * all the time (or there are RT and DL tasks only).
|
|
|
|
- *
|
|
|
|
- * As a workaround for that issue, this function is called by the RT and DL
|
|
|
|
- * sched classes to trigger extra cpufreq updates to prevent it from stalling,
|
|
|
|
- * but that really is a band-aid. Going forward it should be replaced with
|
|
|
|
- * solutions targeted more specifically at RT and DL tasks.
|
|
|
|
- */
|
|
|
|
-static inline void cpufreq_trigger_update(u64 time)
|
|
|
|
-{
|
|
|
|
- cpufreq_update_util(time, ULONG_MAX, 0);
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
-struct update_util_data {
|
|
|
|
- void (*func)(struct update_util_data *data,
|
|
|
|
- u64 time, unsigned long util, unsigned long max);
|
|
|
|
-};
|
|
|
|
-
|
|
|
|
-void cpufreq_set_update_util_data(int cpu, struct update_util_data *data);
|
|
|
|
-
|
|
|
|
unsigned int cpufreq_get(unsigned int cpu);
|
|
unsigned int cpufreq_get(unsigned int cpu);
|
|
unsigned int cpufreq_quick_get(unsigned int cpu);
|
|
unsigned int cpufreq_quick_get(unsigned int cpu);
|
|
unsigned int cpufreq_quick_get_max(unsigned int cpu);
|
|
unsigned int cpufreq_quick_get_max(unsigned int cpu);
|
|
@@ -187,10 +157,6 @@ int cpufreq_update_policy(unsigned int cpu);
|
|
bool have_governor_per_policy(void);
|
|
bool have_governor_per_policy(void);
|
|
struct kobject *get_governor_parent_kobj(struct cpufreq_policy *policy);
|
|
struct kobject *get_governor_parent_kobj(struct cpufreq_policy *policy);
|
|
#else
|
|
#else
|
|
-static inline void cpufreq_update_util(u64 time, unsigned long util,
|
|
|
|
- unsigned long max) {}
|
|
|
|
-static inline void cpufreq_trigger_update(u64 time) {}
|
|
|
|
-
|
|
|
|
static inline unsigned int cpufreq_get(unsigned int cpu)
|
|
static inline unsigned int cpufreq_get(unsigned int cpu)
|
|
{
|
|
{
|
|
return 0;
|
|
return 0;
|