cpufreq_ondemand.c 17.1 KB
Newer Older
L
Linus Torvalds 已提交
1 2 3 4 5 6 7 8 9 10 11 12
/*
 *  drivers/cpufreq/cpufreq_ondemand.c
 *
 *  Copyright (C)  2001 Russell King
 *            (C)  2003 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>.
 *                      Jun Nakajima <jun.nakajima@intel.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */

13 14
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

L
Linus Torvalds 已提交
15
#include <linux/cpufreq.h>
16 17
#include <linux/init.h>
#include <linux/kernel.h>
L
Linus Torvalds 已提交
18
#include <linux/kernel_stat.h>
19 20
#include <linux/kobject.h>
#include <linux/module.h>
21
#include <linux/mutex.h>
22
#include <linux/percpu-defs.h>
23
#include <linux/slab.h>
24
#include <linux/sysfs.h>
25
#include <linux/tick.h>
26
#include <linux/types.h>
L
Linus Torvalds 已提交
27

28
#include "cpufreq_governor.h"
L
Linus Torvalds 已提交
29

30
/* On-demand governor macros */
31
#define DEF_FREQUENCY_DOWN_DIFFERENTIAL		(10)
L
Linus Torvalds 已提交
32
#define DEF_FREQUENCY_UP_THRESHOLD		(80)
33 34
#define DEF_SAMPLING_DOWN_FACTOR		(1)
#define MAX_SAMPLING_DOWN_FACTOR		(100000)
35 36
#define MICRO_FREQUENCY_DOWN_DIFFERENTIAL	(3)
#define MICRO_FREQUENCY_UP_THRESHOLD		(95)
37
#define MICRO_FREQUENCY_MIN_SAMPLE_RATE		(10000)
38
#define MIN_FREQUENCY_UP_THRESHOLD		(11)
L
Linus Torvalds 已提交
39 40
#define MAX_FREQUENCY_UP_THRESHOLD		(100)

41
static DEFINE_PER_CPU(struct od_cpu_dbs_info_s, od_cpu_dbs_info);
L
Linus Torvalds 已提交
42

43 44 45 46
#ifndef CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND
static struct cpufreq_governor cpufreq_gov_ondemand;
#endif

47
static void ondemand_powersave_bias_init_cpu(int cpu)
48
{
49
	struct od_cpu_dbs_info_s *dbs_info = &per_cpu(od_cpu_dbs_info, cpu);
50

51 52 53
	dbs_info->freq_table = cpufreq_frequency_get_table(cpu);
	dbs_info->freq_lo = 0;
}
54

55 56 57 58 59
/*
 * Not all CPUs want IO time to be accounted as busy; this depends on how
 * efficient idling at a higher frequency/voltage is.
 * Pavel Machek says this is not so for various generations of AMD and old
 * Intel systems.
60
 * Mike Chan (android.com) claims this is also not true for ARM.
61 62 63 64 65 66 67
 * Because of this, whitelist specific known (series) of CPUs by default, and
 * leave all others up to the user.
 */
static int should_io_be_busy(void)
{
#if defined(CONFIG_X86)
	/*
68
	 * For Intel, Core 2 (model 15) and later have an efficient idle.
69 70 71 72 73 74 75
	 */
	if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL &&
			boot_cpu_data.x86 == 6 &&
			boot_cpu_data.x86_model >= 15)
		return 1;
#endif
	return 0;
76 77
}

78 79 80 81 82
/*
 * Find right freq to be set now with powersave_bias on.
 * Returns the freq_hi to be used right now and will set freq_hi_jiffies,
 * freq_lo, and freq_lo_jiffies in percpu area for averaging freqs.
 */
83
static unsigned int powersave_bias_target(struct cpufreq_policy *policy,
84
		unsigned int freq_next, unsigned int relation)
85 86 87 88 89
{
	unsigned int freq_req, freq_reduc, freq_avg;
	unsigned int freq_hi, freq_lo;
	unsigned int index = 0;
	unsigned int jiffies_total, jiffies_hi, jiffies_lo;
90
	struct od_cpu_dbs_info_s *dbs_info = &per_cpu(od_cpu_dbs_info,
91
						   policy->cpu);
92 93
	struct dbs_data *dbs_data = policy->governor_data;
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
94 95 96 97 98 99 100 101 102 103

	if (!dbs_info->freq_table) {
		dbs_info->freq_lo = 0;
		dbs_info->freq_lo_jiffies = 0;
		return freq_next;
	}

	cpufreq_frequency_table_target(policy, dbs_info->freq_table, freq_next,
			relation, &index);
	freq_req = dbs_info->freq_table[index].frequency;
104
	freq_reduc = freq_req * od_tuners->powersave_bias / 1000;
105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122
	freq_avg = freq_req - freq_reduc;

	/* Find freq bounds for freq_avg in freq_table */
	index = 0;
	cpufreq_frequency_table_target(policy, dbs_info->freq_table, freq_avg,
			CPUFREQ_RELATION_H, &index);
	freq_lo = dbs_info->freq_table[index].frequency;
	index = 0;
	cpufreq_frequency_table_target(policy, dbs_info->freq_table, freq_avg,
			CPUFREQ_RELATION_L, &index);
	freq_hi = dbs_info->freq_table[index].frequency;

	/* Find out how long we have to be in hi and lo freqs */
	if (freq_hi == freq_lo) {
		dbs_info->freq_lo = 0;
		dbs_info->freq_lo_jiffies = 0;
		return freq_lo;
	}
123
	jiffies_total = usecs_to_jiffies(od_tuners->sampling_rate);
124 125 126 127 128 129 130 131 132 133 134 135 136 137
	jiffies_hi = (freq_avg - freq_lo) * jiffies_total;
	jiffies_hi += ((freq_hi - freq_lo) / 2);
	jiffies_hi /= (freq_hi - freq_lo);
	jiffies_lo = jiffies_total - jiffies_hi;
	dbs_info->freq_lo = freq_lo;
	dbs_info->freq_lo_jiffies = jiffies_lo;
	dbs_info->freq_hi_jiffies = jiffies_hi;
	return freq_hi;
}

static void ondemand_powersave_bias_init(void)
{
	int i;
	for_each_online_cpu(i) {
138
		ondemand_powersave_bias_init_cpu(i);
139 140 141
	}
}

142 143
static void dbs_freq_increase(struct cpufreq_policy *p, unsigned int freq)
{
144 145 146 147
	struct dbs_data *dbs_data = p->governor_data;
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;

	if (od_tuners->powersave_bias)
148 149 150
		freq = powersave_bias_target(p, freq, CPUFREQ_RELATION_H);
	else if (p->cur == p->max)
		return;
151

152
	__cpufreq_driver_target(p, freq, od_tuners->powersave_bias ?
153 154 155 156 157
			CPUFREQ_RELATION_L : CPUFREQ_RELATION_H);
}

/*
 * Every sampling_rate, we check, if current idle time is less than 20%
158 159
 * (default), then we try to increase frequency. Every sampling_rate, we look
 * for the lowest frequency which can sustain the load while keeping idle time
160 161 162 163 164 165
 * over 30%. If such a frequency exist, we try to decrease to this frequency.
 *
 * Any frequency increase takes it to the maximum frequency. Frequency reduction
 * happens at minimum steps of 5% (default) of current frequency
 */
static void od_check_cpu(int cpu, unsigned int load_freq)
L
Linus Torvalds 已提交
166
{
167 168
	struct od_cpu_dbs_info_s *dbs_info = &per_cpu(od_cpu_dbs_info, cpu);
	struct cpufreq_policy *policy = dbs_info->cdbs.cur_policy;
169 170
	struct dbs_data *dbs_data = policy->governor_data;
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
171 172 173 174

	dbs_info->freq_lo = 0;

	/* Check for frequency increase */
175
	if (load_freq > od_tuners->up_threshold * policy->cur) {
176 177 178
		/* If switching to max speed, apply sampling_down_factor */
		if (policy->cur < policy->max)
			dbs_info->rate_mult =
179
				od_tuners->sampling_down_factor;
180 181 182 183 184 185 186 187 188 189 190 191 192 193
		dbs_freq_increase(policy, policy->max);
		return;
	}

	/* Check for frequency decrease */
	/* if we cannot reduce the frequency anymore, break out early */
	if (policy->cur == policy->min)
		return;

	/*
	 * The optimal frequency is the frequency that is the lowest that can
	 * support the current CPU usage without triggering the up policy. To be
	 * safe, we focus 10 points under the threshold.
	 */
194 195
	if (load_freq < od_tuners->adj_up_threshold
			* policy->cur) {
196
		unsigned int freq_next;
197
		freq_next = load_freq / od_tuners->adj_up_threshold;
198 199 200 201 202 203 204

		/* No longer fully busy, reset rate_mult */
		dbs_info->rate_mult = 1;

		if (freq_next < policy->min)
			freq_next = policy->min;

205
		if (!od_tuners->powersave_bias) {
206 207 208 209 210 211 212 213 214
			__cpufreq_driver_target(policy, freq_next,
					CPUFREQ_RELATION_L);
		} else {
			int freq = powersave_bias_target(policy, freq_next,
					CPUFREQ_RELATION_L);
			__cpufreq_driver_target(policy, freq,
					CPUFREQ_RELATION_L);
		}
	}
L
Linus Torvalds 已提交
215 216
}

217
static void od_dbs_timer(struct work_struct *work)
218
{
219 220
	struct od_cpu_dbs_info_s *dbs_info =
		container_of(work, struct od_cpu_dbs_info_s, cdbs.work.work);
221
	unsigned int cpu = dbs_info->cdbs.cur_policy->cpu;
222 223
	struct od_cpu_dbs_info_s *core_dbs_info = &per_cpu(od_cpu_dbs_info,
			cpu);
224 225
	struct dbs_data *dbs_data = dbs_info->cdbs.cur_policy->governor_data;
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
226
	int delay = 0, sample_type = core_dbs_info->sample_type;
227
	bool modify_all = true;
228 229

	mutex_lock(&core_dbs_info->cdbs.timer_mutex);
230 231
	if (!need_load_eval(&core_dbs_info->cdbs, od_tuners->sampling_rate)) {
		modify_all = false;
232
		goto max_delay;
233
	}
L
Linus Torvalds 已提交
234

235
	/* Common NORMAL_SAMPLE setup */
236
	core_dbs_info->sample_type = OD_NORMAL_SAMPLE;
237
	if (sample_type == OD_SUB_SAMPLE) {
238
		delay = core_dbs_info->freq_lo_jiffies;
239 240
		__cpufreq_driver_target(core_dbs_info->cdbs.cur_policy,
				core_dbs_info->freq_lo, CPUFREQ_RELATION_H);
241
	} else {
242
		dbs_check_cpu(dbs_data, cpu);
243
		if (core_dbs_info->freq_lo) {
244
			/* Setup timer for SUB_SAMPLE */
245 246
			core_dbs_info->sample_type = OD_SUB_SAMPLE;
			delay = core_dbs_info->freq_hi_jiffies;
247 248 249
		}
	}

250 251 252 253 254
max_delay:
	if (!delay)
		delay = delay_for_sampling_rate(od_tuners->sampling_rate
				* core_dbs_info->rate_mult);

255
	gov_queue_work(dbs_data, dbs_info->cdbs.cur_policy, delay, modify_all);
256
	mutex_unlock(&core_dbs_info->cdbs.timer_mutex);
257 258
}

259
/************************** sysfs interface ************************/
260
static struct common_dbs_data od_dbs_cdata;
L
Linus Torvalds 已提交
261

262 263 264 265
/**
 * update_sampling_rate - update sampling rate effective immediately if needed.
 * @new_rate: new sampling rate
 *
266
 * If new rate is smaller than the old, simply updating
267 268 269 270 271 272 273
 * dbs_tuners_int.sampling_rate might not be appropriate. For example, if the
 * original sampling_rate was 1 second and the requested new sampling rate is 10
 * ms because the user needs immediate reaction from ondemand governor, but not
 * sure if higher frequency will be required or not, then, the governor may
 * change the sampling rate too late; up to 1 second later. Thus, if we are
 * reducing the sampling rate, we need to make the new value effective
 * immediately.
274
 */
275 276
static void update_sampling_rate(struct dbs_data *dbs_data,
		unsigned int new_rate)
277
{
278
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
279 280
	int cpu;

281 282
	od_tuners->sampling_rate = new_rate = max(new_rate,
			dbs_data->min_sampling_rate);
283 284 285

	for_each_online_cpu(cpu) {
		struct cpufreq_policy *policy;
286
		struct od_cpu_dbs_info_s *dbs_info;
287 288 289 290 291
		unsigned long next_sampling, appointed_at;

		policy = cpufreq_cpu_get(cpu);
		if (!policy)
			continue;
292 293 294 295
		if (policy->governor != &cpufreq_gov_ondemand) {
			cpufreq_cpu_put(policy);
			continue;
		}
296
		dbs_info = &per_cpu(od_cpu_dbs_info, cpu);
297 298
		cpufreq_cpu_put(policy);

299
		mutex_lock(&dbs_info->cdbs.timer_mutex);
300

301 302
		if (!delayed_work_pending(&dbs_info->cdbs.work)) {
			mutex_unlock(&dbs_info->cdbs.timer_mutex);
303 304 305
			continue;
		}

306 307
		next_sampling = jiffies + usecs_to_jiffies(new_rate);
		appointed_at = dbs_info->cdbs.work.timer.expires;
308 309 310

		if (time_before(next_sampling, appointed_at)) {

311 312 313
			mutex_unlock(&dbs_info->cdbs.timer_mutex);
			cancel_delayed_work_sync(&dbs_info->cdbs.work);
			mutex_lock(&dbs_info->cdbs.timer_mutex);
314

315 316
			gov_queue_work(dbs_data, dbs_info->cdbs.cur_policy,
					usecs_to_jiffies(new_rate), true);
317 318

		}
319
		mutex_unlock(&dbs_info->cdbs.timer_mutex);
320 321 322
	}
}

323 324
static ssize_t store_sampling_rate(struct dbs_data *dbs_data, const char *buf,
		size_t count)
L
Linus Torvalds 已提交
325 326 327
{
	unsigned int input;
	int ret;
328
	ret = sscanf(buf, "%u", &input);
329 330
	if (ret != 1)
		return -EINVAL;
331 332

	update_sampling_rate(dbs_data, input);
L
Linus Torvalds 已提交
333 334 335
	return count;
}

336 337
static ssize_t store_io_is_busy(struct dbs_data *dbs_data, const char *buf,
		size_t count)
338
{
339
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
340 341 342 343 344 345
	unsigned int input;
	int ret;

	ret = sscanf(buf, "%u", &input);
	if (ret != 1)
		return -EINVAL;
346
	od_tuners->io_is_busy = !!input;
347 348 349
	return count;
}

350 351
static ssize_t store_up_threshold(struct dbs_data *dbs_data, const char *buf,
		size_t count)
L
Linus Torvalds 已提交
352
{
353
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
L
Linus Torvalds 已提交
354 355
	unsigned int input;
	int ret;
356
	ret = sscanf(buf, "%u", &input);
L
Linus Torvalds 已提交
357

358
	if (ret != 1 || input > MAX_FREQUENCY_UP_THRESHOLD ||
359
			input < MIN_FREQUENCY_UP_THRESHOLD) {
L
Linus Torvalds 已提交
360 361
		return -EINVAL;
	}
362
	/* Calculate the new adj_up_threshold */
363 364
	od_tuners->adj_up_threshold += input;
	od_tuners->adj_up_threshold -= od_tuners->up_threshold;
365

366
	od_tuners->up_threshold = input;
L
Linus Torvalds 已提交
367 368 369
	return count;
}

370 371
static ssize_t store_sampling_down_factor(struct dbs_data *dbs_data,
		const char *buf, size_t count)
372
{
373
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
374 375 376 377 378 379
	unsigned int input, j;
	int ret;
	ret = sscanf(buf, "%u", &input);

	if (ret != 1 || input > MAX_SAMPLING_DOWN_FACTOR || input < 1)
		return -EINVAL;
380
	od_tuners->sampling_down_factor = input;
381 382 383

	/* Reset down sampling multiplier in case it was active */
	for_each_online_cpu(j) {
384 385
		struct od_cpu_dbs_info_s *dbs_info = &per_cpu(od_cpu_dbs_info,
				j);
386 387 388 389 390
		dbs_info->rate_mult = 1;
	}
	return count;
}

391 392
static ssize_t store_ignore_nice(struct dbs_data *dbs_data, const char *buf,
		size_t count)
393
{
394
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
395 396 397 398
	unsigned int input;
	int ret;

	unsigned int j;
399

400
	ret = sscanf(buf, "%u", &input);
401
	if (ret != 1)
402 403
		return -EINVAL;

404
	if (input > 1)
405
		input = 1;
406

407
	if (input == od_tuners->ignore_nice) { /* nothing to do */
408 409
		return count;
	}
410
	od_tuners->ignore_nice = input;
411

412
	/* we need to re-evaluate prev_cpu_idle */
413
	for_each_online_cpu(j) {
414
		struct od_cpu_dbs_info_s *dbs_info;
415
		dbs_info = &per_cpu(od_cpu_dbs_info, j);
416 417
		dbs_info->cdbs.prev_cpu_idle = get_cpu_idle_time(j,
						&dbs_info->cdbs.prev_cpu_wall);
418
		if (od_tuners->ignore_nice)
419 420
			dbs_info->cdbs.prev_cpu_nice =
				kcpustat_cpu(j).cpustat[CPUTIME_NICE];
421

422 423 424 425
	}
	return count;
}

426 427
static ssize_t store_powersave_bias(struct dbs_data *dbs_data, const char *buf,
		size_t count)
428
{
429
	struct od_dbs_tuners *od_tuners = dbs_data->tuners;
430 431 432 433 434 435 436 437 438 439
	unsigned int input;
	int ret;
	ret = sscanf(buf, "%u", &input);

	if (ret != 1)
		return -EINVAL;

	if (input > 1000)
		input = 1000;

440
	od_tuners->powersave_bias = input;
441 442 443 444
	ondemand_powersave_bias_init();
	return count;
}

445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468
show_store_one(od, sampling_rate);
show_store_one(od, io_is_busy);
show_store_one(od, up_threshold);
show_store_one(od, sampling_down_factor);
show_store_one(od, ignore_nice);
show_store_one(od, powersave_bias);
declare_show_sampling_rate_min(od);

gov_sys_pol_attr_rw(sampling_rate);
gov_sys_pol_attr_rw(io_is_busy);
gov_sys_pol_attr_rw(up_threshold);
gov_sys_pol_attr_rw(sampling_down_factor);
gov_sys_pol_attr_rw(ignore_nice);
gov_sys_pol_attr_rw(powersave_bias);
gov_sys_pol_attr_ro(sampling_rate_min);

static struct attribute *dbs_attributes_gov_sys[] = {
	&sampling_rate_min_gov_sys.attr,
	&sampling_rate_gov_sys.attr,
	&up_threshold_gov_sys.attr,
	&sampling_down_factor_gov_sys.attr,
	&ignore_nice_gov_sys.attr,
	&powersave_bias_gov_sys.attr,
	&io_is_busy_gov_sys.attr,
L
Linus Torvalds 已提交
469 470 471
	NULL
};

472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489
static struct attribute_group od_attr_group_gov_sys = {
	.attrs = dbs_attributes_gov_sys,
	.name = "ondemand",
};

static struct attribute *dbs_attributes_gov_pol[] = {
	&sampling_rate_min_gov_pol.attr,
	&sampling_rate_gov_pol.attr,
	&up_threshold_gov_pol.attr,
	&sampling_down_factor_gov_pol.attr,
	&ignore_nice_gov_pol.attr,
	&powersave_bias_gov_pol.attr,
	&io_is_busy_gov_pol.attr,
	NULL
};

static struct attribute_group od_attr_group_gov_pol = {
	.attrs = dbs_attributes_gov_pol,
L
Linus Torvalds 已提交
490 491 492 493 494
	.name = "ondemand",
};

/************************** sysfs end ************************/

495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546
static int od_init(struct dbs_data *dbs_data)
{
	struct od_dbs_tuners *tuners;
	u64 idle_time;
	int cpu;

	tuners = kzalloc(sizeof(struct od_dbs_tuners), GFP_KERNEL);
	if (!tuners) {
		pr_err("%s: kzalloc failed\n", __func__);
		return -ENOMEM;
	}

	cpu = get_cpu();
	idle_time = get_cpu_idle_time_us(cpu, NULL);
	put_cpu();
	if (idle_time != -1ULL) {
		/* Idle micro accounting is supported. Use finer thresholds */
		tuners->up_threshold = MICRO_FREQUENCY_UP_THRESHOLD;
		tuners->adj_up_threshold = MICRO_FREQUENCY_UP_THRESHOLD -
			MICRO_FREQUENCY_DOWN_DIFFERENTIAL;
		/*
		 * In nohz/micro accounting case we set the minimum frequency
		 * not depending on HZ, but fixed (very low). The deferred
		 * timer might skip some samples if idle/sleeping as needed.
		*/
		dbs_data->min_sampling_rate = MICRO_FREQUENCY_MIN_SAMPLE_RATE;
	} else {
		tuners->up_threshold = DEF_FREQUENCY_UP_THRESHOLD;
		tuners->adj_up_threshold = DEF_FREQUENCY_UP_THRESHOLD -
			DEF_FREQUENCY_DOWN_DIFFERENTIAL;

		/* For correct statistics, we need 10 ticks for each measure */
		dbs_data->min_sampling_rate = MIN_SAMPLING_RATE_RATIO *
			jiffies_to_usecs(10);
	}

	tuners->sampling_down_factor = DEF_SAMPLING_DOWN_FACTOR;
	tuners->ignore_nice = 0;
	tuners->powersave_bias = 0;
	tuners->io_is_busy = should_io_be_busy();

	dbs_data->tuners = tuners;
	pr_info("%s: tuners %p\n", __func__, tuners);
	mutex_init(&dbs_data->mutex);
	return 0;
}

static void od_exit(struct dbs_data *dbs_data)
{
	kfree(dbs_data->tuners);
}

547
define_get_cpu_dbs_routines(od_cpu_dbs_info);
548

549 550 551 552 553
static struct od_ops od_ops = {
	.powersave_bias_init_cpu = ondemand_powersave_bias_init_cpu,
	.powersave_bias_target = powersave_bias_target,
	.freq_increase = dbs_freq_increase,
};
554

555
static struct common_dbs_data od_dbs_cdata = {
556
	.governor = GOV_ONDEMAND,
557 558
	.attr_group_gov_sys = &od_attr_group_gov_sys,
	.attr_group_gov_pol = &od_attr_group_gov_pol,
559 560 561 562 563
	.get_cpu_cdbs = get_cpu_cdbs,
	.get_cpu_dbs_info_s = get_cpu_dbs_info_s,
	.gov_dbs_timer = od_dbs_timer,
	.gov_check_cpu = od_check_cpu,
	.gov_ops = &od_ops,
564 565
	.init = od_init,
	.exit = od_exit,
566
};
L
Linus Torvalds 已提交
567

568 569
static int od_cpufreq_governor_dbs(struct cpufreq_policy *policy,
		unsigned int event)
L
Linus Torvalds 已提交
570
{
571
	return cpufreq_governor_dbs(policy, &od_dbs_cdata, event);
L
Linus Torvalds 已提交
572 573
}

574 575
#ifndef CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND
static
576
#endif
577 578 579 580 581 582
struct cpufreq_governor cpufreq_gov_ondemand = {
	.name			= "ondemand",
	.governor		= od_cpufreq_governor_dbs,
	.max_transition_latency	= TRANSITION_LATENCY_LIMIT,
	.owner			= THIS_MODULE,
};
L
Linus Torvalds 已提交
583 584 585

static int __init cpufreq_gov_dbs_init(void)
{
586
	return cpufreq_register_governor(&cpufreq_gov_ondemand);
L
Linus Torvalds 已提交
587 588 589 590
}

static void __exit cpufreq_gov_dbs_exit(void)
{
591
	cpufreq_unregister_governor(&cpufreq_gov_ondemand);
L
Linus Torvalds 已提交
592 593
}

594 595 596
MODULE_AUTHOR("Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>");
MODULE_AUTHOR("Alexey Starikovskiy <alexey.y.starikovskiy@intel.com>");
MODULE_DESCRIPTION("'cpufreq_ondemand' - A dynamic cpufreq governor for "
597
	"Low Latency Frequency Transition capable processors");
598
MODULE_LICENSE("GPL");
L
Linus Torvalds 已提交
599

600 601 602
#ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND
fs_initcall(cpufreq_gov_dbs_init);
#else
L
Linus Torvalds 已提交
603
module_init(cpufreq_gov_dbs_init);
604
#endif
L
Linus Torvalds 已提交
605
module_exit(cpufreq_gov_dbs_exit);