cpufreq_schedutil.c 19.0 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11
/*
 * CPUFreq governor based on scheduler-provided CPU utilization data.
 *
 * Copyright (C) 2016, Intel Corporation
 * Author: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */

12 13
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

14 15
#include "sched.h"

16 17
#include <trace/events/power.h>

18
struct sugov_tunables {
19 20
	struct gov_attr_set	attr_set;
	unsigned int		rate_limit_us;
21 22 23
};

struct sugov_policy {
24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43
	struct cpufreq_policy	*policy;

	struct sugov_tunables	*tunables;
	struct list_head	tunables_hook;

	raw_spinlock_t		update_lock;	/* For shared policies */
	u64			last_freq_update_time;
	s64			freq_update_delay_ns;
	unsigned int		next_freq;
	unsigned int		cached_raw_freq;

	/* The next fields are only needed if fast switch cannot be used: */
	struct			irq_work irq_work;
	struct			kthread_work work;
	struct			mutex work_lock;
	struct			kthread_worker worker;
	struct task_struct	*thread;
	bool			work_in_progress;

	bool			need_freq_update;
44 45 46
};

struct sugov_cpu {
47 48 49
	struct update_util_data	update_util;
	struct sugov_policy	*sg_policy;
	unsigned int		cpu;
50

51 52 53
	bool			iowait_boost_pending;
	unsigned int		iowait_boost;
	unsigned int		iowait_boost_max;
54
	u64 last_update;
55

56 57 58 59
	/* The fields below are only needed when sharing a policy: */
	unsigned long		util_cfs;
	unsigned long		util_dl;
	unsigned long		max;
60

61
	/* The field below is for single-CPU policies only: */
62
#ifdef CONFIG_NO_HZ_COMMON
63
	unsigned long		saved_idle_calls;
64
#endif
65 66 67 68 69 70 71 72 73 74
};

static DEFINE_PER_CPU(struct sugov_cpu, sugov_cpu);

/************************ Governor internals ***********************/

static bool sugov_should_update_freq(struct sugov_policy *sg_policy, u64 time)
{
	s64 delta_ns;

75 76
	/*
	 * Since cpufreq_update_util() is called with rq->lock held for
77
	 * the @target_cpu, our per-CPU data is fully serialized.
78
	 *
79
	 * However, drivers cannot in general deal with cross-CPU
80
	 * requests, so while get_next_freq() will work, our
81
	 * sugov_update_commit() call may not for the fast switching platforms.
82 83 84 85
	 *
	 * Hence stop here for remote requests if they aren't supported
	 * by the hardware, as calculating the frequency is pointless if
	 * we cannot in fact act on it.
86 87 88 89
	 *
	 * For the slow switching platforms, the kthread is always scheduled on
	 * the right set of CPUs and any CPU can find the next frequency and
	 * schedule the kthread.
90
	 */
91 92
	if (sg_policy->policy->fast_switch_enabled &&
	    !cpufreq_can_do_remote_dvfs(sg_policy->policy))
93 94
		return false;

95 96 97 98 99 100 101 102 103 104 105 106 107 108
	if (sg_policy->work_in_progress)
		return false;

	if (unlikely(sg_policy->need_freq_update)) {
		sg_policy->need_freq_update = false;
		/*
		 * This happens when limits change, so forget the previous
		 * next_freq value and force an update.
		 */
		sg_policy->next_freq = UINT_MAX;
		return true;
	}

	delta_ns = time - sg_policy->last_freq_update_time;
109

110 111 112 113 114 115 116 117
	return delta_ns >= sg_policy->freq_update_delay_ns;
}

static void sugov_update_commit(struct sugov_policy *sg_policy, u64 time,
				unsigned int next_freq)
{
	struct cpufreq_policy *policy = sg_policy->policy;

118 119 120 121
	if (sg_policy->next_freq == next_freq)
		return;

	sg_policy->next_freq = next_freq;
122 123 124 125
	sg_policy->last_freq_update_time = time;

	if (policy->fast_switch_enabled) {
		next_freq = cpufreq_driver_fast_switch(policy, next_freq);
126
		if (!next_freq)
127 128 129 130
			return;

		policy->cur = next_freq;
		trace_cpu_frequency(next_freq, smp_processor_id());
131
	} else {
132 133 134 135 136 137 138
		sg_policy->work_in_progress = true;
		irq_work_queue(&sg_policy->irq_work);
	}
}

/**
 * get_next_freq - Compute a new frequency for a given cpufreq policy.
139
 * @sg_policy: schedutil policy object to compute the new frequency for.
140 141 142 143 144 145 146 147 148 149 150 151 152 153
 * @util: Current CPU utilization.
 * @max: CPU capacity.
 *
 * If the utilization is frequency-invariant, choose the new frequency to be
 * proportional to it, that is
 *
 * next_freq = C * max_freq * util / max
 *
 * Otherwise, approximate the would-be frequency-invariant utilization by
 * util_raw * (curr_freq / max_freq) which leads to
 *
 * next_freq = C * curr_freq * util_raw / max
 *
 * Take C = 1.25 for the frequency tipping point at (util / max) = 0.8.
154 155 156 157
 *
 * The lowest driver-supported frequency which is equal or greater than the raw
 * next_freq (as calculated above) is returned, subject to policy min/max and
 * cpufreq driver limitations.
158
 */
159 160
static unsigned int get_next_freq(struct sugov_policy *sg_policy,
				  unsigned long util, unsigned long max)
161
{
162
	struct cpufreq_policy *policy = sg_policy->policy;
163 164 165
	unsigned int freq = arch_scale_freq_invariant() ?
				policy->cpuinfo.max_freq : policy->cur;

166 167
	freq = (freq + (freq >> 2)) * util / max;

168
	if (freq == sg_policy->cached_raw_freq && sg_policy->next_freq != UINT_MAX)
169
		return sg_policy->next_freq;
170
	sg_policy->cached_raw_freq = freq;
171
	return cpufreq_driver_resolve_freq(policy, freq);
172 173
}

174
static void sugov_get_util(struct sugov_cpu *sg_cpu)
175
{
176
	struct rq *rq = cpu_rq(sg_cpu->cpu);
177

178 179 180 181
	sg_cpu->max = arch_scale_cpu_capacity(NULL, sg_cpu->cpu);
	sg_cpu->util_cfs = cpu_util_cfs(rq);
	sg_cpu->util_dl  = cpu_util_dl(rq);
}
182

183 184
static unsigned long sugov_aggregate_util(struct sugov_cpu *sg_cpu)
{
185 186 187 188 189 190 191 192 193 194 195
	struct rq *rq = cpu_rq(sg_cpu->cpu);
	unsigned long util;

	if (rq->rt.rt_nr_running) {
		util = sg_cpu->max;
	} else {
		util = sg_cpu->util_dl;
		if (rq->cfs.h_nr_running)
			util += sg_cpu->util_cfs;
	}

196 197 198 199 200
	/*
	 * Ideally we would like to set util_dl as min/guaranteed freq and
	 * util_cfs + util_dl as requested freq. However, cpufreq is not yet
	 * ready for such an interface. So, we only do the latter for now.
	 */
201
	return min(util, sg_cpu->max);
202 203
}

204
static void sugov_set_iowait_boost(struct sugov_cpu *sg_cpu, u64 time, unsigned int flags)
205
{
206
	if (flags & SCHED_CPUFREQ_IOWAIT) {
207 208 209 210 211 212 213 214 215 216 217 218
		if (sg_cpu->iowait_boost_pending)
			return;

		sg_cpu->iowait_boost_pending = true;

		if (sg_cpu->iowait_boost) {
			sg_cpu->iowait_boost <<= 1;
			if (sg_cpu->iowait_boost > sg_cpu->iowait_boost_max)
				sg_cpu->iowait_boost = sg_cpu->iowait_boost_max;
		} else {
			sg_cpu->iowait_boost = sg_cpu->sg_policy->policy->min;
		}
219 220 221 222
	} else if (sg_cpu->iowait_boost) {
		s64 delta_ns = time - sg_cpu->last_update;

		/* Clear iowait_boost if the CPU apprears to have been idle. */
223
		if (delta_ns > TICK_NSEC) {
224
			sg_cpu->iowait_boost = 0;
225 226
			sg_cpu->iowait_boost_pending = false;
		}
227 228 229 230 231 232
	}
}

static void sugov_iowait_boost(struct sugov_cpu *sg_cpu, unsigned long *util,
			       unsigned long *max)
{
233
	unsigned int boost_util, boost_max;
234

235
	if (!sg_cpu->iowait_boost)
236 237
		return;

238 239 240 241 242 243 244 245 246 247 248 249 250
	if (sg_cpu->iowait_boost_pending) {
		sg_cpu->iowait_boost_pending = false;
	} else {
		sg_cpu->iowait_boost >>= 1;
		if (sg_cpu->iowait_boost < sg_cpu->sg_policy->policy->min) {
			sg_cpu->iowait_boost = 0;
			return;
		}
	}

	boost_util = sg_cpu->iowait_boost;
	boost_max = sg_cpu->iowait_boost_max;

251 252 253 254 255 256
	if (*util * boost_max < *max * boost_util) {
		*util = boost_util;
		*max = boost_max;
	}
}

257 258 259
#ifdef CONFIG_NO_HZ_COMMON
static bool sugov_cpu_is_busy(struct sugov_cpu *sg_cpu)
{
260
	unsigned long idle_calls = tick_nohz_get_idle_calls_cpu(sg_cpu->cpu);
261 262 263 264 265 266 267 268 269
	bool ret = idle_calls == sg_cpu->saved_idle_calls;

	sg_cpu->saved_idle_calls = idle_calls;
	return ret;
}
#else
static inline bool sugov_cpu_is_busy(struct sugov_cpu *sg_cpu) { return false; }
#endif /* CONFIG_NO_HZ_COMMON */

270
static void sugov_update_single(struct update_util_data *hook, u64 time,
271
				unsigned int flags)
272 273 274
{
	struct sugov_cpu *sg_cpu = container_of(hook, struct sugov_cpu, update_util);
	struct sugov_policy *sg_policy = sg_cpu->sg_policy;
275
	unsigned long util, max;
276
	unsigned int next_f;
277
	bool busy;
278

279
	sugov_set_iowait_boost(sg_cpu, time, flags);
280 281
	sg_cpu->last_update = time;

282 283 284
	if (!sugov_should_update_freq(sg_policy, time))
		return;

285 286
	busy = sugov_cpu_is_busy(sg_cpu);

287 288 289 290 291 292 293 294 295 296 297
	sugov_get_util(sg_cpu);
	max = sg_cpu->max;
	util = sugov_aggregate_util(sg_cpu);
	sugov_iowait_boost(sg_cpu, &util, &max);
	next_f = get_next_freq(sg_policy, util, max);
	/*
	 * Do not reduce the frequency if the CPU has not been idle
	 * recently, as the reduction is likely to be premature then.
	 */
	if (busy && next_f < sg_policy->next_freq) {
		next_f = sg_policy->next_freq;
298

299 300
		/* Reset cached freq as next_freq has changed */
		sg_policy->cached_raw_freq = 0;
301
	}
302

303 304 305
	sugov_update_commit(sg_policy, time, next_f);
}

306
static unsigned int sugov_next_freq_shared(struct sugov_cpu *sg_cpu, u64 time)
307
{
308
	struct sugov_policy *sg_policy = sg_cpu->sg_policy;
309
	struct cpufreq_policy *policy = sg_policy->policy;
310
	unsigned long util = 0, max = 1;
311 312 313
	unsigned int j;

	for_each_cpu(j, policy->cpus) {
314
		struct sugov_cpu *j_sg_cpu = &per_cpu(sugov_cpu, j);
315 316 317
		unsigned long j_util, j_max;
		s64 delta_ns;

318 319
		sugov_get_util(j_sg_cpu);

320
		/*
321 322 323 324 325 326 327
		 * If the CFS CPU utilization was last updated before the
		 * previous frequency update and the time elapsed between the
		 * last update of the CPU utilization and the last frequency
		 * update is long enough, reset iowait_boost and util_cfs, as
		 * they are now probably stale. However, still consider the
		 * CPU contribution if it has some DEADLINE utilization
		 * (util_dl).
328
		 */
329
		delta_ns = time - j_sg_cpu->last_update;
330 331
		if (delta_ns > TICK_NSEC) {
			j_sg_cpu->iowait_boost = 0;
332
			j_sg_cpu->iowait_boost_pending = false;
333
		}
334 335

		j_max = j_sg_cpu->max;
336
		j_util = sugov_aggregate_util(j_sg_cpu);
337
		sugov_iowait_boost(j_sg_cpu, &j_util, &j_max);
338 339 340 341 342 343
		if (j_util * max > j_max * util) {
			util = j_util;
			max = j_max;
		}
	}

344
	return get_next_freq(sg_policy, util, max);
345 346
}

347 348
static void
sugov_update_shared(struct update_util_data *hook, u64 time, unsigned int flags)
349 350 351 352 353 354 355
{
	struct sugov_cpu *sg_cpu = container_of(hook, struct sugov_cpu, update_util);
	struct sugov_policy *sg_policy = sg_cpu->sg_policy;
	unsigned int next_f;

	raw_spin_lock(&sg_policy->update_lock);

356
	sugov_set_iowait_boost(sg_cpu, time, flags);
357 358 359
	sg_cpu->last_update = time;

	if (sugov_should_update_freq(sg_policy, time)) {
360
		next_f = sugov_next_freq_shared(sg_cpu, time);
361 362 363 364 365 366
		sugov_update_commit(sg_policy, time, next_f);
	}

	raw_spin_unlock(&sg_policy->update_lock);
}

367
static void sugov_work(struct kthread_work *work)
368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383
{
	struct sugov_policy *sg_policy = container_of(work, struct sugov_policy, work);

	mutex_lock(&sg_policy->work_lock);
	__cpufreq_driver_target(sg_policy->policy, sg_policy->next_freq,
				CPUFREQ_RELATION_L);
	mutex_unlock(&sg_policy->work_lock);

	sg_policy->work_in_progress = false;
}

static void sugov_irq_work(struct irq_work *irq_work)
{
	struct sugov_policy *sg_policy;

	sg_policy = container_of(irq_work, struct sugov_policy, irq_work);
384 385

	/*
386 387 388
	 * For RT tasks, the schedutil governor shoots the frequency to maximum.
	 * Special care must be taken to ensure that this kthread doesn't result
	 * in the same behavior.
389 390
	 *
	 * This is (mostly) guaranteed by the work_in_progress flag. The flag is
391 392
	 * updated only at the end of the sugov_work() function and before that
	 * the schedutil governor rejects all other frequency scaling requests.
393
	 *
394
	 * There is a very rare case though, where the RT thread yields right
395 396 397 398
	 * after the work_in_progress flag is cleared. The effects of that are
	 * neglected for now.
	 */
	kthread_queue_work(&sg_policy->worker, &sg_policy->work);
399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417
}

/************************** sysfs interface ************************/

static struct sugov_tunables *global_tunables;
static DEFINE_MUTEX(global_tunables_lock);

static inline struct sugov_tunables *to_sugov_tunables(struct gov_attr_set *attr_set)
{
	return container_of(attr_set, struct sugov_tunables, attr_set);
}

static ssize_t rate_limit_us_show(struct gov_attr_set *attr_set, char *buf)
{
	struct sugov_tunables *tunables = to_sugov_tunables(attr_set);

	return sprintf(buf, "%u\n", tunables->rate_limit_us);
}

418 419
static ssize_t
rate_limit_us_store(struct gov_attr_set *attr_set, const char *buf, size_t count)
420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469
{
	struct sugov_tunables *tunables = to_sugov_tunables(attr_set);
	struct sugov_policy *sg_policy;
	unsigned int rate_limit_us;

	if (kstrtouint(buf, 10, &rate_limit_us))
		return -EINVAL;

	tunables->rate_limit_us = rate_limit_us;

	list_for_each_entry(sg_policy, &attr_set->policy_list, tunables_hook)
		sg_policy->freq_update_delay_ns = rate_limit_us * NSEC_PER_USEC;

	return count;
}

static struct governor_attr rate_limit_us = __ATTR_RW(rate_limit_us);

static struct attribute *sugov_attributes[] = {
	&rate_limit_us.attr,
	NULL
};

static struct kobj_type sugov_tunables_ktype = {
	.default_attrs = sugov_attributes,
	.sysfs_ops = &governor_sysfs_ops,
};

/********************** cpufreq governor interface *********************/

static struct cpufreq_governor schedutil_gov;

static struct sugov_policy *sugov_policy_alloc(struct cpufreq_policy *policy)
{
	struct sugov_policy *sg_policy;

	sg_policy = kzalloc(sizeof(*sg_policy), GFP_KERNEL);
	if (!sg_policy)
		return NULL;

	sg_policy->policy = policy;
	raw_spin_lock_init(&sg_policy->update_lock);
	return sg_policy;
}

static void sugov_policy_free(struct sugov_policy *sg_policy)
{
	kfree(sg_policy);
}

470 471 472
static int sugov_kthread_create(struct sugov_policy *sg_policy)
{
	struct task_struct *thread;
473
	struct sched_attr attr = {
474 475 476 477 478
		.size		= sizeof(struct sched_attr),
		.sched_policy	= SCHED_DEADLINE,
		.sched_flags	= SCHED_FLAG_SUGOV,
		.sched_nice	= 0,
		.sched_priority	= 0,
479 480 481 482 483 484 485 486
		/*
		 * Fake (unused) bandwidth; workaround to "fix"
		 * priority inheritance.
		 */
		.sched_runtime	=  1000000,
		.sched_deadline = 10000000,
		.sched_period	= 10000000,
	};
487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503
	struct cpufreq_policy *policy = sg_policy->policy;
	int ret;

	/* kthread only required for slow path */
	if (policy->fast_switch_enabled)
		return 0;

	kthread_init_work(&sg_policy->work, sugov_work);
	kthread_init_worker(&sg_policy->worker);
	thread = kthread_create(kthread_worker_fn, &sg_policy->worker,
				"sugov:%d",
				cpumask_first(policy->related_cpus));
	if (IS_ERR(thread)) {
		pr_err("failed to create sugov thread: %ld\n", PTR_ERR(thread));
		return PTR_ERR(thread);
	}

504
	ret = sched_setattr_nocheck(thread, &attr);
505 506
	if (ret) {
		kthread_stop(thread);
507
		pr_warn("%s: failed to set SCHED_DEADLINE\n", __func__);
508 509 510 511
		return ret;
	}

	sg_policy->thread = thread;
512 513 514 515 516

	/* Kthread is bound to all CPUs by default */
	if (!policy->dvfs_possible_from_any_cpu)
		kthread_bind_mask(thread, policy->related_cpus);

517 518 519
	init_irq_work(&sg_policy->irq_work, sugov_irq_work);
	mutex_init(&sg_policy->work_lock);

520 521 522 523 524 525 526 527 528 529 530 531 532
	wake_up_process(thread);

	return 0;
}

static void sugov_kthread_stop(struct sugov_policy *sg_policy)
{
	/* kthread only required for slow path */
	if (sg_policy->policy->fast_switch_enabled)
		return;

	kthread_flush_worker(&sg_policy->worker);
	kthread_stop(sg_policy->thread);
533
	mutex_destroy(&sg_policy->work_lock);
534 535
}

536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566
static struct sugov_tunables *sugov_tunables_alloc(struct sugov_policy *sg_policy)
{
	struct sugov_tunables *tunables;

	tunables = kzalloc(sizeof(*tunables), GFP_KERNEL);
	if (tunables) {
		gov_attr_set_init(&tunables->attr_set, &sg_policy->tunables_hook);
		if (!have_governor_per_policy())
			global_tunables = tunables;
	}
	return tunables;
}

static void sugov_tunables_free(struct sugov_tunables *tunables)
{
	if (!have_governor_per_policy())
		global_tunables = NULL;

	kfree(tunables);
}

static int sugov_init(struct cpufreq_policy *policy)
{
	struct sugov_policy *sg_policy;
	struct sugov_tunables *tunables;
	int ret = 0;

	/* State should be equivalent to EXIT */
	if (policy->governor_data)
		return -EBUSY;

567 568
	cpufreq_enable_fast_switch(policy);

569
	sg_policy = sugov_policy_alloc(policy);
570 571 572 573
	if (!sg_policy) {
		ret = -ENOMEM;
		goto disable_fast_switch;
	}
574

575 576 577 578
	ret = sugov_kthread_create(sg_policy);
	if (ret)
		goto free_sg_policy;

579 580 581 582 583
	mutex_lock(&global_tunables_lock);

	if (global_tunables) {
		if (WARN_ON(have_governor_per_policy())) {
			ret = -EINVAL;
584
			goto stop_kthread;
585 586 587 588 589 590 591 592 593 594 595
		}
		policy->governor_data = sg_policy;
		sg_policy->tunables = global_tunables;

		gov_attr_set_get(&global_tunables->attr_set, &sg_policy->tunables_hook);
		goto out;
	}

	tunables = sugov_tunables_alloc(sg_policy);
	if (!tunables) {
		ret = -ENOMEM;
596
		goto stop_kthread;
597 598
	}

599
	tunables->rate_limit_us = cpufreq_policy_transition_delay_us(policy);
600 601 602 603 604 605 606 607 608 609

	policy->governor_data = sg_policy;
	sg_policy->tunables = tunables;

	ret = kobject_init_and_add(&tunables->attr_set.kobj, &sugov_tunables_ktype,
				   get_governor_parent_kobj(policy), "%s",
				   schedutil_gov.name);
	if (ret)
		goto fail;

610
out:
611 612 613
	mutex_unlock(&global_tunables_lock);
	return 0;

614
fail:
615 616 617
	policy->governor_data = NULL;
	sugov_tunables_free(tunables);

618 619 620
stop_kthread:
	sugov_kthread_stop(sg_policy);

621
free_sg_policy:
622 623 624
	mutex_unlock(&global_tunables_lock);

	sugov_policy_free(sg_policy);
625 626 627 628

disable_fast_switch:
	cpufreq_disable_fast_switch(policy);

629
	pr_err("initialization failed (error %d)\n", ret);
630 631 632
	return ret;
}

633
static void sugov_exit(struct cpufreq_policy *policy)
634 635 636 637 638 639 640 641 642 643 644 645 646 647
{
	struct sugov_policy *sg_policy = policy->governor_data;
	struct sugov_tunables *tunables = sg_policy->tunables;
	unsigned int count;

	mutex_lock(&global_tunables_lock);

	count = gov_attr_set_put(&tunables->attr_set, &sg_policy->tunables_hook);
	policy->governor_data = NULL;
	if (!count)
		sugov_tunables_free(tunables);

	mutex_unlock(&global_tunables_lock);

648
	sugov_kthread_stop(sg_policy);
649
	sugov_policy_free(sg_policy);
650
	cpufreq_disable_fast_switch(policy);
651 652 653 654 655 656 657
}

static int sugov_start(struct cpufreq_policy *policy)
{
	struct sugov_policy *sg_policy = policy->governor_data;
	unsigned int cpu;

658 659 660 661 662 663
	sg_policy->freq_update_delay_ns	= sg_policy->tunables->rate_limit_us * NSEC_PER_USEC;
	sg_policy->last_freq_update_time	= 0;
	sg_policy->next_freq			= UINT_MAX;
	sg_policy->work_in_progress		= false;
	sg_policy->need_freq_update		= false;
	sg_policy->cached_raw_freq		= 0;
664 665 666 667

	for_each_cpu(cpu, policy->cpus) {
		struct sugov_cpu *sg_cpu = &per_cpu(sugov_cpu, cpu);

668
		memset(sg_cpu, 0, sizeof(*sg_cpu));
669 670 671
		sg_cpu->cpu			= cpu;
		sg_cpu->sg_policy		= sg_policy;
		sg_cpu->iowait_boost_max	= policy->cpuinfo.max_freq;
672 673 674 675 676
	}

	for_each_cpu(cpu, policy->cpus) {
		struct sugov_cpu *sg_cpu = &per_cpu(sugov_cpu, cpu);

677 678 679 680
		cpufreq_add_update_util_hook(cpu, &sg_cpu->update_util,
					     policy_is_shared(policy) ?
							sugov_update_shared :
							sugov_update_single);
681 682 683 684
	}
	return 0;
}

685
static void sugov_stop(struct cpufreq_policy *policy)
686 687 688 689 690 691 692 693 694
{
	struct sugov_policy *sg_policy = policy->governor_data;
	unsigned int cpu;

	for_each_cpu(cpu, policy->cpus)
		cpufreq_remove_update_util_hook(cpu);

	synchronize_sched();

695 696 697 698
	if (!policy->fast_switch_enabled) {
		irq_work_sync(&sg_policy->irq_work);
		kthread_cancel_work_sync(&sg_policy->work);
	}
699 700
}

701
static void sugov_limits(struct cpufreq_policy *policy)
702 703 704 705 706
{
	struct sugov_policy *sg_policy = policy->governor_data;

	if (!policy->fast_switch_enabled) {
		mutex_lock(&sg_policy->work_lock);
707
		cpufreq_policy_apply_limits(policy);
708 709 710 711 712 713 714
		mutex_unlock(&sg_policy->work_lock);
	}

	sg_policy->need_freq_update = true;
}

static struct cpufreq_governor schedutil_gov = {
715 716 717 718 719 720 721 722
	.name			= "schedutil",
	.owner			= THIS_MODULE,
	.dynamic_switching	= true,
	.init			= sugov_init,
	.exit			= sugov_exit,
	.start			= sugov_start,
	.stop			= sugov_stop,
	.limits			= sugov_limits,
723 724 725 726 727 728 729 730
};

#ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_SCHEDUTIL
struct cpufreq_governor *cpufreq_default_governor(void)
{
	return &schedutil_gov;
}
#endif
731 732 733 734 735 736

static int __init sugov_register(void)
{
	return cpufreq_register_governor(&schedutil_gov);
}
fs_initcall(sugov_register);