cpuidle44xx.c 7.6 KB
Newer Older
1
/*
2
 * OMAP4+ CPU idle Routines
3
 *
4
 * Copyright (C) 2011-2013 Texas Instruments, Inc.
5 6 7 8 9 10 11 12 13 14 15 16
 * Santosh Shilimkar <santosh.shilimkar@ti.com>
 * Rajendra Nayak <rnayak@ti.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */

#include <linux/sched.h>
#include <linux/cpuidle.h>
#include <linux/cpu_pm.h>
#include <linux/export.h>
17
#include <linux/tick.h>
18

19
#include <asm/cpuidle.h>
20 21 22 23

#include "common.h"
#include "pm.h"
#include "prm.h"
24
#include "soc.h"
25
#include "clockdomain.h"
26

27 28
#define MAX_CPUS	2

29
/* Machine specific information */
30
struct idle_statedata {
31 32 33
	u32 cpu_state;
	u32 mpu_logic_state;
	u32 mpu_state;
34
	u32 mpu_state_vote;
35 36
};

37
static struct idle_statedata omap4_idle_data[] = {
38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53
	{
		.cpu_state = PWRDM_POWER_ON,
		.mpu_state = PWRDM_POWER_ON,
		.mpu_logic_state = PWRDM_POWER_RET,
	},
	{
		.cpu_state = PWRDM_POWER_OFF,
		.mpu_state = PWRDM_POWER_RET,
		.mpu_logic_state = PWRDM_POWER_RET,
	},
	{
		.cpu_state = PWRDM_POWER_OFF,
		.mpu_state = PWRDM_POWER_RET,
		.mpu_logic_state = PWRDM_POWER_OFF,
	},
};
54

55 56 57 58 59 60 61 62 63 64 65 66 67
static struct idle_statedata omap5_idle_data[] = {
	{
		.cpu_state = PWRDM_POWER_ON,
		.mpu_state = PWRDM_POWER_ON,
		.mpu_logic_state = PWRDM_POWER_ON,
	},
	{
		.cpu_state = PWRDM_POWER_RET,
		.mpu_state = PWRDM_POWER_RET,
		.mpu_logic_state = PWRDM_POWER_RET,
	},
};

68 69
static struct powerdomain *mpu_pd, *cpu_pd[MAX_CPUS];
static struct clockdomain *cpu_clkdm[MAX_CPUS];
70

71
static atomic_t abort_barrier;
72
static bool cpu_done[MAX_CPUS];
73
static struct idle_statedata *state_ptr = &omap4_idle_data[0];
74
static DEFINE_RAW_SPINLOCK(mpu_lock);
75

76 77
/* Private functions */

78
/**
79
 * omap_enter_idle_[simple/coupled] - OMAP4PLUS cpuidle entry functions
80 81 82 83 84 85 86 87
 * @dev: cpuidle device
 * @drv: cpuidle driver
 * @index: the index of state to be entered
 *
 * Called from the CPUidle framework to program the device to the
 * specified low power state selected by the governor.
 * Returns the amount of time spent in the low power state.
 */
88
static int omap_enter_idle_simple(struct cpuidle_device *dev,
89 90 91 92 93 94 95
			struct cpuidle_driver *drv,
			int index)
{
	omap_do_wfi();
	return index;
}

96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121
static int omap_enter_idle_smp(struct cpuidle_device *dev,
			       struct cpuidle_driver *drv,
			       int index)
{
	struct idle_statedata *cx = state_ptr + index;
	unsigned long flag;

	raw_spin_lock_irqsave(&mpu_lock, flag);
	cx->mpu_state_vote++;
	if (cx->mpu_state_vote == num_online_cpus()) {
		pwrdm_set_logic_retst(mpu_pd, cx->mpu_logic_state);
		omap_set_pwrdm_state(mpu_pd, cx->mpu_state);
	}
	raw_spin_unlock_irqrestore(&mpu_lock, flag);

	omap4_enter_lowpower(dev->cpu, cx->cpu_state);

	raw_spin_lock_irqsave(&mpu_lock, flag);
	if (cx->mpu_state_vote == num_online_cpus())
		omap_set_pwrdm_state(mpu_pd, PWRDM_POWER_ON);
	cx->mpu_state_vote--;
	raw_spin_unlock_irqrestore(&mpu_lock, flag);

	return index;
}

122
static int omap_enter_idle_coupled(struct cpuidle_device *dev,
123 124 125
			struct cpuidle_driver *drv,
			int index)
{
126
	struct idle_statedata *cx = state_ptr + index;
127
	u32 mpuss_can_lose_context = 0;
128 129

	/*
130
	 * CPU0 has to wait and stay ON until CPU1 is OFF state.
131 132 133 134
	 * This is necessary to honour hardware recommondation
	 * of triggeing all the possible low power modes once CPU1 is
	 * out of coherency and in OFF mode.
	 */
135
	if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
136
		while (pwrdm_read_pwrst(cpu_pd[1]) != PWRDM_POWER_OFF) {
137
			cpu_relax();
138 139 140 141 142 143 144 145 146 147 148 149

			/*
			 * CPU1 could have already entered & exited idle
			 * without hitting off because of a wakeup
			 * or a failed attempt to hit off mode.  Check for
			 * that here, otherwise we could spin forever
			 * waiting for CPU1 off.
			 */
			if (cpu_done[1])
			    goto fail;

		}
150 151
	}

152 153 154
	mpuss_can_lose_context = (cx->mpu_state == PWRDM_POWER_RET) &&
				 (cx->mpu_logic_state == PWRDM_POWER_OFF);

155 156 157 158
	/* Enter broadcast mode for periodic timers */
	tick_broadcast_enable();

	/* Enter broadcast mode for one-shot timers */
159
	tick_broadcast_enter();
160

161 162 163 164
	/*
	 * Call idle CPU PM enter notifier chain so that
	 * VFP and per CPU interrupt context is saved.
	 */
165 166 167 168 169 170 171 172 173 174
	cpu_pm_enter();

	if (dev->cpu == 0) {
		pwrdm_set_logic_retst(mpu_pd, cx->mpu_logic_state);
		omap_set_pwrdm_state(mpu_pd, cx->mpu_state);

		/*
		 * Call idle CPU cluster PM enter notifier chain
		 * to save GIC and wakeupgen context.
		 */
175 176
		if (mpuss_can_lose_context)
			cpu_cluster_pm_enter();
177
	}
178 179

	omap4_enter_lowpower(dev->cpu, cx->cpu_state);
180
	cpu_done[dev->cpu] = true;
181

182 183
	/* Wakeup CPU1 only if it is not offlined */
	if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
184 185 186 187 188

		if (IS_PM44XX_ERRATUM(PM_OMAP4_ROM_SMP_BOOT_ERRATUM_GICD) &&
		    mpuss_can_lose_context)
			gic_dist_disable();

189
		clkdm_deny_idle(cpu_clkdm[1]);
190
		omap_set_pwrdm_state(cpu_pd[1], PWRDM_POWER_ON);
191
		clkdm_allow_idle(cpu_clkdm[1]);
192 193 194 195 196 197 198 199 200

		if (IS_PM44XX_ERRATUM(PM_OMAP4_ROM_SMP_BOOT_ERRATUM_GICD) &&
		    mpuss_can_lose_context) {
			while (gic_dist_disabled()) {
				udelay(1);
				cpu_relax();
			}
			gic_timer_retrigger();
		}
201
	}
202 203 204

	/*
	 * Call idle CPU PM exit notifier chain to restore
205
	 * VFP and per CPU IRQ context.
206
	 */
207
	cpu_pm_exit();
208 209 210 211 212

	/*
	 * Call idle CPU cluster PM exit notifier chain
	 * to restore GIC and wakeupgen context.
	 */
213
	if (dev->cpu == 0 && mpuss_can_lose_context)
214 215
		cpu_cluster_pm_exit();

216
	tick_broadcast_exit();
217

218 219 220
fail:
	cpuidle_coupled_parallel_barrier(dev, &abort_barrier);
	cpu_done[dev->cpu] = false;
221

222 223 224
	return index;
}

225
static struct cpuidle_driver omap4_idle_driver = {
226 227
	.name				= "omap4_idle",
	.owner				= THIS_MODULE,
228 229 230 231 232
	.states = {
		{
			/* C1 - CPU0 ON + CPU1 ON + MPU ON */
			.exit_latency = 2 + 2,
			.target_residency = 5,
233
			.enter = omap_enter_idle_simple,
234
			.name = "C1",
235
			.desc = "CPUx ON, MPUSS ON"
236 237
		},
		{
238
			/* C2 - CPU0 OFF + CPU1 OFF + MPU CSWR */
239 240
			.exit_latency = 328 + 440,
			.target_residency = 960,
241
			.flags = CPUIDLE_FLAG_COUPLED,
242
			.enter = omap_enter_idle_coupled,
243
			.name = "C2",
244
			.desc = "CPUx OFF, MPUSS CSWR",
245 246 247 248 249
		},
		{
			/* C3 - CPU0 OFF + CPU1 OFF + MPU OSWR */
			.exit_latency = 460 + 518,
			.target_residency = 1100,
250
			.flags = CPUIDLE_FLAG_COUPLED,
251
			.enter = omap_enter_idle_coupled,
252
			.name = "C3",
253
			.desc = "CPUx OFF, MPUSS OSWR",
254 255
		},
	},
256
	.state_count = ARRAY_SIZE(omap4_idle_data),
257
	.safe_state_index = 0,
258 259
};

260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285
static struct cpuidle_driver omap5_idle_driver = {
	.name				= "omap5_idle",
	.owner				= THIS_MODULE,
	.states = {
		{
			/* C1 - CPU0 ON + CPU1 ON + MPU ON */
			.exit_latency = 2 + 2,
			.target_residency = 5,
			.enter = omap_enter_idle_simple,
			.name = "C1",
			.desc = "CPUx WFI, MPUSS ON"
		},
		{
			/* C2 - CPU0 RET + CPU1 RET + MPU CSWR */
			.exit_latency = 48 + 60,
			.target_residency = 100,
			.flags = CPUIDLE_FLAG_TIMER_STOP,
			.enter = omap_enter_idle_smp,
			.name = "C2",
			.desc = "CPUx CSWR, MPUSS CSWR",
		},
	},
	.state_count = ARRAY_SIZE(omap5_idle_data),
	.safe_state_index = 0,
};

286
/* Public functions */
287

288
/**
289
 * omap4_idle_init - Init routine for OMAP4+ idle
290
 *
291
 * Registers the OMAP4+ specific cpuidle driver to the cpuidle
292 293 294 295
 * framework with the valid set of states.
 */
int __init omap4_idle_init(void)
{
296 297 298 299 300 301 302 303 304 305
	struct cpuidle_driver *idle_driver;

	if (soc_is_omap54xx()) {
		state_ptr = &omap5_idle_data[0];
		idle_driver = &omap5_idle_driver;
	} else {
		state_ptr = &omap4_idle_data[0];
		idle_driver = &omap4_idle_driver;
	}

306
	mpu_pd = pwrdm_lookup("mpu_pwrdm");
307 308 309
	cpu_pd[0] = pwrdm_lookup("cpu0_pwrdm");
	cpu_pd[1] = pwrdm_lookup("cpu1_pwrdm");
	if ((!mpu_pd) || (!cpu_pd[0]) || (!cpu_pd[1]))
310 311
		return -ENODEV;

312 313 314
	cpu_clkdm[0] = clkdm_lookup("mpu0_clkdm");
	cpu_clkdm[1] = clkdm_lookup("mpu1_clkdm");
	if (!cpu_clkdm[0] || !cpu_clkdm[1])
315 316
		return -ENODEV;

317
	return cpuidle_register(idle_driver, cpu_online_mask);
318
}