cpuidle44xx.c 5.9 KB
Newer Older
1
/*
2
 * OMAP4+ CPU idle Routines
3
 *
4
 * Copyright (C) 2011-2013 Texas Instruments, Inc.
5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
 * Santosh Shilimkar <santosh.shilimkar@ti.com>
 * Rajendra Nayak <rnayak@ti.com>
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 as
 * published by the Free Software Foundation.
 */

#include <linux/sched.h>
#include <linux/cpuidle.h>
#include <linux/cpu_pm.h>
#include <linux/export.h>

#include <asm/proc-fns.h>

#include "common.h"
#include "pm.h"
#include "prm.h"
23
#include "clockdomain.h"
24

25
/* Machine specific information */
26
struct idle_statedata {
27 28 29 30 31
	u32 cpu_state;
	u32 mpu_logic_state;
	u32 mpu_state;
};

32
static struct idle_statedata omap4_idle_data[] = {
33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48
	{
		.cpu_state = PWRDM_POWER_ON,
		.mpu_state = PWRDM_POWER_ON,
		.mpu_logic_state = PWRDM_POWER_RET,
	},
	{
		.cpu_state = PWRDM_POWER_OFF,
		.mpu_state = PWRDM_POWER_RET,
		.mpu_logic_state = PWRDM_POWER_RET,
	},
	{
		.cpu_state = PWRDM_POWER_OFF,
		.mpu_state = PWRDM_POWER_RET,
		.mpu_logic_state = PWRDM_POWER_OFF,
	},
};
49

50 51
static struct powerdomain *mpu_pd, *cpu_pd[NR_CPUS];
static struct clockdomain *cpu_clkdm[NR_CPUS];
52

53 54
static atomic_t abort_barrier;
static bool cpu_done[NR_CPUS];
55
static struct idle_statedata *state_ptr = &omap4_idle_data[0];
56

57 58
/* Private functions */

59
/**
60
 * omap_enter_idle_[simple/coupled] - OMAP4PLUS cpuidle entry functions
61 62 63 64 65 66 67 68
 * @dev: cpuidle device
 * @drv: cpuidle driver
 * @index: the index of state to be entered
 *
 * Called from the CPUidle framework to program the device to the
 * specified low power state selected by the governor.
 * Returns the amount of time spent in the low power state.
 */
69
static int omap_enter_idle_simple(struct cpuidle_device *dev,
70 71 72 73 74 75 76 77 78 79
			struct cpuidle_driver *drv,
			int index)
{
	local_fiq_disable();
	omap_do_wfi();
	local_fiq_enable();

	return index;
}

80
static int omap_enter_idle_coupled(struct cpuidle_device *dev,
81 82 83
			struct cpuidle_driver *drv,
			int index)
{
84
	struct idle_statedata *cx = state_ptr + index;
85 86 87 88

	local_fiq_disable();

	/*
89
	 * CPU0 has to wait and stay ON until CPU1 is OFF state.
90 91 92 93
	 * This is necessary to honour hardware recommondation
	 * of triggeing all the possible low power modes once CPU1 is
	 * out of coherency and in OFF mode.
	 */
94
	if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
95
		while (pwrdm_read_pwrst(cpu_pd[1]) != PWRDM_POWER_OFF) {
96
			cpu_relax();
97 98 99 100 101 102 103 104 105 106 107 108

			/*
			 * CPU1 could have already entered & exited idle
			 * without hitting off because of a wakeup
			 * or a failed attempt to hit off mode.  Check for
			 * that here, otherwise we could spin forever
			 * waiting for CPU1 off.
			 */
			if (cpu_done[1])
			    goto fail;

		}
109 110 111 112 113 114
	}

	/*
	 * Call idle CPU PM enter notifier chain so that
	 * VFP and per CPU interrupt context is saved.
	 */
115 116 117 118 119 120 121 122 123 124 125 126 127 128
	cpu_pm_enter();

	if (dev->cpu == 0) {
		pwrdm_set_logic_retst(mpu_pd, cx->mpu_logic_state);
		omap_set_pwrdm_state(mpu_pd, cx->mpu_state);

		/*
		 * Call idle CPU cluster PM enter notifier chain
		 * to save GIC and wakeupgen context.
		 */
		if ((cx->mpu_state == PWRDM_POWER_RET) &&
			(cx->mpu_logic_state == PWRDM_POWER_OFF))
				cpu_cluster_pm_enter();
	}
129 130

	omap4_enter_lowpower(dev->cpu, cx->cpu_state);
131
	cpu_done[dev->cpu] = true;
132

133 134 135 136 137
	/* Wakeup CPU1 only if it is not offlined */
	if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
		clkdm_wakeup(cpu_clkdm[1]);
		clkdm_allow_idle(cpu_clkdm[1]);
	}
138 139 140

	/*
	 * Call idle CPU PM exit notifier chain to restore
141
	 * VFP and per CPU IRQ context.
142
	 */
143
	cpu_pm_exit();
144 145 146 147 148

	/*
	 * Call idle CPU cluster PM exit notifier chain
	 * to restore GIC and wakeupgen context.
	 */
149 150
	if ((cx->mpu_state == PWRDM_POWER_RET) &&
		(cx->mpu_logic_state == PWRDM_POWER_OFF))
151 152
		cpu_cluster_pm_exit();

153 154 155
fail:
	cpuidle_coupled_parallel_barrier(dev, &abort_barrier);
	cpu_done[dev->cpu] = false;
156

157 158 159 160 161
	local_fiq_enable();

	return index;
}

162
static DEFINE_PER_CPU(struct cpuidle_device, omap_idle_dev);
163

164
static struct cpuidle_driver omap4_idle_driver = {
165 166
	.name				= "omap4_idle",
	.owner				= THIS_MODULE,
167 168 169 170 171 172
	.states = {
		{
			/* C1 - CPU0 ON + CPU1 ON + MPU ON */
			.exit_latency = 2 + 2,
			.target_residency = 5,
			.flags = CPUIDLE_FLAG_TIME_VALID,
173
			.enter = omap_enter_idle_simple,
174
			.name = "C1",
175
			.desc = "CPUx ON, MPUSS ON"
176 177
		},
		{
178
			/* C2 - CPU0 OFF + CPU1 OFF + MPU CSWR */
179 180
			.exit_latency = 328 + 440,
			.target_residency = 960,
181 182
			.flags = CPUIDLE_FLAG_TIME_VALID | CPUIDLE_FLAG_COUPLED |
			         CPUIDLE_FLAG_TIMER_STOP,
183
			.enter = omap_enter_idle_coupled,
184
			.name = "C2",
185
			.desc = "CPUx OFF, MPUSS CSWR",
186 187 188 189 190
		},
		{
			/* C3 - CPU0 OFF + CPU1 OFF + MPU OSWR */
			.exit_latency = 460 + 518,
			.target_residency = 1100,
191 192
			.flags = CPUIDLE_FLAG_TIME_VALID | CPUIDLE_FLAG_COUPLED |
			         CPUIDLE_FLAG_TIMER_STOP,
193
			.enter = omap_enter_idle_coupled,
194
			.name = "C3",
195
			.desc = "CPUx OFF, MPUSS OSWR",
196 197
		},
	},
198
	.state_count = ARRAY_SIZE(omap4_idle_data),
199
	.safe_state_index = 0,
200 201
};

202
/* Public functions */
203

204
/**
205
 * omap4_idle_init - Init routine for OMAP4+ idle
206
 *
207
 * Registers the OMAP4+ specific cpuidle driver to the cpuidle
208 209 210 211 212 213 214 215
 * framework with the valid set of states.
 */
int __init omap4_idle_init(void)
{
	struct cpuidle_device *dev;
	unsigned int cpu_id = 0;

	mpu_pd = pwrdm_lookup("mpu_pwrdm");
216 217 218
	cpu_pd[0] = pwrdm_lookup("cpu0_pwrdm");
	cpu_pd[1] = pwrdm_lookup("cpu1_pwrdm");
	if ((!mpu_pd) || (!cpu_pd[0]) || (!cpu_pd[1]))
219 220
		return -ENODEV;

221 222 223
	cpu_clkdm[0] = clkdm_lookup("mpu0_clkdm");
	cpu_clkdm[1] = clkdm_lookup("mpu1_clkdm");
	if (!cpu_clkdm[0] || !cpu_clkdm[1])
224 225
		return -ENODEV;

226 227 228 229
	if (cpuidle_register_driver(&omap4_idle_driver)) {
		pr_err("%s: CPUidle driver register failed\n", __func__);
		return -EIO;
	}
230

231
	for_each_cpu(cpu_id, cpu_online_mask) {
232
		dev = &per_cpu(omap_idle_dev, cpu_id);
233
		dev->cpu = cpu_id;
234
#ifdef CONFIG_ARCH_NEEDS_CPU_IDLE_COUPLED
235
		dev->coupled_cpus = *cpu_online_mask;
236
#endif
237 238
		if (cpuidle_register_device(dev)) {
			pr_err("%s: CPUidle register failed\n", __func__);
239
			cpuidle_unregister_driver(&omap4_idle_driver);
240 241
			return -EIO;
		}
242
	}
243 244 245

	return 0;
}