Linux Audio

Check our new training course

Loading...
v3.5.6
 
  1/*
  2 * linux/arch/arm/mach-omap2/cpuidle34xx.c
  3 *
  4 * OMAP3 CPU IDLE Routines
  5 *
  6 * Copyright (C) 2008 Texas Instruments, Inc.
  7 * Rajendra Nayak <rnayak@ti.com>
  8 *
  9 * Copyright (C) 2007 Texas Instruments, Inc.
 10 * Karthik Dasu <karthik-dp@ti.com>
 11 *
 12 * Copyright (C) 2006 Nokia Corporation
 13 * Tony Lindgren <tony@atomide.com>
 14 *
 15 * Copyright (C) 2005 Texas Instruments, Inc.
 16 * Richard Woodruff <r-woodruff2@ti.com>
 17 *
 18 * Based on pm.c for omap2
 19 *
 20 * This program is free software; you can redistribute it and/or modify
 21 * it under the terms of the GNU General Public License version 2 as
 22 * published by the Free Software Foundation.
 23 */
 24
 25#include <linux/sched.h>
 26#include <linux/cpuidle.h>
 27#include <linux/export.h>
 28#include <linux/cpu_pm.h>
 
 29
 30#include <plat/prcm.h>
 31#include <plat/irqs.h>
 32#include "powerdomain.h"
 33#include "clockdomain.h"
 34
 35#include "pm.h"
 36#include "control.h"
 37#include "common.h"
 38
 39#ifdef CONFIG_CPU_IDLE
 40
 41/* Mach specific information to be recorded in the C-state driver_data */
 42struct omap3_idle_statedata {
 43	u32 mpu_state;
 44	u32 core_state;
 
 
 45};
 46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 47static struct omap3_idle_statedata omap3_idle_data[] = {
 48	{
 49		.mpu_state = PWRDM_POWER_ON,
 50		.core_state = PWRDM_POWER_ON,
 
 
 
 51	},
 52	{
 53		.mpu_state = PWRDM_POWER_ON,
 54		.core_state = PWRDM_POWER_ON,
 
 55	},
 56	{
 57		.mpu_state = PWRDM_POWER_RET,
 58		.core_state = PWRDM_POWER_ON,
 
 59	},
 60	{
 61		.mpu_state = PWRDM_POWER_OFF,
 62		.core_state = PWRDM_POWER_ON,
 
 63	},
 64	{
 65		.mpu_state = PWRDM_POWER_RET,
 66		.core_state = PWRDM_POWER_RET,
 
 67	},
 68	{
 69		.mpu_state = PWRDM_POWER_OFF,
 70		.core_state = PWRDM_POWER_RET,
 
 71	},
 72	{
 73		.mpu_state = PWRDM_POWER_OFF,
 74		.core_state = PWRDM_POWER_OFF,
 
 75	},
 76};
 77
 78static struct powerdomain *mpu_pd, *core_pd, *per_pd, *cam_pd;
 79
 80static int _cpuidle_allow_idle(struct powerdomain *pwrdm,
 81				struct clockdomain *clkdm)
 82{
 83	clkdm_allow_idle(clkdm);
 84	return 0;
 85}
 86
 87static int _cpuidle_deny_idle(struct powerdomain *pwrdm,
 88				struct clockdomain *clkdm)
 89{
 90	clkdm_deny_idle(clkdm);
 91	return 0;
 92}
 93
 94static int __omap3_enter_idle(struct cpuidle_device *dev,
 95				struct cpuidle_driver *drv,
 96				int index)
 97{
 98	struct omap3_idle_statedata *cx = &omap3_idle_data[index];
 99	u32 mpu_state = cx->mpu_state, core_state = cx->core_state;
100
101	local_fiq_disable();
102
103	pwrdm_set_next_pwrst(mpu_pd, mpu_state);
104	pwrdm_set_next_pwrst(core_pd, core_state);
105
106	if (omap_irq_pending() || need_resched())
107		goto return_sleep_time;
108
109	/* Deny idle for C1 */
110	if (index == 0) {
111		pwrdm_for_each_clkdm(mpu_pd, _cpuidle_deny_idle);
112		pwrdm_for_each_clkdm(core_pd, _cpuidle_deny_idle);
 
 
113	}
114
115	/*
116	 * Call idle CPU PM enter notifier chain so that
117	 * VFP context is saved.
118	 */
119	if (mpu_state == PWRDM_POWER_OFF)
120		cpu_pm_enter();
 
 
 
121
122	/* Execute ARM wfi */
123	omap_sram_idle();
124
125	/*
126	 * Call idle CPU PM enter notifier chain to restore
127	 * VFP context.
128	 */
129	if (pwrdm_read_prev_pwrst(mpu_pd) == PWRDM_POWER_OFF)
 
130		cpu_pm_exit();
131
 
132	/* Re-allow idle for C1 */
133	if (index == 0) {
134		pwrdm_for_each_clkdm(mpu_pd, _cpuidle_allow_idle);
135		pwrdm_for_each_clkdm(core_pd, _cpuidle_allow_idle);
136	}
137
138return_sleep_time:
139
140	local_fiq_enable();
141
142	return index;
143}
144
145/**
146 * omap3_enter_idle - Programs OMAP3 to enter the specified state
147 * @dev: cpuidle device
148 * @drv: cpuidle driver
149 * @index: the index of state to be entered
150 *
151 * Called from the CPUidle framework to program the device to the
152 * specified target state selected by the governor.
153 */
154static inline int omap3_enter_idle(struct cpuidle_device *dev,
155				struct cpuidle_driver *drv,
156				int index)
157{
158	return cpuidle_wrap_enter(dev, drv, index, __omap3_enter_idle);
159}
160
161/**
162 * next_valid_state - Find next valid C-state
163 * @dev: cpuidle device
164 * @drv: cpuidle driver
165 * @index: Index of currently selected c-state
166 *
167 * If the state corresponding to index is valid, index is returned back
168 * to the caller. Else, this function searches for a lower c-state which is
169 * still valid (as defined in omap3_power_states[]) and returns its index.
170 *
171 * A state is valid if the 'valid' field is enabled and
172 * if it satisfies the enable_off_mode condition.
173 */
174static int next_valid_state(struct cpuidle_device *dev,
175			    struct cpuidle_driver *drv, int index)
176{
177	struct omap3_idle_statedata *cx = &omap3_idle_data[index];
178	u32 mpu_deepest_state = PWRDM_POWER_RET;
179	u32 core_deepest_state = PWRDM_POWER_RET;
180	int idx;
181	int next_index = -1;
182
183	if (enable_off_mode) {
184		mpu_deepest_state = PWRDM_POWER_OFF;
185		/*
186		 * Erratum i583: valable for ES rev < Es1.2 on 3630.
187		 * CORE OFF mode is not supported in a stable form, restrict
188		 * instead the CORE state to RET.
189		 */
190		if (!IS_PM34XX_ERRATUM(PM_SDRC_WAKEUP_ERRATUM_i583))
191			core_deepest_state = PWRDM_POWER_OFF;
192	}
193
194	/* Check if current state is valid */
195	if ((cx->mpu_state >= mpu_deepest_state) &&
196	    (cx->core_state >= core_deepest_state))
197		return index;
198
199	/*
200	 * Drop to next valid state.
201	 * Start search from the next (lower) state.
202	 */
203	for (idx = index - 1; idx >= 0; idx--) {
204		cx =  &omap3_idle_data[idx];
205		if ((cx->mpu_state >= mpu_deepest_state) &&
206		    (cx->core_state >= core_deepest_state)) {
207			next_index = idx;
208			break;
209		}
210	}
211
212	/*
213	 * C1 is always valid.
214	 * So, no need to check for 'next_index == -1' outside
215	 * this loop.
216	 */
217
218	return next_index;
219}
220
221/**
222 * omap3_enter_idle_bm - Checks for any bus activity
223 * @dev: cpuidle device
224 * @drv: cpuidle driver
225 * @index: array index of target state to be programmed
226 *
227 * This function checks for any pending activity and then programs
228 * the device to the specified or a safer state.
229 */
230static int omap3_enter_idle_bm(struct cpuidle_device *dev,
231				struct cpuidle_driver *drv,
232			       int index)
233{
234	int new_state_idx;
235	u32 core_next_state, per_next_state = 0, per_saved_state = 0, cam_state;
236	struct omap3_idle_statedata *cx;
237	int ret;
238
239	/*
240	 * Prevent idle completely if CAM is active.
241	 * CAM does not have wakeup capability in OMAP3.
242	 */
243	cam_state = pwrdm_read_pwrst(cam_pd);
244	if (cam_state == PWRDM_POWER_ON) {
245		new_state_idx = drv->safe_state_index;
246		goto select_state;
247	}
248
249	/*
250	 * FIXME: we currently manage device-specific idle states
251	 *        for PER and CORE in combination with CPU-specific
252	 *        idle states.  This is wrong, and device-specific
253	 *        idle management needs to be separated out into
254	 *        its own code.
255	 */
256
257	/*
258	 * Prevent PER off if CORE is not in retention or off as this
259	 * would disable PER wakeups completely.
260	 */
261	cx = &omap3_idle_data[index];
262	core_next_state = cx->core_state;
263	per_next_state = per_saved_state = pwrdm_read_next_pwrst(per_pd);
264	if ((per_next_state == PWRDM_POWER_OFF) &&
265	    (core_next_state > PWRDM_POWER_RET))
266		per_next_state = PWRDM_POWER_RET;
267
268	/* Are we changing PER target state? */
269	if (per_next_state != per_saved_state)
 
 
270		pwrdm_set_next_pwrst(per_pd, per_next_state);
 
271
272	new_state_idx = next_valid_state(dev, drv, index);
273
274select_state:
275	ret = omap3_enter_idle(dev, drv, new_state_idx);
276
277	/* Restore original PER state if it was modified */
278	if (per_next_state != per_saved_state)
279		pwrdm_set_next_pwrst(per_pd, per_saved_state);
280
281	return ret;
282}
283
284DEFINE_PER_CPU(struct cpuidle_device, omap3_idle_dev);
285
286struct cpuidle_driver omap3_idle_driver = {
287	.name = 	"omap3_idle",
288	.owner = 	THIS_MODULE,
289	.states = {
290		{
291			.enter		  = omap3_enter_idle,
 
292			.exit_latency	  = 2 + 2,
293			.target_residency = 5,
294			.flags		  = CPUIDLE_FLAG_TIME_VALID,
295			.name		  = "C1",
296			.desc		  = "MPU ON + CORE ON",
297		},
298		{
 
299			.enter		  = omap3_enter_idle_bm,
300			.exit_latency	  = 10 + 10,
301			.target_residency = 30,
302			.flags		  = CPUIDLE_FLAG_TIME_VALID,
303			.name		  = "C2",
304			.desc		  = "MPU ON + CORE ON",
305		},
306		{
 
307			.enter		  = omap3_enter_idle_bm,
308			.exit_latency	  = 50 + 50,
309			.target_residency = 300,
310			.flags		  = CPUIDLE_FLAG_TIME_VALID,
311			.name		  = "C3",
312			.desc		  = "MPU RET + CORE ON",
313		},
314		{
 
315			.enter		  = omap3_enter_idle_bm,
316			.exit_latency	  = 1500 + 1800,
317			.target_residency = 4000,
318			.flags		  = CPUIDLE_FLAG_TIME_VALID,
319			.name		  = "C4",
320			.desc		  = "MPU OFF + CORE ON",
321		},
322		{
 
323			.enter		  = omap3_enter_idle_bm,
324			.exit_latency	  = 2500 + 7500,
325			.target_residency = 12000,
326			.flags		  = CPUIDLE_FLAG_TIME_VALID,
327			.name		  = "C5",
328			.desc		  = "MPU RET + CORE RET",
329		},
330		{
 
331			.enter		  = omap3_enter_idle_bm,
332			.exit_latency	  = 3000 + 8500,
333			.target_residency = 15000,
334			.flags		  = CPUIDLE_FLAG_TIME_VALID,
335			.name		  = "C6",
336			.desc		  = "MPU OFF + CORE RET",
337		},
338		{
 
339			.enter		  = omap3_enter_idle_bm,
340			.exit_latency	  = 10000 + 30000,
341			.target_residency = 30000,
342			.flags		  = CPUIDLE_FLAG_TIME_VALID,
343			.name		  = "C7",
344			.desc		  = "MPU OFF + CORE OFF",
345		},
346	},
347	.state_count = ARRAY_SIZE(omap3_idle_data),
348	.safe_state_index = 0,
349};
350
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
351/**
352 * omap3_idle_init - Init routine for OMAP3 idle
353 *
354 * Registers the OMAP3 specific cpuidle driver to the cpuidle
355 * framework with the valid set of states.
356 */
357int __init omap3_idle_init(void)
358{
359	struct cpuidle_device *dev;
360
361	mpu_pd = pwrdm_lookup("mpu_pwrdm");
362	core_pd = pwrdm_lookup("core_pwrdm");
363	per_pd = pwrdm_lookup("per_pwrdm");
364	cam_pd = pwrdm_lookup("cam_pwrdm");
365
366	if (!mpu_pd || !core_pd || !per_pd || !cam_pd)
367		return -ENODEV;
368
369	cpuidle_register_driver(&omap3_idle_driver);
370
371	dev = &per_cpu(omap3_idle_dev, smp_processor_id());
372	dev->cpu = 0;
373
374	if (cpuidle_register_device(dev)) {
375		printk(KERN_ERR "%s: CPUidle register device failed\n",
376		       __func__);
377		return -EIO;
378	}
379
380	return 0;
381}
382#else
383int __init omap3_idle_init(void)
384{
385	return 0;
386}
387#endif /* CONFIG_CPU_IDLE */
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * linux/arch/arm/mach-omap2/cpuidle34xx.c
  4 *
  5 * OMAP3 CPU IDLE Routines
  6 *
  7 * Copyright (C) 2008 Texas Instruments, Inc.
  8 * Rajendra Nayak <rnayak@ti.com>
  9 *
 10 * Copyright (C) 2007 Texas Instruments, Inc.
 11 * Karthik Dasu <karthik-dp@ti.com>
 12 *
 13 * Copyright (C) 2006 Nokia Corporation
 14 * Tony Lindgren <tony@atomide.com>
 15 *
 16 * Copyright (C) 2005 Texas Instruments, Inc.
 17 * Richard Woodruff <r-woodruff2@ti.com>
 18 *
 19 * Based on pm.c for omap2
 
 
 
 
 20 */
 21
 22#include <linux/sched.h>
 23#include <linux/cpuidle.h>
 24#include <linux/export.h>
 25#include <linux/cpu_pm.h>
 26#include <asm/cpuidle.h>
 27
 
 
 28#include "powerdomain.h"
 29#include "clockdomain.h"
 30
 31#include "pm.h"
 32#include "control.h"
 33#include "common.h"
 34#include "soc.h"
 
 35
 36/* Mach specific information to be recorded in the C-state driver_data */
 37struct omap3_idle_statedata {
 38	u8 mpu_state;
 39	u8 core_state;
 40	u8 per_min_state;
 41	u8 flags;
 42};
 43
 44static struct powerdomain *mpu_pd, *core_pd, *per_pd, *cam_pd;
 45
 46/*
 47 * Possible flag bits for struct omap3_idle_statedata.flags:
 48 *
 49 * OMAP_CPUIDLE_CX_NO_CLKDM_IDLE: don't allow the MPU clockdomain to go
 50 *    inactive.  This in turn prevents the MPU DPLL from entering autoidle
 51 *    mode, so wakeup latency is greatly reduced, at the cost of additional
 52 *    energy consumption.  This also prevents the CORE clockdomain from
 53 *    entering idle.
 54 */
 55#define OMAP_CPUIDLE_CX_NO_CLKDM_IDLE		BIT(0)
 56
 57/*
 58 * Prevent PER OFF if CORE is not in RETention or OFF as this would
 59 * disable PER wakeups completely.
 60 */
 61static struct omap3_idle_statedata omap3_idle_data[] = {
 62	{
 63		.mpu_state = PWRDM_POWER_ON,
 64		.core_state = PWRDM_POWER_ON,
 65		/* In C1 do not allow PER state lower than CORE state */
 66		.per_min_state = PWRDM_POWER_ON,
 67		.flags = OMAP_CPUIDLE_CX_NO_CLKDM_IDLE,
 68	},
 69	{
 70		.mpu_state = PWRDM_POWER_ON,
 71		.core_state = PWRDM_POWER_ON,
 72		.per_min_state = PWRDM_POWER_RET,
 73	},
 74	{
 75		.mpu_state = PWRDM_POWER_RET,
 76		.core_state = PWRDM_POWER_ON,
 77		.per_min_state = PWRDM_POWER_RET,
 78	},
 79	{
 80		.mpu_state = PWRDM_POWER_OFF,
 81		.core_state = PWRDM_POWER_ON,
 82		.per_min_state = PWRDM_POWER_RET,
 83	},
 84	{
 85		.mpu_state = PWRDM_POWER_RET,
 86		.core_state = PWRDM_POWER_RET,
 87		.per_min_state = PWRDM_POWER_OFF,
 88	},
 89	{
 90		.mpu_state = PWRDM_POWER_OFF,
 91		.core_state = PWRDM_POWER_RET,
 92		.per_min_state = PWRDM_POWER_OFF,
 93	},
 94	{
 95		.mpu_state = PWRDM_POWER_OFF,
 96		.core_state = PWRDM_POWER_OFF,
 97		.per_min_state = PWRDM_POWER_OFF,
 98	},
 99};
100
101/**
102 * omap3_enter_idle - Programs OMAP3 to enter the specified state
103 * @dev: cpuidle device
104 * @drv: cpuidle driver
105 * @index: the index of state to be entered
106 */
107static int omap3_enter_idle(struct cpuidle_device *dev,
108			    struct cpuidle_driver *drv,
109			    int index)
 
 
 
 
 
 
 
 
 
 
110{
111	struct omap3_idle_statedata *cx = &omap3_idle_data[index];
112	int error;
 
 
 
 
 
113
114	if (omap_irq_pending() || need_resched())
115		goto return_sleep_time;
116
117	/* Deny idle for C1 */
118	if (cx->flags & OMAP_CPUIDLE_CX_NO_CLKDM_IDLE) {
119		clkdm_deny_idle(mpu_pd->pwrdm_clkdms[0]);
120	} else {
121		pwrdm_set_next_pwrst(mpu_pd, cx->mpu_state);
122		pwrdm_set_next_pwrst(core_pd, cx->core_state);
123	}
124
125	/*
126	 * Call idle CPU PM enter notifier chain so that
127	 * VFP context is saved.
128	 */
129	if (cx->mpu_state == PWRDM_POWER_OFF) {
130		error = cpu_pm_enter();
131		if (error)
132			goto out_clkdm_set;
133	}
134
135	/* Execute ARM wfi */
136	omap_sram_idle(true);
137
138	/*
139	 * Call idle CPU PM enter notifier chain to restore
140	 * VFP context.
141	 */
142	if (cx->mpu_state == PWRDM_POWER_OFF &&
143	    pwrdm_read_prev_pwrst(mpu_pd) == PWRDM_POWER_OFF)
144		cpu_pm_exit();
145
146out_clkdm_set:
147	/* Re-allow idle for C1 */
148	if (cx->flags & OMAP_CPUIDLE_CX_NO_CLKDM_IDLE)
149		clkdm_allow_idle(mpu_pd->pwrdm_clkdms[0]);
 
 
150
151return_sleep_time:
152
 
 
153	return index;
154}
155
156/**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157 * next_valid_state - Find next valid C-state
158 * @dev: cpuidle device
159 * @drv: cpuidle driver
160 * @index: Index of currently selected c-state
161 *
162 * If the state corresponding to index is valid, index is returned back
163 * to the caller. Else, this function searches for a lower c-state which is
164 * still valid (as defined in omap3_power_states[]) and returns its index.
165 *
166 * A state is valid if the 'valid' field is enabled and
167 * if it satisfies the enable_off_mode condition.
168 */
169static int next_valid_state(struct cpuidle_device *dev,
170			    struct cpuidle_driver *drv, int index)
171{
172	struct omap3_idle_statedata *cx = &omap3_idle_data[index];
173	u32 mpu_deepest_state = PWRDM_POWER_RET;
174	u32 core_deepest_state = PWRDM_POWER_RET;
175	int idx;
176	int next_index = 0; /* C1 is the default value */
177
178	if (enable_off_mode) {
179		mpu_deepest_state = PWRDM_POWER_OFF;
180		/*
181		 * Erratum i583: valable for ES rev < Es1.2 on 3630.
182		 * CORE OFF mode is not supported in a stable form, restrict
183		 * instead the CORE state to RET.
184		 */
185		if (!IS_PM34XX_ERRATUM(PM_SDRC_WAKEUP_ERRATUM_i583))
186			core_deepest_state = PWRDM_POWER_OFF;
187	}
188
189	/* Check if current state is valid */
190	if ((cx->mpu_state >= mpu_deepest_state) &&
191	    (cx->core_state >= core_deepest_state))
192		return index;
193
194	/*
195	 * Drop to next valid state.
196	 * Start search from the next (lower) state.
197	 */
198	for (idx = index - 1; idx >= 0; idx--) {
199		cx = &omap3_idle_data[idx];
200		if ((cx->mpu_state >= mpu_deepest_state) &&
201		    (cx->core_state >= core_deepest_state)) {
202			next_index = idx;
203			break;
204		}
205	}
206
 
 
 
 
 
 
207	return next_index;
208}
209
210/**
211 * omap3_enter_idle_bm - Checks for any bus activity
212 * @dev: cpuidle device
213 * @drv: cpuidle driver
214 * @index: array index of target state to be programmed
215 *
216 * This function checks for any pending activity and then programs
217 * the device to the specified or a safer state.
218 */
219static int omap3_enter_idle_bm(struct cpuidle_device *dev,
220			       struct cpuidle_driver *drv,
221			       int index)
222{
223	int new_state_idx, ret;
224	u8 per_next_state, per_saved_state;
225	struct omap3_idle_statedata *cx;
 
226
227	/*
228	 * Use only C1 if CAM is active.
229	 * CAM does not have wakeup capability in OMAP3.
230	 */
231	if (pwrdm_read_pwrst(cam_pd) == PWRDM_POWER_ON)
 
232		new_state_idx = drv->safe_state_index;
233	else
234		new_state_idx = next_valid_state(dev, drv, index);
235
236	/*
237	 * FIXME: we currently manage device-specific idle states
238	 *        for PER and CORE in combination with CPU-specific
239	 *        idle states.  This is wrong, and device-specific
240	 *        idle management needs to be separated out into
241	 *        its own code.
242	 */
243
244	/* Program PER state */
245	cx = &omap3_idle_data[new_state_idx];
 
 
 
 
 
 
 
 
246
247	per_next_state = pwrdm_read_next_pwrst(per_pd);
248	per_saved_state = per_next_state;
249	if (per_next_state < cx->per_min_state) {
250		per_next_state = cx->per_min_state;
251		pwrdm_set_next_pwrst(per_pd, per_next_state);
252	}
253
 
 
 
254	ret = omap3_enter_idle(dev, drv, new_state_idx);
255
256	/* Restore original PER state if it was modified */
257	if (per_next_state != per_saved_state)
258		pwrdm_set_next_pwrst(per_pd, per_saved_state);
259
260	return ret;
261}
262
263static struct cpuidle_driver omap3_idle_driver = {
264	.name             = "omap3_idle",
265	.owner            = THIS_MODULE,
 
 
266	.states = {
267		{
268			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
269			.enter		  = omap3_enter_idle_bm,
270			.exit_latency	  = 2 + 2,
271			.target_residency = 5,
 
272			.name		  = "C1",
273			.desc		  = "MPU ON + CORE ON",
274		},
275		{
276			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
277			.enter		  = omap3_enter_idle_bm,
278			.exit_latency	  = 10 + 10,
279			.target_residency = 30,
 
280			.name		  = "C2",
281			.desc		  = "MPU ON + CORE ON",
282		},
283		{
284			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
285			.enter		  = omap3_enter_idle_bm,
286			.exit_latency	  = 50 + 50,
287			.target_residency = 300,
 
288			.name		  = "C3",
289			.desc		  = "MPU RET + CORE ON",
290		},
291		{
292			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
293			.enter		  = omap3_enter_idle_bm,
294			.exit_latency	  = 1500 + 1800,
295			.target_residency = 4000,
 
296			.name		  = "C4",
297			.desc		  = "MPU OFF + CORE ON",
298		},
299		{
300			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
301			.enter		  = omap3_enter_idle_bm,
302			.exit_latency	  = 2500 + 7500,
303			.target_residency = 12000,
 
304			.name		  = "C5",
305			.desc		  = "MPU RET + CORE RET",
306		},
307		{
308			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
309			.enter		  = omap3_enter_idle_bm,
310			.exit_latency	  = 3000 + 8500,
311			.target_residency = 15000,
 
312			.name		  = "C6",
313			.desc		  = "MPU OFF + CORE RET",
314		},
315		{
316			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
317			.enter		  = omap3_enter_idle_bm,
318			.exit_latency	  = 10000 + 30000,
319			.target_residency = 30000,
 
320			.name		  = "C7",
321			.desc		  = "MPU OFF + CORE OFF",
322		},
323	},
324	.state_count = ARRAY_SIZE(omap3_idle_data),
325	.safe_state_index = 0,
326};
327
328/*
329 * Numbers based on measurements made in October 2009 for PM optimized kernel
330 * with CPU freq enabled on device Nokia N900. Assumes OPP2 (main idle OPP,
331 * and worst case latencies).
332 */
333static struct cpuidle_driver omap3430_idle_driver = {
334	.name             = "omap3430_idle",
335	.owner            = THIS_MODULE,
336	.states = {
337		{
338			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
339			.enter		  = omap3_enter_idle_bm,
340			.exit_latency	  = 110 + 162,
341			.target_residency = 5,
342			.name		  = "C1",
343			.desc		  = "MPU ON + CORE ON",
344		},
345		{
346			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
347			.enter		  = omap3_enter_idle_bm,
348			.exit_latency	  = 106 + 180,
349			.target_residency = 309,
350			.name		  = "C2",
351			.desc		  = "MPU ON + CORE ON",
352		},
353		{
354			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
355			.enter		  = omap3_enter_idle_bm,
356			.exit_latency	  = 107 + 410,
357			.target_residency = 46057,
358			.name		  = "C3",
359			.desc		  = "MPU RET + CORE ON",
360		},
361		{
362			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
363			.enter		  = omap3_enter_idle_bm,
364			.exit_latency	  = 121 + 3374,
365			.target_residency = 46057,
366			.name		  = "C4",
367			.desc		  = "MPU OFF + CORE ON",
368		},
369		{
370			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
371			.enter		  = omap3_enter_idle_bm,
372			.exit_latency	  = 855 + 1146,
373			.target_residency = 46057,
374			.name		  = "C5",
375			.desc		  = "MPU RET + CORE RET",
376		},
377		{
378			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
379			.enter		  = omap3_enter_idle_bm,
380			.exit_latency	  = 7580 + 4134,
381			.target_residency = 484329,
382			.name		  = "C6",
383			.desc		  = "MPU OFF + CORE RET",
384		},
385		{
386			.flags		  = CPUIDLE_FLAG_RCU_IDLE,
387			.enter		  = omap3_enter_idle_bm,
388			.exit_latency	  = 7505 + 15274,
389			.target_residency = 484329,
390			.name		  = "C7",
391			.desc		  = "MPU OFF + CORE OFF",
392		},
393	},
394	.state_count = ARRAY_SIZE(omap3_idle_data),
395	.safe_state_index = 0,
396};
397
398/* Public functions */
399
400/**
401 * omap3_idle_init - Init routine for OMAP3 idle
402 *
403 * Registers the OMAP3 specific cpuidle driver to the cpuidle
404 * framework with the valid set of states.
405 */
406int __init omap3_idle_init(void)
407{
 
 
408	mpu_pd = pwrdm_lookup("mpu_pwrdm");
409	core_pd = pwrdm_lookup("core_pwrdm");
410	per_pd = pwrdm_lookup("per_pwrdm");
411	cam_pd = pwrdm_lookup("cam_pwrdm");
412
413	if (!mpu_pd || !core_pd || !per_pd || !cam_pd)
414		return -ENODEV;
415
416	if (cpu_is_omap3430())
417		return cpuidle_register(&omap3430_idle_driver, NULL);
418	else
419		return cpuidle_register(&omap3_idle_driver, NULL);
 
 
 
 
 
 
 
 
 
 
 
 
 
420}