CPU hotplug, re-create sysfs directory and symlinks
[linux-2.6/linux-acpi-2.6/ibm-acpi-2.6.git] / drivers / cpufreq / cpufreq.c
blob7c10f96c5ae981361875e50ea36b028f59b31d32
1 /*
2 * linux/drivers/cpufreq/cpufreq.c
4 * Copyright (C) 2001 Russell King
5 * (C) 2002 - 2003 Dominik Brodowski <linux@brodo.de>
7 * Oct 2005 - Ashok Raj <ashok.raj@intel.com>
8 * Added handling for CPU hotplug
9 * Feb 2006 - Jacob Shin <jacob.shin@amd.com>
10 * Fix handling for CPU hotplug -- affected CPUs
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of the GNU General Public License version 2 as
14 * published by the Free Software Foundation.
18 #include <linux/kernel.h>
19 #include <linux/module.h>
20 #include <linux/init.h>
21 #include <linux/notifier.h>
22 #include <linux/cpufreq.h>
23 #include <linux/delay.h>
24 #include <linux/interrupt.h>
25 #include <linux/spinlock.h>
26 #include <linux/device.h>
27 #include <linux/slab.h>
28 #include <linux/cpu.h>
29 #include <linux/completion.h>
30 #include <linux/mutex.h>
31 #include <linux/syscore_ops.h>
33 #include <trace/events/power.h>
35 #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_CORE, \
36 "cpufreq-core", msg)
38 /**
39 * The "cpufreq driver" - the arch- or hardware-dependent low
40 * level driver of CPUFreq support, and its spinlock. This lock
41 * also protects the cpufreq_cpu_data array.
43 static struct cpufreq_driver *cpufreq_driver;
44 static DEFINE_PER_CPU(struct cpufreq_policy *, cpufreq_cpu_data);
45 #ifdef CONFIG_HOTPLUG_CPU
46 /* This one keeps track of the previously set governor of a removed CPU */
47 static DEFINE_PER_CPU(char[CPUFREQ_NAME_LEN], cpufreq_cpu_governor);
48 #endif
49 static DEFINE_SPINLOCK(cpufreq_driver_lock);
52 * cpu_policy_rwsem is a per CPU reader-writer semaphore designed to cure
53 * all cpufreq/hotplug/workqueue/etc related lock issues.
55 * The rules for this semaphore:
56 * - Any routine that wants to read from the policy structure will
57 * do a down_read on this semaphore.
58 * - Any routine that will write to the policy structure and/or may take away
59 * the policy altogether (eg. CPU hotplug), will hold this lock in write
60 * mode before doing so.
62 * Additional rules:
63 * - All holders of the lock should check to make sure that the CPU they
64 * are concerned with are online after they get the lock.
65 * - Governor routines that can be called in cpufreq hotplug path should not
66 * take this sem as top level hotplug notifier handler takes this.
67 * - Lock should not be held across
68 * __cpufreq_governor(data, CPUFREQ_GOV_STOP);
70 static DEFINE_PER_CPU(int, cpufreq_policy_cpu);
71 static DEFINE_PER_CPU(struct rw_semaphore, cpu_policy_rwsem);
73 #define lock_policy_rwsem(mode, cpu) \
74 static int lock_policy_rwsem_##mode \
75 (int cpu) \
76 { \
77 int policy_cpu = per_cpu(cpufreq_policy_cpu, cpu); \
78 BUG_ON(policy_cpu == -1); \
79 down_##mode(&per_cpu(cpu_policy_rwsem, policy_cpu)); \
80 if (unlikely(!cpu_online(cpu))) { \
81 up_##mode(&per_cpu(cpu_policy_rwsem, policy_cpu)); \
82 return -1; \
83 } \
85 return 0; \
88 lock_policy_rwsem(read, cpu);
90 lock_policy_rwsem(write, cpu);
92 static void unlock_policy_rwsem_read(int cpu)
94 int policy_cpu = per_cpu(cpufreq_policy_cpu, cpu);
95 BUG_ON(policy_cpu == -1);
96 up_read(&per_cpu(cpu_policy_rwsem, policy_cpu));
99 static void unlock_policy_rwsem_write(int cpu)
101 int policy_cpu = per_cpu(cpufreq_policy_cpu, cpu);
102 BUG_ON(policy_cpu == -1);
103 up_write(&per_cpu(cpu_policy_rwsem, policy_cpu));
107 /* internal prototypes */
108 static int __cpufreq_governor(struct cpufreq_policy *policy,
109 unsigned int event);
110 static unsigned int __cpufreq_get(unsigned int cpu);
111 static void handle_update(struct work_struct *work);
114 * Two notifier lists: the "policy" list is involved in the
115 * validation process for a new CPU frequency policy; the
116 * "transition" list for kernel code that needs to handle
117 * changes to devices when the CPU clock speed changes.
118 * The mutex locks both lists.
120 static BLOCKING_NOTIFIER_HEAD(cpufreq_policy_notifier_list);
121 static struct srcu_notifier_head cpufreq_transition_notifier_list;
123 static bool init_cpufreq_transition_notifier_list_called;
124 static int __init init_cpufreq_transition_notifier_list(void)
126 srcu_init_notifier_head(&cpufreq_transition_notifier_list);
127 init_cpufreq_transition_notifier_list_called = true;
128 return 0;
130 pure_initcall(init_cpufreq_transition_notifier_list);
132 static LIST_HEAD(cpufreq_governor_list);
133 static DEFINE_MUTEX(cpufreq_governor_mutex);
135 struct cpufreq_policy *cpufreq_cpu_get(unsigned int cpu)
137 struct cpufreq_policy *data;
138 unsigned long flags;
140 if (cpu >= nr_cpu_ids)
141 goto err_out;
143 /* get the cpufreq driver */
144 spin_lock_irqsave(&cpufreq_driver_lock, flags);
146 if (!cpufreq_driver)
147 goto err_out_unlock;
149 if (!try_module_get(cpufreq_driver->owner))
150 goto err_out_unlock;
153 /* get the CPU */
154 data = per_cpu(cpufreq_cpu_data, cpu);
156 if (!data)
157 goto err_out_put_module;
159 if (!kobject_get(&data->kobj))
160 goto err_out_put_module;
162 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
163 return data;
165 err_out_put_module:
166 module_put(cpufreq_driver->owner);
167 err_out_unlock:
168 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
169 err_out:
170 return NULL;
172 EXPORT_SYMBOL_GPL(cpufreq_cpu_get);
175 void cpufreq_cpu_put(struct cpufreq_policy *data)
177 kobject_put(&data->kobj);
178 module_put(cpufreq_driver->owner);
180 EXPORT_SYMBOL_GPL(cpufreq_cpu_put);
183 /*********************************************************************
184 * UNIFIED DEBUG HELPERS *
185 *********************************************************************/
186 #ifdef CONFIG_CPU_FREQ_DEBUG
188 /* what part(s) of the CPUfreq subsystem are debugged? */
189 static unsigned int debug;
191 /* is the debug output ratelimit'ed using printk_ratelimit? User can
192 * set or modify this value.
194 static unsigned int debug_ratelimit = 1;
196 /* is the printk_ratelimit'ing enabled? It's enabled after a successful
197 * loading of a cpufreq driver, temporarily disabled when a new policy
198 * is set, and disabled upon cpufreq driver removal
200 static unsigned int disable_ratelimit = 1;
201 static DEFINE_SPINLOCK(disable_ratelimit_lock);
203 static void cpufreq_debug_enable_ratelimit(void)
205 unsigned long flags;
207 spin_lock_irqsave(&disable_ratelimit_lock, flags);
208 if (disable_ratelimit)
209 disable_ratelimit--;
210 spin_unlock_irqrestore(&disable_ratelimit_lock, flags);
213 static void cpufreq_debug_disable_ratelimit(void)
215 unsigned long flags;
217 spin_lock_irqsave(&disable_ratelimit_lock, flags);
218 disable_ratelimit++;
219 spin_unlock_irqrestore(&disable_ratelimit_lock, flags);
222 void cpufreq_debug_printk(unsigned int type, const char *prefix,
223 const char *fmt, ...)
225 char s[256];
226 va_list args;
227 unsigned int len;
228 unsigned long flags;
230 WARN_ON(!prefix);
231 if (type & debug) {
232 spin_lock_irqsave(&disable_ratelimit_lock, flags);
233 if (!disable_ratelimit && debug_ratelimit
234 && !printk_ratelimit()) {
235 spin_unlock_irqrestore(&disable_ratelimit_lock, flags);
236 return;
238 spin_unlock_irqrestore(&disable_ratelimit_lock, flags);
240 len = snprintf(s, 256, KERN_DEBUG "%s: ", prefix);
242 va_start(args, fmt);
243 len += vsnprintf(&s[len], (256 - len), fmt, args);
244 va_end(args);
246 printk(s);
248 WARN_ON(len < 5);
251 EXPORT_SYMBOL(cpufreq_debug_printk);
254 module_param(debug, uint, 0644);
255 MODULE_PARM_DESC(debug, "CPUfreq debugging: add 1 to debug core,"
256 " 2 to debug drivers, and 4 to debug governors.");
258 module_param(debug_ratelimit, uint, 0644);
259 MODULE_PARM_DESC(debug_ratelimit, "CPUfreq debugging:"
260 " set to 0 to disable ratelimiting.");
262 #else /* !CONFIG_CPU_FREQ_DEBUG */
264 static inline void cpufreq_debug_enable_ratelimit(void) { return; }
265 static inline void cpufreq_debug_disable_ratelimit(void) { return; }
267 #endif /* CONFIG_CPU_FREQ_DEBUG */
270 /*********************************************************************
271 * EXTERNALLY AFFECTING FREQUENCY CHANGES *
272 *********************************************************************/
275 * adjust_jiffies - adjust the system "loops_per_jiffy"
277 * This function alters the system "loops_per_jiffy" for the clock
278 * speed change. Note that loops_per_jiffy cannot be updated on SMP
279 * systems as each CPU might be scaled differently. So, use the arch
280 * per-CPU loops_per_jiffy value wherever possible.
282 #ifndef CONFIG_SMP
283 static unsigned long l_p_j_ref;
284 static unsigned int l_p_j_ref_freq;
286 static void adjust_jiffies(unsigned long val, struct cpufreq_freqs *ci)
288 if (ci->flags & CPUFREQ_CONST_LOOPS)
289 return;
291 if (!l_p_j_ref_freq) {
292 l_p_j_ref = loops_per_jiffy;
293 l_p_j_ref_freq = ci->old;
294 dprintk("saving %lu as reference value for loops_per_jiffy; "
295 "freq is %u kHz\n", l_p_j_ref, l_p_j_ref_freq);
297 if ((val == CPUFREQ_PRECHANGE && ci->old < ci->new) ||
298 (val == CPUFREQ_POSTCHANGE && ci->old > ci->new) ||
299 (val == CPUFREQ_RESUMECHANGE || val == CPUFREQ_SUSPENDCHANGE)) {
300 loops_per_jiffy = cpufreq_scale(l_p_j_ref, l_p_j_ref_freq,
301 ci->new);
302 dprintk("scaling loops_per_jiffy to %lu "
303 "for frequency %u kHz\n", loops_per_jiffy, ci->new);
306 #else
307 static inline void adjust_jiffies(unsigned long val, struct cpufreq_freqs *ci)
309 return;
311 #endif
315 * cpufreq_notify_transition - call notifier chain and adjust_jiffies
316 * on frequency transition.
318 * This function calls the transition notifiers and the "adjust_jiffies"
319 * function. It is called twice on all CPU frequency changes that have
320 * external effects.
322 void cpufreq_notify_transition(struct cpufreq_freqs *freqs, unsigned int state)
324 struct cpufreq_policy *policy;
326 BUG_ON(irqs_disabled());
328 freqs->flags = cpufreq_driver->flags;
329 dprintk("notification %u of frequency transition to %u kHz\n",
330 state, freqs->new);
332 policy = per_cpu(cpufreq_cpu_data, freqs->cpu);
333 switch (state) {
335 case CPUFREQ_PRECHANGE:
336 /* detect if the driver reported a value as "old frequency"
337 * which is not equal to what the cpufreq core thinks is
338 * "old frequency".
340 if (!(cpufreq_driver->flags & CPUFREQ_CONST_LOOPS)) {
341 if ((policy) && (policy->cpu == freqs->cpu) &&
342 (policy->cur) && (policy->cur != freqs->old)) {
343 dprintk("Warning: CPU frequency is"
344 " %u, cpufreq assumed %u kHz.\n",
345 freqs->old, policy->cur);
346 freqs->old = policy->cur;
349 srcu_notifier_call_chain(&cpufreq_transition_notifier_list,
350 CPUFREQ_PRECHANGE, freqs);
351 adjust_jiffies(CPUFREQ_PRECHANGE, freqs);
352 break;
354 case CPUFREQ_POSTCHANGE:
355 adjust_jiffies(CPUFREQ_POSTCHANGE, freqs);
356 dprintk("FREQ: %lu - CPU: %lu", (unsigned long)freqs->new,
357 (unsigned long)freqs->cpu);
358 trace_power_frequency(POWER_PSTATE, freqs->new, freqs->cpu);
359 trace_cpu_frequency(freqs->new, freqs->cpu);
360 srcu_notifier_call_chain(&cpufreq_transition_notifier_list,
361 CPUFREQ_POSTCHANGE, freqs);
362 if (likely(policy) && likely(policy->cpu == freqs->cpu))
363 policy->cur = freqs->new;
364 break;
367 EXPORT_SYMBOL_GPL(cpufreq_notify_transition);
371 /*********************************************************************
372 * SYSFS INTERFACE *
373 *********************************************************************/
375 static struct cpufreq_governor *__find_governor(const char *str_governor)
377 struct cpufreq_governor *t;
379 list_for_each_entry(t, &cpufreq_governor_list, governor_list)
380 if (!strnicmp(str_governor, t->name, CPUFREQ_NAME_LEN))
381 return t;
383 return NULL;
387 * cpufreq_parse_governor - parse a governor string
389 static int cpufreq_parse_governor(char *str_governor, unsigned int *policy,
390 struct cpufreq_governor **governor)
392 int err = -EINVAL;
394 if (!cpufreq_driver)
395 goto out;
397 if (cpufreq_driver->setpolicy) {
398 if (!strnicmp(str_governor, "performance", CPUFREQ_NAME_LEN)) {
399 *policy = CPUFREQ_POLICY_PERFORMANCE;
400 err = 0;
401 } else if (!strnicmp(str_governor, "powersave",
402 CPUFREQ_NAME_LEN)) {
403 *policy = CPUFREQ_POLICY_POWERSAVE;
404 err = 0;
406 } else if (cpufreq_driver->target) {
407 struct cpufreq_governor *t;
409 mutex_lock(&cpufreq_governor_mutex);
411 t = __find_governor(str_governor);
413 if (t == NULL) {
414 char *name = kasprintf(GFP_KERNEL, "cpufreq_%s",
415 str_governor);
417 if (name) {
418 int ret;
420 mutex_unlock(&cpufreq_governor_mutex);
421 ret = request_module("%s", name);
422 mutex_lock(&cpufreq_governor_mutex);
424 if (ret == 0)
425 t = __find_governor(str_governor);
428 kfree(name);
431 if (t != NULL) {
432 *governor = t;
433 err = 0;
436 mutex_unlock(&cpufreq_governor_mutex);
438 out:
439 return err;
444 * cpufreq_per_cpu_attr_read() / show_##file_name() -
445 * print out cpufreq information
447 * Write out information from cpufreq_driver->policy[cpu]; object must be
448 * "unsigned int".
451 #define show_one(file_name, object) \
452 static ssize_t show_##file_name \
453 (struct cpufreq_policy *policy, char *buf) \
455 return sprintf(buf, "%u\n", policy->object); \
458 show_one(cpuinfo_min_freq, cpuinfo.min_freq);
459 show_one(cpuinfo_max_freq, cpuinfo.max_freq);
460 show_one(cpuinfo_transition_latency, cpuinfo.transition_latency);
461 show_one(scaling_min_freq, min);
462 show_one(scaling_max_freq, max);
463 show_one(scaling_cur_freq, cur);
465 static int __cpufreq_set_policy(struct cpufreq_policy *data,
466 struct cpufreq_policy *policy);
469 * cpufreq_per_cpu_attr_write() / store_##file_name() - sysfs write access
471 #define store_one(file_name, object) \
472 static ssize_t store_##file_name \
473 (struct cpufreq_policy *policy, const char *buf, size_t count) \
475 unsigned int ret = -EINVAL; \
476 struct cpufreq_policy new_policy; \
478 ret = cpufreq_get_policy(&new_policy, policy->cpu); \
479 if (ret) \
480 return -EINVAL; \
482 ret = sscanf(buf, "%u", &new_policy.object); \
483 if (ret != 1) \
484 return -EINVAL; \
486 ret = __cpufreq_set_policy(policy, &new_policy); \
487 policy->user_policy.object = policy->object; \
489 return ret ? ret : count; \
492 store_one(scaling_min_freq, min);
493 store_one(scaling_max_freq, max);
496 * show_cpuinfo_cur_freq - current CPU frequency as detected by hardware
498 static ssize_t show_cpuinfo_cur_freq(struct cpufreq_policy *policy,
499 char *buf)
501 unsigned int cur_freq = __cpufreq_get(policy->cpu);
502 if (!cur_freq)
503 return sprintf(buf, "<unknown>");
504 return sprintf(buf, "%u\n", cur_freq);
509 * show_scaling_governor - show the current policy for the specified CPU
511 static ssize_t show_scaling_governor(struct cpufreq_policy *policy, char *buf)
513 if (policy->policy == CPUFREQ_POLICY_POWERSAVE)
514 return sprintf(buf, "powersave\n");
515 else if (policy->policy == CPUFREQ_POLICY_PERFORMANCE)
516 return sprintf(buf, "performance\n");
517 else if (policy->governor)
518 return scnprintf(buf, CPUFREQ_NAME_LEN, "%s\n",
519 policy->governor->name);
520 return -EINVAL;
525 * store_scaling_governor - store policy for the specified CPU
527 static ssize_t store_scaling_governor(struct cpufreq_policy *policy,
528 const char *buf, size_t count)
530 unsigned int ret = -EINVAL;
531 char str_governor[16];
532 struct cpufreq_policy new_policy;
534 ret = cpufreq_get_policy(&new_policy, policy->cpu);
535 if (ret)
536 return ret;
538 ret = sscanf(buf, "%15s", str_governor);
539 if (ret != 1)
540 return -EINVAL;
542 if (cpufreq_parse_governor(str_governor, &new_policy.policy,
543 &new_policy.governor))
544 return -EINVAL;
546 /* Do not use cpufreq_set_policy here or the user_policy.max
547 will be wrongly overridden */
548 ret = __cpufreq_set_policy(policy, &new_policy);
550 policy->user_policy.policy = policy->policy;
551 policy->user_policy.governor = policy->governor;
553 if (ret)
554 return ret;
555 else
556 return count;
560 * show_scaling_driver - show the cpufreq driver currently loaded
562 static ssize_t show_scaling_driver(struct cpufreq_policy *policy, char *buf)
564 return scnprintf(buf, CPUFREQ_NAME_LEN, "%s\n", cpufreq_driver->name);
568 * show_scaling_available_governors - show the available CPUfreq governors
570 static ssize_t show_scaling_available_governors(struct cpufreq_policy *policy,
571 char *buf)
573 ssize_t i = 0;
574 struct cpufreq_governor *t;
576 if (!cpufreq_driver->target) {
577 i += sprintf(buf, "performance powersave");
578 goto out;
581 list_for_each_entry(t, &cpufreq_governor_list, governor_list) {
582 if (i >= (ssize_t) ((PAGE_SIZE / sizeof(char))
583 - (CPUFREQ_NAME_LEN + 2)))
584 goto out;
585 i += scnprintf(&buf[i], CPUFREQ_NAME_LEN, "%s ", t->name);
587 out:
588 i += sprintf(&buf[i], "\n");
589 return i;
592 static ssize_t show_cpus(const struct cpumask *mask, char *buf)
594 ssize_t i = 0;
595 unsigned int cpu;
597 for_each_cpu(cpu, mask) {
598 if (i)
599 i += scnprintf(&buf[i], (PAGE_SIZE - i - 2), " ");
600 i += scnprintf(&buf[i], (PAGE_SIZE - i - 2), "%u", cpu);
601 if (i >= (PAGE_SIZE - 5))
602 break;
604 i += sprintf(&buf[i], "\n");
605 return i;
609 * show_related_cpus - show the CPUs affected by each transition even if
610 * hw coordination is in use
612 static ssize_t show_related_cpus(struct cpufreq_policy *policy, char *buf)
614 if (cpumask_empty(policy->related_cpus))
615 return show_cpus(policy->cpus, buf);
616 return show_cpus(policy->related_cpus, buf);
620 * show_affected_cpus - show the CPUs affected by each transition
622 static ssize_t show_affected_cpus(struct cpufreq_policy *policy, char *buf)
624 return show_cpus(policy->cpus, buf);
627 static ssize_t store_scaling_setspeed(struct cpufreq_policy *policy,
628 const char *buf, size_t count)
630 unsigned int freq = 0;
631 unsigned int ret;
633 if (!policy->governor || !policy->governor->store_setspeed)
634 return -EINVAL;
636 ret = sscanf(buf, "%u", &freq);
637 if (ret != 1)
638 return -EINVAL;
640 policy->governor->store_setspeed(policy, freq);
642 return count;
645 static ssize_t show_scaling_setspeed(struct cpufreq_policy *policy, char *buf)
647 if (!policy->governor || !policy->governor->show_setspeed)
648 return sprintf(buf, "<unsupported>\n");
650 return policy->governor->show_setspeed(policy, buf);
654 * show_scaling_driver - show the current cpufreq HW/BIOS limitation
656 static ssize_t show_bios_limit(struct cpufreq_policy *policy, char *buf)
658 unsigned int limit;
659 int ret;
660 if (cpufreq_driver->bios_limit) {
661 ret = cpufreq_driver->bios_limit(policy->cpu, &limit);
662 if (!ret)
663 return sprintf(buf, "%u\n", limit);
665 return sprintf(buf, "%u\n", policy->cpuinfo.max_freq);
668 cpufreq_freq_attr_ro_perm(cpuinfo_cur_freq, 0400);
669 cpufreq_freq_attr_ro(cpuinfo_min_freq);
670 cpufreq_freq_attr_ro(cpuinfo_max_freq);
671 cpufreq_freq_attr_ro(cpuinfo_transition_latency);
672 cpufreq_freq_attr_ro(scaling_available_governors);
673 cpufreq_freq_attr_ro(scaling_driver);
674 cpufreq_freq_attr_ro(scaling_cur_freq);
675 cpufreq_freq_attr_ro(bios_limit);
676 cpufreq_freq_attr_ro(related_cpus);
677 cpufreq_freq_attr_ro(affected_cpus);
678 cpufreq_freq_attr_rw(scaling_min_freq);
679 cpufreq_freq_attr_rw(scaling_max_freq);
680 cpufreq_freq_attr_rw(scaling_governor);
681 cpufreq_freq_attr_rw(scaling_setspeed);
683 static struct attribute *default_attrs[] = {
684 &cpuinfo_min_freq.attr,
685 &cpuinfo_max_freq.attr,
686 &cpuinfo_transition_latency.attr,
687 &scaling_min_freq.attr,
688 &scaling_max_freq.attr,
689 &affected_cpus.attr,
690 &related_cpus.attr,
691 &scaling_governor.attr,
692 &scaling_driver.attr,
693 &scaling_available_governors.attr,
694 &scaling_setspeed.attr,
695 NULL
698 struct kobject *cpufreq_global_kobject;
699 EXPORT_SYMBOL(cpufreq_global_kobject);
701 #define to_policy(k) container_of(k, struct cpufreq_policy, kobj)
702 #define to_attr(a) container_of(a, struct freq_attr, attr)
704 static ssize_t show(struct kobject *kobj, struct attribute *attr, char *buf)
706 struct cpufreq_policy *policy = to_policy(kobj);
707 struct freq_attr *fattr = to_attr(attr);
708 ssize_t ret = -EINVAL;
709 policy = cpufreq_cpu_get(policy->cpu);
710 if (!policy)
711 goto no_policy;
713 if (lock_policy_rwsem_read(policy->cpu) < 0)
714 goto fail;
716 if (fattr->show)
717 ret = fattr->show(policy, buf);
718 else
719 ret = -EIO;
721 unlock_policy_rwsem_read(policy->cpu);
722 fail:
723 cpufreq_cpu_put(policy);
724 no_policy:
725 return ret;
728 static ssize_t store(struct kobject *kobj, struct attribute *attr,
729 const char *buf, size_t count)
731 struct cpufreq_policy *policy = to_policy(kobj);
732 struct freq_attr *fattr = to_attr(attr);
733 ssize_t ret = -EINVAL;
734 policy = cpufreq_cpu_get(policy->cpu);
735 if (!policy)
736 goto no_policy;
738 if (lock_policy_rwsem_write(policy->cpu) < 0)
739 goto fail;
741 if (fattr->store)
742 ret = fattr->store(policy, buf, count);
743 else
744 ret = -EIO;
746 unlock_policy_rwsem_write(policy->cpu);
747 fail:
748 cpufreq_cpu_put(policy);
749 no_policy:
750 return ret;
753 static void cpufreq_sysfs_release(struct kobject *kobj)
755 struct cpufreq_policy *policy = to_policy(kobj);
756 dprintk("last reference is dropped\n");
757 complete(&policy->kobj_unregister);
760 static const struct sysfs_ops sysfs_ops = {
761 .show = show,
762 .store = store,
765 static struct kobj_type ktype_cpufreq = {
766 .sysfs_ops = &sysfs_ops,
767 .default_attrs = default_attrs,
768 .release = cpufreq_sysfs_release,
772 * Returns:
773 * Negative: Failure
774 * 0: Success
775 * Positive: When we have a managed CPU and the sysfs got symlinked
777 static int cpufreq_add_dev_policy(unsigned int cpu,
778 struct cpufreq_policy *policy,
779 struct sys_device *sys_dev)
781 int ret = 0;
782 #ifdef CONFIG_SMP
783 unsigned long flags;
784 unsigned int j;
785 #ifdef CONFIG_HOTPLUG_CPU
786 struct cpufreq_governor *gov;
788 gov = __find_governor(per_cpu(cpufreq_cpu_governor, cpu));
789 if (gov) {
790 policy->governor = gov;
791 dprintk("Restoring governor %s for cpu %d\n",
792 policy->governor->name, cpu);
794 #endif
796 for_each_cpu(j, policy->cpus) {
797 struct cpufreq_policy *managed_policy;
799 if (cpu == j)
800 continue;
802 /* Check for existing affected CPUs.
803 * They may not be aware of it due to CPU Hotplug.
804 * cpufreq_cpu_put is called when the device is removed
805 * in __cpufreq_remove_dev()
807 managed_policy = cpufreq_cpu_get(j);
808 if (unlikely(managed_policy)) {
810 /* Set proper policy_cpu */
811 unlock_policy_rwsem_write(cpu);
812 per_cpu(cpufreq_policy_cpu, cpu) = managed_policy->cpu;
814 if (lock_policy_rwsem_write(cpu) < 0) {
815 /* Should not go through policy unlock path */
816 if (cpufreq_driver->exit)
817 cpufreq_driver->exit(policy);
818 cpufreq_cpu_put(managed_policy);
819 return -EBUSY;
822 spin_lock_irqsave(&cpufreq_driver_lock, flags);
823 cpumask_copy(managed_policy->cpus, policy->cpus);
824 per_cpu(cpufreq_cpu_data, cpu) = managed_policy;
825 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
827 dprintk("CPU already managed, adding link\n");
828 ret = sysfs_create_link(&sys_dev->kobj,
829 &managed_policy->kobj,
830 "cpufreq");
831 if (ret)
832 cpufreq_cpu_put(managed_policy);
834 * Success. We only needed to be added to the mask.
835 * Call driver->exit() because only the cpu parent of
836 * the kobj needed to call init().
838 if (cpufreq_driver->exit)
839 cpufreq_driver->exit(policy);
841 if (!ret)
842 return 1;
843 else
844 return ret;
847 #endif
848 return ret;
852 /* symlink affected CPUs */
853 static int cpufreq_add_dev_symlink(unsigned int cpu,
854 struct cpufreq_policy *policy)
856 unsigned int j;
857 int ret = 0;
859 for_each_cpu(j, policy->cpus) {
860 struct cpufreq_policy *managed_policy;
861 struct sys_device *cpu_sys_dev;
863 if (j == cpu)
864 continue;
865 if (!cpu_online(j))
866 continue;
868 dprintk("CPU %u already managed, adding link\n", j);
869 managed_policy = cpufreq_cpu_get(cpu);
870 cpu_sys_dev = get_cpu_sysdev(j);
871 ret = sysfs_create_link(&cpu_sys_dev->kobj, &policy->kobj,
872 "cpufreq");
873 if (ret) {
874 cpufreq_cpu_put(managed_policy);
875 return ret;
878 return ret;
881 static int cpufreq_add_dev_interface(unsigned int cpu,
882 struct cpufreq_policy *policy,
883 struct sys_device *sys_dev)
885 struct cpufreq_policy new_policy;
886 struct freq_attr **drv_attr;
887 unsigned long flags;
888 int ret = 0;
889 unsigned int j;
891 /* prepare interface data */
892 ret = kobject_init_and_add(&policy->kobj, &ktype_cpufreq,
893 &sys_dev->kobj, "cpufreq");
894 if (ret)
895 return ret;
897 /* set up files for this cpu device */
898 drv_attr = cpufreq_driver->attr;
899 while ((drv_attr) && (*drv_attr)) {
900 ret = sysfs_create_file(&policy->kobj, &((*drv_attr)->attr));
901 if (ret)
902 goto err_out_kobj_put;
903 drv_attr++;
905 if (cpufreq_driver->get) {
906 ret = sysfs_create_file(&policy->kobj, &cpuinfo_cur_freq.attr);
907 if (ret)
908 goto err_out_kobj_put;
910 if (cpufreq_driver->target) {
911 ret = sysfs_create_file(&policy->kobj, &scaling_cur_freq.attr);
912 if (ret)
913 goto err_out_kobj_put;
915 if (cpufreq_driver->bios_limit) {
916 ret = sysfs_create_file(&policy->kobj, &bios_limit.attr);
917 if (ret)
918 goto err_out_kobj_put;
921 spin_lock_irqsave(&cpufreq_driver_lock, flags);
922 for_each_cpu(j, policy->cpus) {
923 if (!cpu_online(j))
924 continue;
925 per_cpu(cpufreq_cpu_data, j) = policy;
926 per_cpu(cpufreq_policy_cpu, j) = policy->cpu;
928 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
930 ret = cpufreq_add_dev_symlink(cpu, policy);
931 if (ret)
932 goto err_out_kobj_put;
934 memcpy(&new_policy, policy, sizeof(struct cpufreq_policy));
935 /* assure that the starting sequence is run in __cpufreq_set_policy */
936 policy->governor = NULL;
938 /* set default policy */
939 ret = __cpufreq_set_policy(policy, &new_policy);
940 policy->user_policy.policy = policy->policy;
941 policy->user_policy.governor = policy->governor;
943 if (ret) {
944 dprintk("setting policy failed\n");
945 if (cpufreq_driver->exit)
946 cpufreq_driver->exit(policy);
948 return ret;
950 err_out_kobj_put:
951 kobject_put(&policy->kobj);
952 wait_for_completion(&policy->kobj_unregister);
953 return ret;
958 * cpufreq_add_dev - add a CPU device
960 * Adds the cpufreq interface for a CPU device.
962 * The Oracle says: try running cpufreq registration/unregistration concurrently
963 * with with cpu hotplugging and all hell will break loose. Tried to clean this
964 * mess up, but more thorough testing is needed. - Mathieu
966 static int cpufreq_add_dev(struct sys_device *sys_dev)
968 unsigned int cpu = sys_dev->id;
969 int ret = 0, found = 0;
970 struct cpufreq_policy *policy;
971 unsigned long flags;
972 unsigned int j;
973 #ifdef CONFIG_HOTPLUG_CPU
974 int sibling;
975 #endif
977 if (cpu_is_offline(cpu))
978 return 0;
980 cpufreq_debug_disable_ratelimit();
981 dprintk("adding CPU %u\n", cpu);
983 #ifdef CONFIG_SMP
984 /* check whether a different CPU already registered this
985 * CPU because it is in the same boat. */
986 policy = cpufreq_cpu_get(cpu);
987 if (unlikely(policy)) {
988 cpufreq_cpu_put(policy);
989 cpufreq_debug_enable_ratelimit();
990 return 0;
992 #endif
994 if (!try_module_get(cpufreq_driver->owner)) {
995 ret = -EINVAL;
996 goto module_out;
999 ret = -ENOMEM;
1000 policy = kzalloc(sizeof(struct cpufreq_policy), GFP_KERNEL);
1001 if (!policy)
1002 goto nomem_out;
1004 if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL))
1005 goto err_free_policy;
1007 if (!zalloc_cpumask_var(&policy->related_cpus, GFP_KERNEL))
1008 goto err_free_cpumask;
1010 policy->cpu = cpu;
1011 cpumask_copy(policy->cpus, cpumask_of(cpu));
1013 /* Initially set CPU itself as the policy_cpu */
1014 per_cpu(cpufreq_policy_cpu, cpu) = cpu;
1015 ret = (lock_policy_rwsem_write(cpu) < 0);
1016 WARN_ON(ret);
1018 init_completion(&policy->kobj_unregister);
1019 INIT_WORK(&policy->update, handle_update);
1021 /* Set governor before ->init, so that driver could check it */
1022 #ifdef CONFIG_HOTPLUG_CPU
1023 for_each_online_cpu(sibling) {
1024 struct cpufreq_policy *cp = per_cpu(cpufreq_cpu_data, sibling);
1025 if (cp && cp->governor &&
1026 (cpumask_test_cpu(cpu, cp->related_cpus))) {
1027 policy->governor = cp->governor;
1028 found = 1;
1029 break;
1032 #endif
1033 if (!found)
1034 policy->governor = CPUFREQ_DEFAULT_GOVERNOR;
1035 /* call driver. From then on the cpufreq must be able
1036 * to accept all calls to ->verify and ->setpolicy for this CPU
1038 ret = cpufreq_driver->init(policy);
1039 if (ret) {
1040 dprintk("initialization failed\n");
1041 goto err_unlock_policy;
1043 policy->user_policy.min = policy->min;
1044 policy->user_policy.max = policy->max;
1046 blocking_notifier_call_chain(&cpufreq_policy_notifier_list,
1047 CPUFREQ_START, policy);
1049 ret = cpufreq_add_dev_policy(cpu, policy, sys_dev);
1050 if (ret) {
1051 if (ret > 0)
1052 /* This is a managed cpu, symlink created,
1053 exit with 0 */
1054 ret = 0;
1055 goto err_unlock_policy;
1058 ret = cpufreq_add_dev_interface(cpu, policy, sys_dev);
1059 if (ret)
1060 goto err_out_unregister;
1062 unlock_policy_rwsem_write(cpu);
1064 kobject_uevent(&policy->kobj, KOBJ_ADD);
1065 module_put(cpufreq_driver->owner);
1066 dprintk("initialization complete\n");
1067 cpufreq_debug_enable_ratelimit();
1069 return 0;
1072 err_out_unregister:
1073 spin_lock_irqsave(&cpufreq_driver_lock, flags);
1074 for_each_cpu(j, policy->cpus)
1075 per_cpu(cpufreq_cpu_data, j) = NULL;
1076 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1078 kobject_put(&policy->kobj);
1079 wait_for_completion(&policy->kobj_unregister);
1081 err_unlock_policy:
1082 unlock_policy_rwsem_write(cpu);
1083 free_cpumask_var(policy->related_cpus);
1084 err_free_cpumask:
1085 free_cpumask_var(policy->cpus);
1086 err_free_policy:
1087 kfree(policy);
1088 nomem_out:
1089 module_put(cpufreq_driver->owner);
1090 module_out:
1091 cpufreq_debug_enable_ratelimit();
1092 return ret;
1097 * __cpufreq_remove_dev - remove a CPU device
1099 * Removes the cpufreq interface for a CPU device.
1100 * Caller should already have policy_rwsem in write mode for this CPU.
1101 * This routine frees the rwsem before returning.
1103 static int __cpufreq_remove_dev(struct sys_device *sys_dev)
1105 unsigned int cpu = sys_dev->id;
1106 unsigned long flags;
1107 struct cpufreq_policy *data;
1108 struct kobject *kobj;
1109 struct completion *cmp;
1110 #ifdef CONFIG_SMP
1111 struct sys_device *cpu_sys_dev;
1112 unsigned int j;
1113 #endif
1115 cpufreq_debug_disable_ratelimit();
1116 dprintk("unregistering CPU %u\n", cpu);
1118 spin_lock_irqsave(&cpufreq_driver_lock, flags);
1119 data = per_cpu(cpufreq_cpu_data, cpu);
1121 if (!data) {
1122 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1123 cpufreq_debug_enable_ratelimit();
1124 unlock_policy_rwsem_write(cpu);
1125 return -EINVAL;
1127 per_cpu(cpufreq_cpu_data, cpu) = NULL;
1130 #ifdef CONFIG_SMP
1131 /* if this isn't the CPU which is the parent of the kobj, we
1132 * only need to unlink, put and exit
1134 if (unlikely(cpu != data->cpu)) {
1135 dprintk("removing link\n");
1136 cpumask_clear_cpu(cpu, data->cpus);
1137 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1138 kobj = &sys_dev->kobj;
1139 cpufreq_cpu_put(data);
1140 cpufreq_debug_enable_ratelimit();
1141 unlock_policy_rwsem_write(cpu);
1142 sysfs_remove_link(kobj, "cpufreq");
1143 return 0;
1145 #endif
1147 #ifdef CONFIG_SMP
1149 #ifdef CONFIG_HOTPLUG_CPU
1150 strncpy(per_cpu(cpufreq_cpu_governor, cpu), data->governor->name,
1151 CPUFREQ_NAME_LEN);
1152 #endif
1154 /* if we have other CPUs still registered, we need to unlink them,
1155 * or else wait_for_completion below will lock up. Clean the
1156 * per_cpu(cpufreq_cpu_data) while holding the lock, and remove
1157 * the sysfs links afterwards.
1159 if (unlikely(cpumask_weight(data->cpus) > 1)) {
1160 for_each_cpu(j, data->cpus) {
1161 if (j == cpu)
1162 continue;
1163 per_cpu(cpufreq_cpu_data, j) = NULL;
1167 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1169 if (unlikely(cpumask_weight(data->cpus) > 1)) {
1170 for_each_cpu(j, data->cpus) {
1171 if (j == cpu)
1172 continue;
1173 dprintk("removing link for cpu %u\n", j);
1174 #ifdef CONFIG_HOTPLUG_CPU
1175 strncpy(per_cpu(cpufreq_cpu_governor, j),
1176 data->governor->name, CPUFREQ_NAME_LEN);
1177 #endif
1178 cpu_sys_dev = get_cpu_sysdev(j);
1179 kobj = &cpu_sys_dev->kobj;
1180 unlock_policy_rwsem_write(cpu);
1181 sysfs_remove_link(kobj, "cpufreq");
1182 lock_policy_rwsem_write(cpu);
1183 cpufreq_cpu_put(data);
1186 #else
1187 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1188 #endif
1190 if (cpufreq_driver->target)
1191 __cpufreq_governor(data, CPUFREQ_GOV_STOP);
1193 kobj = &data->kobj;
1194 cmp = &data->kobj_unregister;
1195 unlock_policy_rwsem_write(cpu);
1196 kobject_put(kobj);
1198 /* we need to make sure that the underlying kobj is actually
1199 * not referenced anymore by anybody before we proceed with
1200 * unloading.
1202 dprintk("waiting for dropping of refcount\n");
1203 wait_for_completion(cmp);
1204 dprintk("wait complete\n");
1206 lock_policy_rwsem_write(cpu);
1207 if (cpufreq_driver->exit)
1208 cpufreq_driver->exit(data);
1209 unlock_policy_rwsem_write(cpu);
1211 cpufreq_debug_enable_ratelimit();
1213 #ifdef CONFIG_HOTPLUG_CPU
1214 /* when the CPU which is the parent of the kobj is hotplugged
1215 * offline, check for siblings, and create cpufreq sysfs interface
1216 * and symlinks
1218 if (unlikely(cpumask_weight(data->cpus) > 1)) {
1219 /* first sibling now owns the new sysfs dir */
1220 cpumask_clear_cpu(cpu, data->cpus);
1221 cpufreq_add_dev(get_cpu_sysdev(cpumask_first(data->cpus)));
1223 /* finally remove our own symlink */
1224 lock_policy_rwsem_write(cpu);
1225 __cpufreq_remove_dev(sys_dev);
1227 #endif
1229 free_cpumask_var(data->related_cpus);
1230 free_cpumask_var(data->cpus);
1231 kfree(data);
1233 return 0;
1237 static int cpufreq_remove_dev(struct sys_device *sys_dev)
1239 unsigned int cpu = sys_dev->id;
1240 int retval;
1242 if (cpu_is_offline(cpu))
1243 return 0;
1245 if (unlikely(lock_policy_rwsem_write(cpu)))
1246 BUG();
1248 retval = __cpufreq_remove_dev(sys_dev);
1249 return retval;
1253 static void handle_update(struct work_struct *work)
1255 struct cpufreq_policy *policy =
1256 container_of(work, struct cpufreq_policy, update);
1257 unsigned int cpu = policy->cpu;
1258 dprintk("handle_update for cpu %u called\n", cpu);
1259 cpufreq_update_policy(cpu);
1263 * cpufreq_out_of_sync - If actual and saved CPU frequency differs, we're in deep trouble.
1264 * @cpu: cpu number
1265 * @old_freq: CPU frequency the kernel thinks the CPU runs at
1266 * @new_freq: CPU frequency the CPU actually runs at
1268 * We adjust to current frequency first, and need to clean up later.
1269 * So either call to cpufreq_update_policy() or schedule handle_update()).
1271 static void cpufreq_out_of_sync(unsigned int cpu, unsigned int old_freq,
1272 unsigned int new_freq)
1274 struct cpufreq_freqs freqs;
1276 dprintk("Warning: CPU frequency out of sync: cpufreq and timing "
1277 "core thinks of %u, is %u kHz.\n", old_freq, new_freq);
1279 freqs.cpu = cpu;
1280 freqs.old = old_freq;
1281 freqs.new = new_freq;
1282 cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE);
1283 cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
1288 * cpufreq_quick_get - get the CPU frequency (in kHz) from policy->cur
1289 * @cpu: CPU number
1291 * This is the last known freq, without actually getting it from the driver.
1292 * Return value will be same as what is shown in scaling_cur_freq in sysfs.
1294 unsigned int cpufreq_quick_get(unsigned int cpu)
1296 struct cpufreq_policy *policy = cpufreq_cpu_get(cpu);
1297 unsigned int ret_freq = 0;
1299 if (policy) {
1300 ret_freq = policy->cur;
1301 cpufreq_cpu_put(policy);
1304 return ret_freq;
1306 EXPORT_SYMBOL(cpufreq_quick_get);
1309 static unsigned int __cpufreq_get(unsigned int cpu)
1311 struct cpufreq_policy *policy = per_cpu(cpufreq_cpu_data, cpu);
1312 unsigned int ret_freq = 0;
1314 if (!cpufreq_driver->get)
1315 return ret_freq;
1317 ret_freq = cpufreq_driver->get(cpu);
1319 if (ret_freq && policy->cur &&
1320 !(cpufreq_driver->flags & CPUFREQ_CONST_LOOPS)) {
1321 /* verify no discrepancy between actual and
1322 saved value exists */
1323 if (unlikely(ret_freq != policy->cur)) {
1324 cpufreq_out_of_sync(cpu, policy->cur, ret_freq);
1325 schedule_work(&policy->update);
1329 return ret_freq;
1333 * cpufreq_get - get the current CPU frequency (in kHz)
1334 * @cpu: CPU number
1336 * Get the CPU current (static) CPU frequency
1338 unsigned int cpufreq_get(unsigned int cpu)
1340 unsigned int ret_freq = 0;
1341 struct cpufreq_policy *policy = cpufreq_cpu_get(cpu);
1343 if (!policy)
1344 goto out;
1346 if (unlikely(lock_policy_rwsem_read(cpu)))
1347 goto out_policy;
1349 ret_freq = __cpufreq_get(cpu);
1351 unlock_policy_rwsem_read(cpu);
1353 out_policy:
1354 cpufreq_cpu_put(policy);
1355 out:
1356 return ret_freq;
1358 EXPORT_SYMBOL(cpufreq_get);
1360 static struct sysdev_driver cpufreq_sysdev_driver = {
1361 .add = cpufreq_add_dev,
1362 .remove = cpufreq_remove_dev,
1367 * cpufreq_bp_suspend - Prepare the boot CPU for system suspend.
1369 * This function is only executed for the boot processor. The other CPUs
1370 * have been put offline by means of CPU hotplug.
1372 static int cpufreq_bp_suspend(void)
1374 int ret = 0;
1376 int cpu = smp_processor_id();
1377 struct cpufreq_policy *cpu_policy;
1379 dprintk("suspending cpu %u\n", cpu);
1381 /* If there's no policy for the boot CPU, we have nothing to do. */
1382 cpu_policy = cpufreq_cpu_get(cpu);
1383 if (!cpu_policy)
1384 return 0;
1386 if (cpufreq_driver->suspend) {
1387 ret = cpufreq_driver->suspend(cpu_policy);
1388 if (ret)
1389 printk(KERN_ERR "cpufreq: suspend failed in ->suspend "
1390 "step on CPU %u\n", cpu_policy->cpu);
1393 cpufreq_cpu_put(cpu_policy);
1394 return ret;
1398 * cpufreq_bp_resume - Restore proper frequency handling of the boot CPU.
1400 * 1.) resume CPUfreq hardware support (cpufreq_driver->resume())
1401 * 2.) schedule call cpufreq_update_policy() ASAP as interrupts are
1402 * restored. It will verify that the current freq is in sync with
1403 * what we believe it to be. This is a bit later than when it
1404 * should be, but nonethteless it's better than calling
1405 * cpufreq_driver->get() here which might re-enable interrupts...
1407 * This function is only executed for the boot CPU. The other CPUs have not
1408 * been turned on yet.
1410 static void cpufreq_bp_resume(void)
1412 int ret = 0;
1414 int cpu = smp_processor_id();
1415 struct cpufreq_policy *cpu_policy;
1417 dprintk("resuming cpu %u\n", cpu);
1419 /* If there's no policy for the boot CPU, we have nothing to do. */
1420 cpu_policy = cpufreq_cpu_get(cpu);
1421 if (!cpu_policy)
1422 return;
1424 if (cpufreq_driver->resume) {
1425 ret = cpufreq_driver->resume(cpu_policy);
1426 if (ret) {
1427 printk(KERN_ERR "cpufreq: resume failed in ->resume "
1428 "step on CPU %u\n", cpu_policy->cpu);
1429 goto fail;
1433 schedule_work(&cpu_policy->update);
1435 fail:
1436 cpufreq_cpu_put(cpu_policy);
1439 static struct syscore_ops cpufreq_syscore_ops = {
1440 .suspend = cpufreq_bp_suspend,
1441 .resume = cpufreq_bp_resume,
1445 /*********************************************************************
1446 * NOTIFIER LISTS INTERFACE *
1447 *********************************************************************/
1450 * cpufreq_register_notifier - register a driver with cpufreq
1451 * @nb: notifier function to register
1452 * @list: CPUFREQ_TRANSITION_NOTIFIER or CPUFREQ_POLICY_NOTIFIER
1454 * Add a driver to one of two lists: either a list of drivers that
1455 * are notified about clock rate changes (once before and once after
1456 * the transition), or a list of drivers that are notified about
1457 * changes in cpufreq policy.
1459 * This function may sleep, and has the same return conditions as
1460 * blocking_notifier_chain_register.
1462 int cpufreq_register_notifier(struct notifier_block *nb, unsigned int list)
1464 int ret;
1466 WARN_ON(!init_cpufreq_transition_notifier_list_called);
1468 switch (list) {
1469 case CPUFREQ_TRANSITION_NOTIFIER:
1470 ret = srcu_notifier_chain_register(
1471 &cpufreq_transition_notifier_list, nb);
1472 break;
1473 case CPUFREQ_POLICY_NOTIFIER:
1474 ret = blocking_notifier_chain_register(
1475 &cpufreq_policy_notifier_list, nb);
1476 break;
1477 default:
1478 ret = -EINVAL;
1481 return ret;
1483 EXPORT_SYMBOL(cpufreq_register_notifier);
1487 * cpufreq_unregister_notifier - unregister a driver with cpufreq
1488 * @nb: notifier block to be unregistered
1489 * @list: CPUFREQ_TRANSITION_NOTIFIER or CPUFREQ_POLICY_NOTIFIER
1491 * Remove a driver from the CPU frequency notifier list.
1493 * This function may sleep, and has the same return conditions as
1494 * blocking_notifier_chain_unregister.
1496 int cpufreq_unregister_notifier(struct notifier_block *nb, unsigned int list)
1498 int ret;
1500 switch (list) {
1501 case CPUFREQ_TRANSITION_NOTIFIER:
1502 ret = srcu_notifier_chain_unregister(
1503 &cpufreq_transition_notifier_list, nb);
1504 break;
1505 case CPUFREQ_POLICY_NOTIFIER:
1506 ret = blocking_notifier_chain_unregister(
1507 &cpufreq_policy_notifier_list, nb);
1508 break;
1509 default:
1510 ret = -EINVAL;
1513 return ret;
1515 EXPORT_SYMBOL(cpufreq_unregister_notifier);
1518 /*********************************************************************
1519 * GOVERNORS *
1520 *********************************************************************/
1523 int __cpufreq_driver_target(struct cpufreq_policy *policy,
1524 unsigned int target_freq,
1525 unsigned int relation)
1527 int retval = -EINVAL;
1529 dprintk("target for CPU %u: %u kHz, relation %u\n", policy->cpu,
1530 target_freq, relation);
1531 if (cpu_online(policy->cpu) && cpufreq_driver->target)
1532 retval = cpufreq_driver->target(policy, target_freq, relation);
1534 return retval;
1536 EXPORT_SYMBOL_GPL(__cpufreq_driver_target);
1538 int cpufreq_driver_target(struct cpufreq_policy *policy,
1539 unsigned int target_freq,
1540 unsigned int relation)
1542 int ret = -EINVAL;
1544 policy = cpufreq_cpu_get(policy->cpu);
1545 if (!policy)
1546 goto no_policy;
1548 if (unlikely(lock_policy_rwsem_write(policy->cpu)))
1549 goto fail;
1551 ret = __cpufreq_driver_target(policy, target_freq, relation);
1553 unlock_policy_rwsem_write(policy->cpu);
1555 fail:
1556 cpufreq_cpu_put(policy);
1557 no_policy:
1558 return ret;
1560 EXPORT_SYMBOL_GPL(cpufreq_driver_target);
1562 int __cpufreq_driver_getavg(struct cpufreq_policy *policy, unsigned int cpu)
1564 int ret = 0;
1566 policy = cpufreq_cpu_get(policy->cpu);
1567 if (!policy)
1568 return -EINVAL;
1570 if (cpu_online(cpu) && cpufreq_driver->getavg)
1571 ret = cpufreq_driver->getavg(policy, cpu);
1573 cpufreq_cpu_put(policy);
1574 return ret;
1576 EXPORT_SYMBOL_GPL(__cpufreq_driver_getavg);
1579 * when "event" is CPUFREQ_GOV_LIMITS
1582 static int __cpufreq_governor(struct cpufreq_policy *policy,
1583 unsigned int event)
1585 int ret;
1587 /* Only must be defined when default governor is known to have latency
1588 restrictions, like e.g. conservative or ondemand.
1589 That this is the case is already ensured in Kconfig
1591 #ifdef CONFIG_CPU_FREQ_GOV_PERFORMANCE
1592 struct cpufreq_governor *gov = &cpufreq_gov_performance;
1593 #else
1594 struct cpufreq_governor *gov = NULL;
1595 #endif
1597 if (policy->governor->max_transition_latency &&
1598 policy->cpuinfo.transition_latency >
1599 policy->governor->max_transition_latency) {
1600 if (!gov)
1601 return -EINVAL;
1602 else {
1603 printk(KERN_WARNING "%s governor failed, too long"
1604 " transition latency of HW, fallback"
1605 " to %s governor\n",
1606 policy->governor->name,
1607 gov->name);
1608 policy->governor = gov;
1612 if (!try_module_get(policy->governor->owner))
1613 return -EINVAL;
1615 dprintk("__cpufreq_governor for CPU %u, event %u\n",
1616 policy->cpu, event);
1617 ret = policy->governor->governor(policy, event);
1619 /* we keep one module reference alive for
1620 each CPU governed by this CPU */
1621 if ((event != CPUFREQ_GOV_START) || ret)
1622 module_put(policy->governor->owner);
1623 if ((event == CPUFREQ_GOV_STOP) && !ret)
1624 module_put(policy->governor->owner);
1626 return ret;
1630 int cpufreq_register_governor(struct cpufreq_governor *governor)
1632 int err;
1634 if (!governor)
1635 return -EINVAL;
1637 mutex_lock(&cpufreq_governor_mutex);
1639 err = -EBUSY;
1640 if (__find_governor(governor->name) == NULL) {
1641 err = 0;
1642 list_add(&governor->governor_list, &cpufreq_governor_list);
1645 mutex_unlock(&cpufreq_governor_mutex);
1646 return err;
1648 EXPORT_SYMBOL_GPL(cpufreq_register_governor);
1651 void cpufreq_unregister_governor(struct cpufreq_governor *governor)
1653 #ifdef CONFIG_HOTPLUG_CPU
1654 int cpu;
1655 #endif
1657 if (!governor)
1658 return;
1660 #ifdef CONFIG_HOTPLUG_CPU
1661 for_each_present_cpu(cpu) {
1662 if (cpu_online(cpu))
1663 continue;
1664 if (!strcmp(per_cpu(cpufreq_cpu_governor, cpu), governor->name))
1665 strcpy(per_cpu(cpufreq_cpu_governor, cpu), "\0");
1667 #endif
1669 mutex_lock(&cpufreq_governor_mutex);
1670 list_del(&governor->governor_list);
1671 mutex_unlock(&cpufreq_governor_mutex);
1672 return;
1674 EXPORT_SYMBOL_GPL(cpufreq_unregister_governor);
1678 /*********************************************************************
1679 * POLICY INTERFACE *
1680 *********************************************************************/
1683 * cpufreq_get_policy - get the current cpufreq_policy
1684 * @policy: struct cpufreq_policy into which the current cpufreq_policy
1685 * is written
1687 * Reads the current cpufreq policy.
1689 int cpufreq_get_policy(struct cpufreq_policy *policy, unsigned int cpu)
1691 struct cpufreq_policy *cpu_policy;
1692 if (!policy)
1693 return -EINVAL;
1695 cpu_policy = cpufreq_cpu_get(cpu);
1696 if (!cpu_policy)
1697 return -EINVAL;
1699 memcpy(policy, cpu_policy, sizeof(struct cpufreq_policy));
1701 cpufreq_cpu_put(cpu_policy);
1702 return 0;
1704 EXPORT_SYMBOL(cpufreq_get_policy);
1708 * data : current policy.
1709 * policy : policy to be set.
1711 static int __cpufreq_set_policy(struct cpufreq_policy *data,
1712 struct cpufreq_policy *policy)
1714 int ret = 0;
1716 cpufreq_debug_disable_ratelimit();
1717 dprintk("setting new policy for CPU %u: %u - %u kHz\n", policy->cpu,
1718 policy->min, policy->max);
1720 memcpy(&policy->cpuinfo, &data->cpuinfo,
1721 sizeof(struct cpufreq_cpuinfo));
1723 if (policy->min > data->max || policy->max < data->min) {
1724 ret = -EINVAL;
1725 goto error_out;
1728 /* verify the cpu speed can be set within this limit */
1729 ret = cpufreq_driver->verify(policy);
1730 if (ret)
1731 goto error_out;
1733 /* adjust if necessary - all reasons */
1734 blocking_notifier_call_chain(&cpufreq_policy_notifier_list,
1735 CPUFREQ_ADJUST, policy);
1737 /* adjust if necessary - hardware incompatibility*/
1738 blocking_notifier_call_chain(&cpufreq_policy_notifier_list,
1739 CPUFREQ_INCOMPATIBLE, policy);
1741 /* verify the cpu speed can be set within this limit,
1742 which might be different to the first one */
1743 ret = cpufreq_driver->verify(policy);
1744 if (ret)
1745 goto error_out;
1747 /* notification of the new policy */
1748 blocking_notifier_call_chain(&cpufreq_policy_notifier_list,
1749 CPUFREQ_NOTIFY, policy);
1751 data->min = policy->min;
1752 data->max = policy->max;
1754 dprintk("new min and max freqs are %u - %u kHz\n",
1755 data->min, data->max);
1757 if (cpufreq_driver->setpolicy) {
1758 data->policy = policy->policy;
1759 dprintk("setting range\n");
1760 ret = cpufreq_driver->setpolicy(policy);
1761 } else {
1762 if (policy->governor != data->governor) {
1763 /* save old, working values */
1764 struct cpufreq_governor *old_gov = data->governor;
1766 dprintk("governor switch\n");
1768 /* end old governor */
1769 if (data->governor)
1770 __cpufreq_governor(data, CPUFREQ_GOV_STOP);
1772 /* start new governor */
1773 data->governor = policy->governor;
1774 if (__cpufreq_governor(data, CPUFREQ_GOV_START)) {
1775 /* new governor failed, so re-start old one */
1776 dprintk("starting governor %s failed\n",
1777 data->governor->name);
1778 if (old_gov) {
1779 data->governor = old_gov;
1780 __cpufreq_governor(data,
1781 CPUFREQ_GOV_START);
1783 ret = -EINVAL;
1784 goto error_out;
1786 /* might be a policy change, too, so fall through */
1788 dprintk("governor: change or update limits\n");
1789 __cpufreq_governor(data, CPUFREQ_GOV_LIMITS);
1792 error_out:
1793 cpufreq_debug_enable_ratelimit();
1794 return ret;
1798 * cpufreq_update_policy - re-evaluate an existing cpufreq policy
1799 * @cpu: CPU which shall be re-evaluated
1801 * Useful for policy notifiers which have different necessities
1802 * at different times.
1804 int cpufreq_update_policy(unsigned int cpu)
1806 struct cpufreq_policy *data = cpufreq_cpu_get(cpu);
1807 struct cpufreq_policy policy;
1808 int ret;
1810 if (!data) {
1811 ret = -ENODEV;
1812 goto no_policy;
1815 if (unlikely(lock_policy_rwsem_write(cpu))) {
1816 ret = -EINVAL;
1817 goto fail;
1820 dprintk("updating policy for CPU %u\n", cpu);
1821 memcpy(&policy, data, sizeof(struct cpufreq_policy));
1822 policy.min = data->user_policy.min;
1823 policy.max = data->user_policy.max;
1824 policy.policy = data->user_policy.policy;
1825 policy.governor = data->user_policy.governor;
1827 /* BIOS might change freq behind our back
1828 -> ask driver for current freq and notify governors about a change */
1829 if (cpufreq_driver->get) {
1830 policy.cur = cpufreq_driver->get(cpu);
1831 if (!data->cur) {
1832 dprintk("Driver did not initialize current freq");
1833 data->cur = policy.cur;
1834 } else {
1835 if (data->cur != policy.cur)
1836 cpufreq_out_of_sync(cpu, data->cur,
1837 policy.cur);
1841 ret = __cpufreq_set_policy(data, &policy);
1843 unlock_policy_rwsem_write(cpu);
1845 fail:
1846 cpufreq_cpu_put(data);
1847 no_policy:
1848 return ret;
1850 EXPORT_SYMBOL(cpufreq_update_policy);
1852 static int __cpuinit cpufreq_cpu_callback(struct notifier_block *nfb,
1853 unsigned long action, void *hcpu)
1855 unsigned int cpu = (unsigned long)hcpu;
1856 struct sys_device *sys_dev;
1858 sys_dev = get_cpu_sysdev(cpu);
1859 if (sys_dev) {
1860 switch (action) {
1861 case CPU_ONLINE:
1862 case CPU_ONLINE_FROZEN:
1863 cpufreq_add_dev(sys_dev);
1864 break;
1865 case CPU_DOWN_PREPARE:
1866 case CPU_DOWN_PREPARE_FROZEN:
1867 if (unlikely(lock_policy_rwsem_write(cpu)))
1868 BUG();
1870 __cpufreq_remove_dev(sys_dev);
1871 break;
1872 case CPU_DOWN_FAILED:
1873 case CPU_DOWN_FAILED_FROZEN:
1874 cpufreq_add_dev(sys_dev);
1875 break;
1878 return NOTIFY_OK;
1881 static struct notifier_block __refdata cpufreq_cpu_notifier = {
1882 .notifier_call = cpufreq_cpu_callback,
1885 /*********************************************************************
1886 * REGISTER / UNREGISTER CPUFREQ DRIVER *
1887 *********************************************************************/
1890 * cpufreq_register_driver - register a CPU Frequency driver
1891 * @driver_data: A struct cpufreq_driver containing the values#
1892 * submitted by the CPU Frequency driver.
1894 * Registers a CPU Frequency driver to this core code. This code
1895 * returns zero on success, -EBUSY when another driver got here first
1896 * (and isn't unregistered in the meantime).
1899 int cpufreq_register_driver(struct cpufreq_driver *driver_data)
1901 unsigned long flags;
1902 int ret;
1904 if (!driver_data || !driver_data->verify || !driver_data->init ||
1905 ((!driver_data->setpolicy) && (!driver_data->target)))
1906 return -EINVAL;
1908 dprintk("trying to register driver %s\n", driver_data->name);
1910 if (driver_data->setpolicy)
1911 driver_data->flags |= CPUFREQ_CONST_LOOPS;
1913 spin_lock_irqsave(&cpufreq_driver_lock, flags);
1914 if (cpufreq_driver) {
1915 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1916 return -EBUSY;
1918 cpufreq_driver = driver_data;
1919 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1921 ret = sysdev_driver_register(&cpu_sysdev_class,
1922 &cpufreq_sysdev_driver);
1923 if (ret)
1924 goto err_null_driver;
1926 if (!(cpufreq_driver->flags & CPUFREQ_STICKY)) {
1927 int i;
1928 ret = -ENODEV;
1930 /* check for at least one working CPU */
1931 for (i = 0; i < nr_cpu_ids; i++)
1932 if (cpu_possible(i) && per_cpu(cpufreq_cpu_data, i)) {
1933 ret = 0;
1934 break;
1937 /* if all ->init() calls failed, unregister */
1938 if (ret) {
1939 dprintk("no CPU initialized for driver %s\n",
1940 driver_data->name);
1941 goto err_sysdev_unreg;
1945 register_hotcpu_notifier(&cpufreq_cpu_notifier);
1946 dprintk("driver %s up and running\n", driver_data->name);
1947 cpufreq_debug_enable_ratelimit();
1949 return 0;
1950 err_sysdev_unreg:
1951 sysdev_driver_unregister(&cpu_sysdev_class,
1952 &cpufreq_sysdev_driver);
1953 err_null_driver:
1954 spin_lock_irqsave(&cpufreq_driver_lock, flags);
1955 cpufreq_driver = NULL;
1956 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1957 return ret;
1959 EXPORT_SYMBOL_GPL(cpufreq_register_driver);
1963 * cpufreq_unregister_driver - unregister the current CPUFreq driver
1965 * Unregister the current CPUFreq driver. Only call this if you have
1966 * the right to do so, i.e. if you have succeeded in initialising before!
1967 * Returns zero if successful, and -EINVAL if the cpufreq_driver is
1968 * currently not initialised.
1970 int cpufreq_unregister_driver(struct cpufreq_driver *driver)
1972 unsigned long flags;
1974 cpufreq_debug_disable_ratelimit();
1976 if (!cpufreq_driver || (driver != cpufreq_driver)) {
1977 cpufreq_debug_enable_ratelimit();
1978 return -EINVAL;
1981 dprintk("unregistering driver %s\n", driver->name);
1983 sysdev_driver_unregister(&cpu_sysdev_class, &cpufreq_sysdev_driver);
1984 unregister_hotcpu_notifier(&cpufreq_cpu_notifier);
1986 spin_lock_irqsave(&cpufreq_driver_lock, flags);
1987 cpufreq_driver = NULL;
1988 spin_unlock_irqrestore(&cpufreq_driver_lock, flags);
1990 return 0;
1992 EXPORT_SYMBOL_GPL(cpufreq_unregister_driver);
1994 static int __init cpufreq_core_init(void)
1996 int cpu;
1998 for_each_possible_cpu(cpu) {
1999 per_cpu(cpufreq_policy_cpu, cpu) = -1;
2000 init_rwsem(&per_cpu(cpu_policy_rwsem, cpu));
2003 cpufreq_global_kobject = kobject_create_and_add("cpufreq",
2004 &cpu_sysdev_class.kset.kobj);
2005 BUG_ON(!cpufreq_global_kobject);
2006 register_syscore_ops(&cpufreq_syscore_ops);
2008 return 0;
2010 core_initcall(cpufreq_core_init);