x86_64: move mm
[linux-2.6/mini2440.git] / kernel / sched_debug.c
blobc3ee38bd3426d91449640d0d5329c7ead4ad038e
1 /*
2 * kernel/time/sched_debug.c
4 * Print the CFS rbtree
6 * Copyright(C) 2007, Red Hat, Inc., Ingo Molnar
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
13 #include <linux/proc_fs.h>
14 #include <linux/sched.h>
15 #include <linux/seq_file.h>
16 #include <linux/kallsyms.h>
17 #include <linux/utsname.h>
20 * This allows printing both to /proc/sched_debug and
21 * to the console
23 #define SEQ_printf(m, x...) \
24 do { \
25 if (m) \
26 seq_printf(m, x); \
27 else \
28 printk(x); \
29 } while (0)
31 static void
32 print_task(struct seq_file *m, struct rq *rq, struct task_struct *p)
34 if (rq->curr == p)
35 SEQ_printf(m, "R");
36 else
37 SEQ_printf(m, " ");
39 SEQ_printf(m, "%15s %5d %15Ld %13Ld %13Ld %9Ld %5d ",
40 p->comm, p->pid,
41 (long long)p->se.fair_key,
42 (long long)(p->se.fair_key - rq->cfs.fair_clock),
43 (long long)p->se.wait_runtime,
44 (long long)(p->nvcsw + p->nivcsw),
45 p->prio);
46 #ifdef CONFIG_SCHEDSTATS
47 SEQ_printf(m, "%15Ld %15Ld %15Ld %15Ld %15Ld\n",
48 (long long)p->se.sum_exec_runtime,
49 (long long)p->se.sum_wait_runtime,
50 (long long)p->se.sum_sleep_runtime,
51 (long long)p->se.wait_runtime_overruns,
52 (long long)p->se.wait_runtime_underruns);
53 #else
54 SEQ_printf(m, "%15Ld %15Ld %15Ld %15Ld %15Ld\n",
55 0LL, 0LL, 0LL, 0LL, 0LL);
56 #endif
59 static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu)
61 struct task_struct *g, *p;
63 SEQ_printf(m,
64 "\nrunnable tasks:\n"
65 " task PID tree-key delta waiting"
66 " switches prio"
67 " sum-exec sum-wait sum-sleep"
68 " wait-overrun wait-underrun\n"
69 "------------------------------------------------------------------"
70 "----------------"
71 "------------------------------------------------"
72 "--------------------------------\n");
74 read_lock_irq(&tasklist_lock);
76 do_each_thread(g, p) {
77 if (!p->se.on_rq || task_cpu(p) != rq_cpu)
78 continue;
80 print_task(m, rq, p);
81 } while_each_thread(g, p);
83 read_unlock_irq(&tasklist_lock);
86 static void
87 print_cfs_rq_runtime_sum(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq)
89 s64 wait_runtime_rq_sum = 0;
90 struct task_struct *p;
91 struct rb_node *curr;
92 unsigned long flags;
93 struct rq *rq = &per_cpu(runqueues, cpu);
95 spin_lock_irqsave(&rq->lock, flags);
96 curr = first_fair(cfs_rq);
97 while (curr) {
98 p = rb_entry(curr, struct task_struct, se.run_node);
99 wait_runtime_rq_sum += p->se.wait_runtime;
101 curr = rb_next(curr);
103 spin_unlock_irqrestore(&rq->lock, flags);
105 SEQ_printf(m, " .%-30s: %Ld\n", "wait_runtime_rq_sum",
106 (long long)wait_runtime_rq_sum);
109 void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq)
111 SEQ_printf(m, "\ncfs_rq\n");
113 #define P(x) \
114 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(cfs_rq->x))
116 P(fair_clock);
117 P(exec_clock);
118 P(wait_runtime);
119 P(wait_runtime_overruns);
120 P(wait_runtime_underruns);
121 P(sleeper_bonus);
122 #undef P
124 print_cfs_rq_runtime_sum(m, cpu, cfs_rq);
127 static void print_cpu(struct seq_file *m, int cpu)
129 struct rq *rq = &per_cpu(runqueues, cpu);
131 #ifdef CONFIG_X86
133 unsigned int freq = cpu_khz ? : 1;
135 SEQ_printf(m, "\ncpu#%d, %u.%03u MHz\n",
136 cpu, freq / 1000, (freq % 1000));
138 #else
139 SEQ_printf(m, "\ncpu#%d\n", cpu);
140 #endif
142 #define P(x) \
143 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rq->x))
145 P(nr_running);
146 SEQ_printf(m, " .%-30s: %lu\n", "load",
147 rq->ls.load.weight);
148 P(ls.delta_fair);
149 P(ls.delta_exec);
150 P(nr_switches);
151 P(nr_load_updates);
152 P(nr_uninterruptible);
153 SEQ_printf(m, " .%-30s: %lu\n", "jiffies", jiffies);
154 P(next_balance);
155 P(curr->pid);
156 P(clock);
157 P(idle_clock);
158 P(prev_clock_raw);
159 P(clock_warps);
160 P(clock_overflows);
161 P(clock_deep_idle_events);
162 P(clock_max_delta);
163 P(cpu_load[0]);
164 P(cpu_load[1]);
165 P(cpu_load[2]);
166 P(cpu_load[3]);
167 P(cpu_load[4]);
168 #undef P
170 print_cfs_stats(m, cpu);
172 print_rq(m, rq, cpu);
175 static int sched_debug_show(struct seq_file *m, void *v)
177 u64 now = ktime_to_ns(ktime_get());
178 int cpu;
180 SEQ_printf(m, "Sched Debug Version: v0.05-v20, %s %.*s\n",
181 init_utsname()->release,
182 (int)strcspn(init_utsname()->version, " "),
183 init_utsname()->version);
185 SEQ_printf(m, "now at %Lu nsecs\n", (unsigned long long)now);
187 for_each_online_cpu(cpu)
188 print_cpu(m, cpu);
190 SEQ_printf(m, "\n");
192 return 0;
195 static void sysrq_sched_debug_show(void)
197 sched_debug_show(NULL, NULL);
200 static int sched_debug_open(struct inode *inode, struct file *filp)
202 return single_open(filp, sched_debug_show, NULL);
205 static struct file_operations sched_debug_fops = {
206 .open = sched_debug_open,
207 .read = seq_read,
208 .llseek = seq_lseek,
209 .release = single_release,
212 static int __init init_sched_debug_procfs(void)
214 struct proc_dir_entry *pe;
216 pe = create_proc_entry("sched_debug", 0644, NULL);
217 if (!pe)
218 return -ENOMEM;
220 pe->proc_fops = &sched_debug_fops;
222 return 0;
225 __initcall(init_sched_debug_procfs);
227 void proc_sched_show_task(struct task_struct *p, struct seq_file *m)
229 unsigned long flags;
230 int num_threads = 1;
232 rcu_read_lock();
233 if (lock_task_sighand(p, &flags)) {
234 num_threads = atomic_read(&p->signal->count);
235 unlock_task_sighand(p, &flags);
237 rcu_read_unlock();
239 SEQ_printf(m, "%s (%d, #threads: %d)\n", p->comm, p->pid, num_threads);
240 SEQ_printf(m, "----------------------------------------------\n");
241 #define P(F) \
242 SEQ_printf(m, "%-25s:%20Ld\n", #F, (long long)p->F)
244 P(se.wait_runtime);
245 P(se.wait_start_fair);
246 P(se.exec_start);
247 P(se.sleep_start_fair);
248 P(se.sum_exec_runtime);
250 #ifdef CONFIG_SCHEDSTATS
251 P(se.wait_start);
252 P(se.sleep_start);
253 P(se.block_start);
254 P(se.sleep_max);
255 P(se.block_max);
256 P(se.exec_max);
257 P(se.wait_max);
258 P(se.wait_runtime_overruns);
259 P(se.wait_runtime_underruns);
260 P(se.sum_wait_runtime);
261 #endif
262 SEQ_printf(m, "%-25s:%20Ld\n",
263 "nr_switches", (long long)(p->nvcsw + p->nivcsw));
264 P(se.load.weight);
265 P(policy);
266 P(prio);
267 #undef P
270 u64 t0, t1;
272 t0 = sched_clock();
273 t1 = sched_clock();
274 SEQ_printf(m, "%-25s:%20Ld\n",
275 "clock-delta", (long long)(t1-t0));
279 void proc_sched_set_task(struct task_struct *p)
281 #ifdef CONFIG_SCHEDSTATS
282 p->se.sleep_max = p->se.block_max = p->se.exec_max = p->se.wait_max = 0;
283 p->se.wait_runtime_overruns = p->se.wait_runtime_underruns = 0;
284 #endif
285 p->se.sum_exec_runtime = 0;
286 p->se.prev_sum_exec_runtime = 0;