Loading...
1#include <linux/cpumask.h>
2#include <linux/fs.h>
3#include <linux/init.h>
4#include <linux/interrupt.h>
5#include <linux/kernel_stat.h>
6#include <linux/proc_fs.h>
7#include <linux/sched.h>
8#include <linux/seq_file.h>
9#include <linux/slab.h>
10#include <linux/time.h>
11#include <linux/irqnr.h>
12#include <asm/cputime.h>
13#include <linux/tick.h>
14
15#ifndef arch_irq_stat_cpu
16#define arch_irq_stat_cpu(cpu) 0
17#endif
18#ifndef arch_irq_stat
19#define arch_irq_stat() 0
20#endif
21
22#ifdef arch_idle_time
23
24static cputime64_t get_idle_time(int cpu)
25{
26 cputime64_t idle;
27
28 idle = kcpustat_cpu(cpu).cpustat[CPUTIME_IDLE];
29 if (cpu_online(cpu) && !nr_iowait_cpu(cpu))
30 idle += arch_idle_time(cpu);
31 return idle;
32}
33
34static cputime64_t get_iowait_time(int cpu)
35{
36 cputime64_t iowait;
37
38 iowait = kcpustat_cpu(cpu).cpustat[CPUTIME_IOWAIT];
39 if (cpu_online(cpu) && nr_iowait_cpu(cpu))
40 iowait += arch_idle_time(cpu);
41 return iowait;
42}
43
44#else
45
46static u64 get_idle_time(int cpu)
47{
48 u64 idle, idle_time = get_cpu_idle_time_us(cpu, NULL);
49
50 if (idle_time == -1ULL)
51 /* !NO_HZ so we can rely on cpustat.idle */
52 idle = kcpustat_cpu(cpu).cpustat[CPUTIME_IDLE];
53 else
54 idle = usecs_to_cputime64(idle_time);
55
56 return idle;
57}
58
59static u64 get_iowait_time(int cpu)
60{
61 u64 iowait, iowait_time = get_cpu_iowait_time_us(cpu, NULL);
62
63 if (iowait_time == -1ULL)
64 /* !NO_HZ so we can rely on cpustat.iowait */
65 iowait = kcpustat_cpu(cpu).cpustat[CPUTIME_IOWAIT];
66 else
67 iowait = usecs_to_cputime64(iowait_time);
68
69 return iowait;
70}
71
72#endif
73
74static int show_stat(struct seq_file *p, void *v)
75{
76 int i, j;
77 unsigned long jif;
78 u64 user, nice, system, idle, iowait, irq, softirq, steal;
79 u64 guest, guest_nice;
80 u64 sum = 0;
81 u64 sum_softirq = 0;
82 unsigned int per_softirq_sums[NR_SOFTIRQS] = {0};
83 struct timespec boottime;
84
85 user = nice = system = idle = iowait =
86 irq = softirq = steal = 0;
87 guest = guest_nice = 0;
88 getboottime(&boottime);
89 jif = boottime.tv_sec;
90
91 for_each_possible_cpu(i) {
92 user += kcpustat_cpu(i).cpustat[CPUTIME_USER];
93 nice += kcpustat_cpu(i).cpustat[CPUTIME_NICE];
94 system += kcpustat_cpu(i).cpustat[CPUTIME_SYSTEM];
95 idle += get_idle_time(i);
96 iowait += get_iowait_time(i);
97 irq += kcpustat_cpu(i).cpustat[CPUTIME_IRQ];
98 softirq += kcpustat_cpu(i).cpustat[CPUTIME_SOFTIRQ];
99 steal += kcpustat_cpu(i).cpustat[CPUTIME_STEAL];
100 guest += kcpustat_cpu(i).cpustat[CPUTIME_GUEST];
101 guest_nice += kcpustat_cpu(i).cpustat[CPUTIME_GUEST_NICE];
102 sum += kstat_cpu_irqs_sum(i);
103 sum += arch_irq_stat_cpu(i);
104
105 for (j = 0; j < NR_SOFTIRQS; j++) {
106 unsigned int softirq_stat = kstat_softirqs_cpu(j, i);
107
108 per_softirq_sums[j] += softirq_stat;
109 sum_softirq += softirq_stat;
110 }
111 }
112 sum += arch_irq_stat();
113
114 seq_puts(p, "cpu ");
115 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(user));
116 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(nice));
117 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(system));
118 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(idle));
119 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(iowait));
120 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(irq));
121 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(softirq));
122 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(steal));
123 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(guest));
124 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(guest_nice));
125 seq_putc(p, '\n');
126
127 for_each_online_cpu(i) {
128 /* Copy values here to work around gcc-2.95.3, gcc-2.96 */
129 user = kcpustat_cpu(i).cpustat[CPUTIME_USER];
130 nice = kcpustat_cpu(i).cpustat[CPUTIME_NICE];
131 system = kcpustat_cpu(i).cpustat[CPUTIME_SYSTEM];
132 idle = get_idle_time(i);
133 iowait = get_iowait_time(i);
134 irq = kcpustat_cpu(i).cpustat[CPUTIME_IRQ];
135 softirq = kcpustat_cpu(i).cpustat[CPUTIME_SOFTIRQ];
136 steal = kcpustat_cpu(i).cpustat[CPUTIME_STEAL];
137 guest = kcpustat_cpu(i).cpustat[CPUTIME_GUEST];
138 guest_nice = kcpustat_cpu(i).cpustat[CPUTIME_GUEST_NICE];
139 seq_printf(p, "cpu%d", i);
140 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(user));
141 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(nice));
142 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(system));
143 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(idle));
144 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(iowait));
145 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(irq));
146 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(softirq));
147 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(steal));
148 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(guest));
149 seq_put_decimal_ull(p, ' ', cputime64_to_clock_t(guest_nice));
150 seq_putc(p, '\n');
151 }
152 seq_printf(p, "intr %llu", (unsigned long long)sum);
153
154 /* sum again ? it could be updated? */
155 for_each_irq_nr(j)
156 seq_put_decimal_ull(p, ' ', kstat_irqs(j));
157
158 seq_printf(p,
159 "\nctxt %llu\n"
160 "btime %lu\n"
161 "processes %lu\n"
162 "procs_running %lu\n"
163 "procs_blocked %lu\n",
164 nr_context_switches(),
165 (unsigned long)jif,
166 total_forks,
167 nr_running(),
168 nr_iowait());
169
170 seq_printf(p, "softirq %llu", (unsigned long long)sum_softirq);
171
172 for (i = 0; i < NR_SOFTIRQS; i++)
173 seq_put_decimal_ull(p, ' ', per_softirq_sums[i]);
174 seq_putc(p, '\n');
175
176 return 0;
177}
178
179static int stat_open(struct inode *inode, struct file *file)
180{
181 unsigned size = 1024 + 128 * num_possible_cpus();
182 char *buf;
183 struct seq_file *m;
184 int res;
185
186 /* minimum size to display an interrupt count : 2 bytes */
187 size += 2 * nr_irqs;
188
189 /* don't ask for more than the kmalloc() max size */
190 if (size > KMALLOC_MAX_SIZE)
191 size = KMALLOC_MAX_SIZE;
192 buf = kmalloc(size, GFP_KERNEL);
193 if (!buf)
194 return -ENOMEM;
195
196 res = single_open(file, show_stat, NULL);
197 if (!res) {
198 m = file->private_data;
199 m->buf = buf;
200 m->size = ksize(buf);
201 } else
202 kfree(buf);
203 return res;
204}
205
206static const struct file_operations proc_stat_operations = {
207 .open = stat_open,
208 .read = seq_read,
209 .llseek = seq_lseek,
210 .release = single_release,
211};
212
213static int __init proc_stat_init(void)
214{
215 proc_create("stat", 0, NULL, &proc_stat_operations);
216 return 0;
217}
218module_init(proc_stat_init);
1#include <linux/cpumask.h>
2#include <linux/fs.h>
3#include <linux/init.h>
4#include <linux/interrupt.h>
5#include <linux/kernel_stat.h>
6#include <linux/proc_fs.h>
7#include <linux/sched.h>
8#include <linux/seq_file.h>
9#include <linux/slab.h>
10#include <linux/time.h>
11#include <linux/irqnr.h>
12#include <asm/cputime.h>
13
14#ifndef arch_irq_stat_cpu
15#define arch_irq_stat_cpu(cpu) 0
16#endif
17#ifndef arch_irq_stat
18#define arch_irq_stat() 0
19#endif
20#ifndef arch_idle_time
21#define arch_idle_time(cpu) 0
22#endif
23
24static int show_stat(struct seq_file *p, void *v)
25{
26 int i, j;
27 unsigned long jif;
28 cputime64_t user, nice, system, idle, iowait, irq, softirq, steal;
29 cputime64_t guest, guest_nice;
30 u64 sum = 0;
31 u64 sum_softirq = 0;
32 unsigned int per_softirq_sums[NR_SOFTIRQS] = {0};
33 struct timespec boottime;
34
35 user = nice = system = idle = iowait =
36 irq = softirq = steal = cputime64_zero;
37 guest = guest_nice = cputime64_zero;
38 getboottime(&boottime);
39 jif = boottime.tv_sec;
40
41 for_each_possible_cpu(i) {
42 user = cputime64_add(user, kstat_cpu(i).cpustat.user);
43 nice = cputime64_add(nice, kstat_cpu(i).cpustat.nice);
44 system = cputime64_add(system, kstat_cpu(i).cpustat.system);
45 idle = cputime64_add(idle, kstat_cpu(i).cpustat.idle);
46 idle = cputime64_add(idle, arch_idle_time(i));
47 iowait = cputime64_add(iowait, kstat_cpu(i).cpustat.iowait);
48 irq = cputime64_add(irq, kstat_cpu(i).cpustat.irq);
49 softirq = cputime64_add(softirq, kstat_cpu(i).cpustat.softirq);
50 steal = cputime64_add(steal, kstat_cpu(i).cpustat.steal);
51 guest = cputime64_add(guest, kstat_cpu(i).cpustat.guest);
52 guest_nice = cputime64_add(guest_nice,
53 kstat_cpu(i).cpustat.guest_nice);
54 sum += kstat_cpu_irqs_sum(i);
55 sum += arch_irq_stat_cpu(i);
56
57 for (j = 0; j < NR_SOFTIRQS; j++) {
58 unsigned int softirq_stat = kstat_softirqs_cpu(j, i);
59
60 per_softirq_sums[j] += softirq_stat;
61 sum_softirq += softirq_stat;
62 }
63 }
64 sum += arch_irq_stat();
65
66 seq_printf(p, "cpu %llu %llu %llu %llu %llu %llu %llu %llu %llu "
67 "%llu\n",
68 (unsigned long long)cputime64_to_clock_t(user),
69 (unsigned long long)cputime64_to_clock_t(nice),
70 (unsigned long long)cputime64_to_clock_t(system),
71 (unsigned long long)cputime64_to_clock_t(idle),
72 (unsigned long long)cputime64_to_clock_t(iowait),
73 (unsigned long long)cputime64_to_clock_t(irq),
74 (unsigned long long)cputime64_to_clock_t(softirq),
75 (unsigned long long)cputime64_to_clock_t(steal),
76 (unsigned long long)cputime64_to_clock_t(guest),
77 (unsigned long long)cputime64_to_clock_t(guest_nice));
78 for_each_online_cpu(i) {
79
80 /* Copy values here to work around gcc-2.95.3, gcc-2.96 */
81 user = kstat_cpu(i).cpustat.user;
82 nice = kstat_cpu(i).cpustat.nice;
83 system = kstat_cpu(i).cpustat.system;
84 idle = kstat_cpu(i).cpustat.idle;
85 idle = cputime64_add(idle, arch_idle_time(i));
86 iowait = kstat_cpu(i).cpustat.iowait;
87 irq = kstat_cpu(i).cpustat.irq;
88 softirq = kstat_cpu(i).cpustat.softirq;
89 steal = kstat_cpu(i).cpustat.steal;
90 guest = kstat_cpu(i).cpustat.guest;
91 guest_nice = kstat_cpu(i).cpustat.guest_nice;
92 seq_printf(p,
93 "cpu%d %llu %llu %llu %llu %llu %llu %llu %llu %llu "
94 "%llu\n",
95 i,
96 (unsigned long long)cputime64_to_clock_t(user),
97 (unsigned long long)cputime64_to_clock_t(nice),
98 (unsigned long long)cputime64_to_clock_t(system),
99 (unsigned long long)cputime64_to_clock_t(idle),
100 (unsigned long long)cputime64_to_clock_t(iowait),
101 (unsigned long long)cputime64_to_clock_t(irq),
102 (unsigned long long)cputime64_to_clock_t(softirq),
103 (unsigned long long)cputime64_to_clock_t(steal),
104 (unsigned long long)cputime64_to_clock_t(guest),
105 (unsigned long long)cputime64_to_clock_t(guest_nice));
106 }
107 seq_printf(p, "intr %llu", (unsigned long long)sum);
108
109 /* sum again ? it could be updated? */
110 for_each_irq_nr(j)
111 seq_printf(p, " %u", kstat_irqs(j));
112
113 seq_printf(p,
114 "\nctxt %llu\n"
115 "btime %lu\n"
116 "processes %lu\n"
117 "procs_running %lu\n"
118 "procs_blocked %lu\n",
119 nr_context_switches(),
120 (unsigned long)jif,
121 total_forks,
122 nr_running(),
123 nr_iowait());
124
125 seq_printf(p, "softirq %llu", (unsigned long long)sum_softirq);
126
127 for (i = 0; i < NR_SOFTIRQS; i++)
128 seq_printf(p, " %u", per_softirq_sums[i]);
129 seq_putc(p, '\n');
130
131 return 0;
132}
133
134static int stat_open(struct inode *inode, struct file *file)
135{
136 unsigned size = 4096 * (1 + num_possible_cpus() / 32);
137 char *buf;
138 struct seq_file *m;
139 int res;
140
141 /* don't ask for more than the kmalloc() max size */
142 if (size > KMALLOC_MAX_SIZE)
143 size = KMALLOC_MAX_SIZE;
144 buf = kmalloc(size, GFP_KERNEL);
145 if (!buf)
146 return -ENOMEM;
147
148 res = single_open(file, show_stat, NULL);
149 if (!res) {
150 m = file->private_data;
151 m->buf = buf;
152 m->size = size;
153 } else
154 kfree(buf);
155 return res;
156}
157
158static const struct file_operations proc_stat_operations = {
159 .open = stat_open,
160 .read = seq_read,
161 .llseek = seq_lseek,
162 .release = single_release,
163};
164
165static int __init proc_stat_init(void)
166{
167 proc_create("stat", 0, NULL, &proc_stat_operations);
168 return 0;
169}
170module_init(proc_stat_init);