]> git.karo-electronics.de Git - karo-tx-linux.git/blob - arch/sparc/kernel/sysfs.c
Merge remote-tracking branch 'userns/for-next'
[karo-tx-linux.git] / arch / sparc / kernel / sysfs.c
1 /* sysfs.c: Toplogy sysfs support code for sparc64.
2  *
3  * Copyright (C) 2007 David S. Miller <davem@davemloft.net>
4  */
5 #include <linux/sched.h>
6 #include <linux/device.h>
7 #include <linux/cpu.h>
8 #include <linux/smp.h>
9 #include <linux/percpu.h>
10 #include <linux/init.h>
11
12 #include <asm/cpudata.h>
13 #include <asm/hypervisor.h>
14 #include <asm/spitfire.h>
15
16 static DEFINE_PER_CPU(struct hv_mmu_statistics, mmu_stats) __attribute__((aligned(64)));
17
18 #define SHOW_MMUSTAT_ULONG(NAME) \
19 static ssize_t show_##NAME(struct device *dev, \
20                         struct device_attribute *attr, char *buf) \
21 { \
22         struct hv_mmu_statistics *p = &per_cpu(mmu_stats, dev->id); \
23         return sprintf(buf, "%lu\n", p->NAME); \
24 } \
25 static DEVICE_ATTR(NAME, 0444, show_##NAME, NULL)
26
27 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_8k_tte);
28 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_8k_tte);
29 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_64k_tte);
30 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_64k_tte);
31 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_4mb_tte);
32 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_4mb_tte);
33 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctx0_256mb_tte);
34 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctx0_256mb_tte);
35 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_8k_tte);
36 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_8k_tte);
37 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_64k_tte);
38 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_64k_tte);
39 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_4mb_tte);
40 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_4mb_tte);
41 SHOW_MMUSTAT_ULONG(immu_tsb_hits_ctxnon0_256mb_tte);
42 SHOW_MMUSTAT_ULONG(immu_tsb_ticks_ctxnon0_256mb_tte);
43 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_8k_tte);
44 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_8k_tte);
45 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_64k_tte);
46 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_64k_tte);
47 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_4mb_tte);
48 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_4mb_tte);
49 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctx0_256mb_tte);
50 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctx0_256mb_tte);
51 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_8k_tte);
52 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_8k_tte);
53 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_64k_tte);
54 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_64k_tte);
55 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_4mb_tte);
56 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_4mb_tte);
57 SHOW_MMUSTAT_ULONG(dmmu_tsb_hits_ctxnon0_256mb_tte);
58 SHOW_MMUSTAT_ULONG(dmmu_tsb_ticks_ctxnon0_256mb_tte);
59
60 static struct attribute *mmu_stat_attrs[] = {
61         &dev_attr_immu_tsb_hits_ctx0_8k_tte.attr,
62         &dev_attr_immu_tsb_ticks_ctx0_8k_tte.attr,
63         &dev_attr_immu_tsb_hits_ctx0_64k_tte.attr,
64         &dev_attr_immu_tsb_ticks_ctx0_64k_tte.attr,
65         &dev_attr_immu_tsb_hits_ctx0_4mb_tte.attr,
66         &dev_attr_immu_tsb_ticks_ctx0_4mb_tte.attr,
67         &dev_attr_immu_tsb_hits_ctx0_256mb_tte.attr,
68         &dev_attr_immu_tsb_ticks_ctx0_256mb_tte.attr,
69         &dev_attr_immu_tsb_hits_ctxnon0_8k_tte.attr,
70         &dev_attr_immu_tsb_ticks_ctxnon0_8k_tte.attr,
71         &dev_attr_immu_tsb_hits_ctxnon0_64k_tte.attr,
72         &dev_attr_immu_tsb_ticks_ctxnon0_64k_tte.attr,
73         &dev_attr_immu_tsb_hits_ctxnon0_4mb_tte.attr,
74         &dev_attr_immu_tsb_ticks_ctxnon0_4mb_tte.attr,
75         &dev_attr_immu_tsb_hits_ctxnon0_256mb_tte.attr,
76         &dev_attr_immu_tsb_ticks_ctxnon0_256mb_tte.attr,
77         &dev_attr_dmmu_tsb_hits_ctx0_8k_tte.attr,
78         &dev_attr_dmmu_tsb_ticks_ctx0_8k_tte.attr,
79         &dev_attr_dmmu_tsb_hits_ctx0_64k_tte.attr,
80         &dev_attr_dmmu_tsb_ticks_ctx0_64k_tte.attr,
81         &dev_attr_dmmu_tsb_hits_ctx0_4mb_tte.attr,
82         &dev_attr_dmmu_tsb_ticks_ctx0_4mb_tte.attr,
83         &dev_attr_dmmu_tsb_hits_ctx0_256mb_tte.attr,
84         &dev_attr_dmmu_tsb_ticks_ctx0_256mb_tte.attr,
85         &dev_attr_dmmu_tsb_hits_ctxnon0_8k_tte.attr,
86         &dev_attr_dmmu_tsb_ticks_ctxnon0_8k_tte.attr,
87         &dev_attr_dmmu_tsb_hits_ctxnon0_64k_tte.attr,
88         &dev_attr_dmmu_tsb_ticks_ctxnon0_64k_tte.attr,
89         &dev_attr_dmmu_tsb_hits_ctxnon0_4mb_tte.attr,
90         &dev_attr_dmmu_tsb_ticks_ctxnon0_4mb_tte.attr,
91         &dev_attr_dmmu_tsb_hits_ctxnon0_256mb_tte.attr,
92         &dev_attr_dmmu_tsb_ticks_ctxnon0_256mb_tte.attr,
93         NULL,
94 };
95
96 static struct attribute_group mmu_stat_group = {
97         .attrs = mmu_stat_attrs,
98         .name = "mmu_stats",
99 };
100
101 /* XXX convert to rusty's on_one_cpu */
102 static unsigned long run_on_cpu(unsigned long cpu,
103                                 unsigned long (*func)(unsigned long),
104                                 unsigned long arg)
105 {
106         cpumask_t old_affinity;
107         unsigned long ret;
108
109         cpumask_copy(&old_affinity, tsk_cpus_allowed(current));
110         /* should return -EINVAL to userspace */
111         if (set_cpus_allowed_ptr(current, cpumask_of(cpu)))
112                 return 0;
113
114         ret = func(arg);
115
116         set_cpus_allowed_ptr(current, &old_affinity);
117
118         return ret;
119 }
120
121 static unsigned long read_mmustat_enable(unsigned long junk)
122 {
123         unsigned long ra = 0;
124
125         sun4v_mmustat_info(&ra);
126
127         return ra != 0;
128 }
129
130 static unsigned long write_mmustat_enable(unsigned long val)
131 {
132         unsigned long ra, orig_ra;
133
134         if (val)
135                 ra = __pa(&per_cpu(mmu_stats, smp_processor_id()));
136         else
137                 ra = 0UL;
138
139         return sun4v_mmustat_conf(ra, &orig_ra);
140 }
141
142 static ssize_t show_mmustat_enable(struct device *s,
143                                 struct device_attribute *attr, char *buf)
144 {
145         unsigned long val = run_on_cpu(s->id, read_mmustat_enable, 0);
146         return sprintf(buf, "%lx\n", val);
147 }
148
149 static ssize_t store_mmustat_enable(struct device *s,
150                         struct device_attribute *attr, const char *buf,
151                         size_t count)
152 {
153         unsigned long val, err;
154         int ret = sscanf(buf, "%ld", &val);
155
156         if (ret != 1)
157                 return -EINVAL;
158
159         err = run_on_cpu(s->id, write_mmustat_enable, val);
160         if (err)
161                 return -EIO;
162
163         return count;
164 }
165
166 static DEVICE_ATTR(mmustat_enable, 0644, show_mmustat_enable, store_mmustat_enable);
167
168 static int mmu_stats_supported;
169
170 static int register_mmu_stats(struct device *s)
171 {
172         if (!mmu_stats_supported)
173                 return 0;
174         device_create_file(s, &dev_attr_mmustat_enable);
175         return sysfs_create_group(&s->kobj, &mmu_stat_group);
176 }
177
178 #ifdef CONFIG_HOTPLUG_CPU
179 static void unregister_mmu_stats(struct device *s)
180 {
181         if (!mmu_stats_supported)
182                 return;
183         sysfs_remove_group(&s->kobj, &mmu_stat_group);
184         device_remove_file(s, &dev_attr_mmustat_enable);
185 }
186 #endif
187
188 #define SHOW_CPUDATA_ULONG_NAME(NAME, MEMBER) \
189 static ssize_t show_##NAME(struct device *dev, \
190                 struct device_attribute *attr, char *buf) \
191 { \
192         cpuinfo_sparc *c = &cpu_data(dev->id); \
193         return sprintf(buf, "%lu\n", c->MEMBER); \
194 }
195
196 #define SHOW_CPUDATA_UINT_NAME(NAME, MEMBER) \
197 static ssize_t show_##NAME(struct device *dev, \
198                 struct device_attribute *attr, char *buf) \
199 { \
200         cpuinfo_sparc *c = &cpu_data(dev->id); \
201         return sprintf(buf, "%u\n", c->MEMBER); \
202 }
203
204 SHOW_CPUDATA_ULONG_NAME(clock_tick, clock_tick);
205 SHOW_CPUDATA_UINT_NAME(l1_dcache_size, dcache_size);
206 SHOW_CPUDATA_UINT_NAME(l1_dcache_line_size, dcache_line_size);
207 SHOW_CPUDATA_UINT_NAME(l1_icache_size, icache_size);
208 SHOW_CPUDATA_UINT_NAME(l1_icache_line_size, icache_line_size);
209 SHOW_CPUDATA_UINT_NAME(l2_cache_size, ecache_size);
210 SHOW_CPUDATA_UINT_NAME(l2_cache_line_size, ecache_line_size);
211
212 static struct device_attribute cpu_core_attrs[] = {
213         __ATTR(clock_tick,          0444, show_clock_tick, NULL),
214         __ATTR(l1_dcache_size,      0444, show_l1_dcache_size, NULL),
215         __ATTR(l1_dcache_line_size, 0444, show_l1_dcache_line_size, NULL),
216         __ATTR(l1_icache_size,      0444, show_l1_icache_size, NULL),
217         __ATTR(l1_icache_line_size, 0444, show_l1_icache_line_size, NULL),
218         __ATTR(l2_cache_size,       0444, show_l2_cache_size, NULL),
219         __ATTR(l2_cache_line_size,  0444, show_l2_cache_line_size, NULL),
220 };
221
222 static DEFINE_PER_CPU(struct cpu, cpu_devices);
223
224 static void register_cpu_online(unsigned int cpu)
225 {
226         struct cpu *c = &per_cpu(cpu_devices, cpu);
227         struct device *s = &c->dev;
228         int i;
229
230         for (i = 0; i < ARRAY_SIZE(cpu_core_attrs); i++)
231                 device_create_file(s, &cpu_core_attrs[i]);
232
233         register_mmu_stats(s);
234 }
235
236 #ifdef CONFIG_HOTPLUG_CPU
237 static void unregister_cpu_online(unsigned int cpu)
238 {
239         struct cpu *c = &per_cpu(cpu_devices, cpu);
240         struct device *s = &c->dev;
241         int i;
242
243         unregister_mmu_stats(s);
244         for (i = 0; i < ARRAY_SIZE(cpu_core_attrs); i++)
245                 device_remove_file(s, &cpu_core_attrs[i]);
246 }
247 #endif
248
249 static int __cpuinit sysfs_cpu_notify(struct notifier_block *self,
250                                       unsigned long action, void *hcpu)
251 {
252         unsigned int cpu = (unsigned int)(long)hcpu;
253
254         switch (action) {
255         case CPU_ONLINE:
256         case CPU_ONLINE_FROZEN:
257                 register_cpu_online(cpu);
258                 break;
259 #ifdef CONFIG_HOTPLUG_CPU
260         case CPU_DEAD:
261         case CPU_DEAD_FROZEN:
262                 unregister_cpu_online(cpu);
263                 break;
264 #endif
265         }
266         return NOTIFY_OK;
267 }
268
269 static struct notifier_block __cpuinitdata sysfs_cpu_nb = {
270         .notifier_call  = sysfs_cpu_notify,
271 };
272
273 static void __init check_mmu_stats(void)
274 {
275         unsigned long dummy1, err;
276
277         if (tlb_type != hypervisor)
278                 return;
279
280         err = sun4v_mmustat_info(&dummy1);
281         if (!err)
282                 mmu_stats_supported = 1;
283 }
284
285 static void register_nodes(void)
286 {
287 #ifdef CONFIG_NUMA
288         int i;
289
290         for (i = 0; i < MAX_NUMNODES; i++)
291                 register_one_node(i);
292 #endif
293 }
294
295 static int __init topology_init(void)
296 {
297         int cpu;
298
299         register_nodes();
300
301         check_mmu_stats();
302
303         register_cpu_notifier(&sysfs_cpu_nb);
304
305         for_each_possible_cpu(cpu) {
306                 struct cpu *c = &per_cpu(cpu_devices, cpu);
307
308                 register_cpu(c, cpu);
309                 if (cpu_online(cpu))
310                         register_cpu_online(cpu);
311         }
312
313         return 0;
314 }
315
316 subsys_initcall(topology_init);