2 * unaligned.c: Unaligned load/store trap handling with special
3 * cases for the kernel to do them more quickly.
5 * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu)
6 * Copyright (C) 1996 Jakub Jelinek (jj@sunsite.mff.cuni.cz)
10 #include <linux/kernel.h>
11 #include <linux/sched.h>
13 #include <linux/module.h>
14 #include <asm/ptrace.h>
15 #include <asm/processor.h>
16 #include <asm/system.h>
17 #include <asm/uaccess.h>
18 #include <linux/smp.h>
19 #include <linux/smp_lock.h>
20 #include <linux/perf_event.h>
23 load, /* ld, ldd, ldh, ldsh */
24 store, /* st, std, sth, stsh */
25 both, /* Swap, ldstub, etc. */
31 static inline enum direction decode_direction(unsigned int insn)
33 unsigned long tmp = (insn >> 21) & 1;
38 if(((insn>>19)&0x3f) == 15)
45 /* 8 = double-word, 4 = word, 2 = half-word */
46 static inline int decode_access_size(unsigned int insn)
48 insn = (insn >> 19) & 3;
57 printk("Impossible unaligned trap. insn=%08x\n", insn);
58 die_if_kernel("Byte sized unaligned access?!?!", current->thread.kregs);
59 return 4; /* just to keep gcc happy. */
63 /* 0x400000 = signed, 0 = unsigned */
64 static inline int decode_signedness(unsigned int insn)
66 return (insn & 0x400000);
69 static inline void maybe_flush_windows(unsigned int rs1, unsigned int rs2,
72 if(rs2 >= 16 || rs1 >= 16 || rd >= 16) {
74 __asm__ __volatile__("save %sp, -0x40, %sp\n\t"
75 "save %sp, -0x40, %sp\n\t"
76 "save %sp, -0x40, %sp\n\t"
77 "save %sp, -0x40, %sp\n\t"
78 "save %sp, -0x40, %sp\n\t"
79 "save %sp, -0x40, %sp\n\t"
80 "save %sp, -0x40, %sp\n\t"
81 "restore; restore; restore; restore;\n\t"
82 "restore; restore; restore;\n\t");
86 static inline int sign_extend_imm13(int imm)
88 return imm << 19 >> 19;
91 static inline unsigned long fetch_reg(unsigned int reg, struct pt_regs *regs)
93 struct reg_window32 *win;
96 return (!reg ? 0 : regs->u_regs[reg]);
98 /* Ho hum, the slightly complicated case. */
99 win = (struct reg_window32 *) regs->u_regs[UREG_FP];
100 return win->locals[reg - 16]; /* yes, I know what this does... */
103 static inline unsigned long safe_fetch_reg(unsigned int reg, struct pt_regs *regs)
105 struct reg_window32 __user *win;
109 return (!reg ? 0 : regs->u_regs[reg]);
111 /* Ho hum, the slightly complicated case. */
112 win = (struct reg_window32 __user *) regs->u_regs[UREG_FP];
114 if ((unsigned long)win & 3)
117 if (get_user(ret, &win->locals[reg - 16]))
123 static inline unsigned long *fetch_reg_addr(unsigned int reg, struct pt_regs *regs)
125 struct reg_window32 *win;
128 return ®s->u_regs[reg];
129 win = (struct reg_window32 *) regs->u_regs[UREG_FP];
130 return &win->locals[reg - 16];
133 static unsigned long compute_effective_address(struct pt_regs *regs,
136 unsigned int rs1 = (insn >> 14) & 0x1f;
137 unsigned int rs2 = insn & 0x1f;
138 unsigned int rd = (insn >> 25) & 0x1f;
141 maybe_flush_windows(rs1, 0, rd);
142 return (fetch_reg(rs1, regs) + sign_extend_imm13(insn));
144 maybe_flush_windows(rs1, rs2, rd);
145 return (fetch_reg(rs1, regs) + fetch_reg(rs2, regs));
149 unsigned long safe_compute_effective_address(struct pt_regs *regs,
152 unsigned int rs1 = (insn >> 14) & 0x1f;
153 unsigned int rs2 = insn & 0x1f;
154 unsigned int rd = (insn >> 25) & 0x1f;
157 maybe_flush_windows(rs1, 0, rd);
158 return (safe_fetch_reg(rs1, regs) + sign_extend_imm13(insn));
160 maybe_flush_windows(rs1, rs2, rd);
161 return (safe_fetch_reg(rs1, regs) + safe_fetch_reg(rs2, regs));
165 /* This is just to make gcc think panic does return... */
166 static void unaligned_panic(char *str)
172 extern int do_int_load(unsigned long *dest_reg, int size,
173 unsigned long *saddr, int is_signed);
174 extern int __do_int_store(unsigned long *dst_addr, int size,
175 unsigned long *src_val);
177 static int do_int_store(int reg_num, int size, unsigned long *dst_addr,
178 struct pt_regs *regs)
180 unsigned long zero[2] = { 0, 0 };
181 unsigned long *src_val;
184 src_val = fetch_reg_addr(reg_num, regs);
188 zero[1] = fetch_reg(1, regs);
190 return __do_int_store(dst_addr, size, src_val);
193 extern void smp_capture(void);
194 extern void smp_release(void);
196 static inline void advance(struct pt_regs *regs)
198 regs->pc = regs->npc;
202 static inline int floating_point_load_or_store_p(unsigned int insn)
204 return (insn >> 24) & 1;
207 static inline int ok_for_kernel(unsigned int insn)
209 return !floating_point_load_or_store_p(insn);
212 static void kernel_mna_trap_fault(struct pt_regs *regs, unsigned int insn)
214 unsigned long g2 = regs->u_regs [UREG_G2];
215 unsigned long fixup = search_extables_range(regs->pc, &g2);
218 unsigned long address = compute_effective_address(regs, insn);
219 if(address < PAGE_SIZE) {
220 printk(KERN_ALERT "Unable to handle kernel NULL pointer dereference in mna handler");
222 printk(KERN_ALERT "Unable to handle kernel paging request in mna handler");
223 printk(KERN_ALERT " at virtual address %08lx\n",address);
224 printk(KERN_ALERT "current->{mm,active_mm}->context = %08lx\n",
225 (current->mm ? current->mm->context :
226 current->active_mm->context));
227 printk(KERN_ALERT "current->{mm,active_mm}->pgd = %08lx\n",
228 (current->mm ? (unsigned long) current->mm->pgd :
229 (unsigned long) current->active_mm->pgd));
230 die_if_kernel("Oops", regs);
234 regs->npc = regs->pc + 4;
235 regs->u_regs [UREG_G2] = g2;
238 asmlinkage void kernel_unaligned_trap(struct pt_regs *regs, unsigned int insn)
240 enum direction dir = decode_direction(insn);
241 int size = decode_access_size(insn);
243 if(!ok_for_kernel(insn) || dir == both) {
244 printk("Unsupported unaligned load/store trap for kernel at <%08lx>.\n",
246 unaligned_panic("Wheee. Kernel does fpu/atomic unaligned load/store.");
248 unsigned long addr = compute_effective_address(regs, insn);
251 perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, addr);
254 err = do_int_load(fetch_reg_addr(((insn>>25)&0x1f),
256 size, (unsigned long *) addr,
257 decode_signedness(insn));
261 err = do_int_store(((insn>>25)&0x1f), size,
262 (unsigned long *) addr, regs);
265 panic("Impossible kernel unaligned trap.");
269 kernel_mna_trap_fault(regs, insn);
275 static inline int ok_for_user(struct pt_regs *regs, unsigned int insn,
279 int check = (dir == load) ? VERIFY_READ : VERIFY_WRITE;
280 int size = ((insn >> 19) & 3) == 3 ? 8 : 4;
282 if ((regs->pc | regs->npc) & 3)
285 /* Must access_ok() in all the necessary places. */
286 #define WINREG_ADDR(regnum) \
287 ((void __user *)(((unsigned long *)regs->u_regs[UREG_FP])+(regnum)))
289 reg = (insn >> 25) & 0x1f;
291 if (!access_ok(check, WINREG_ADDR(reg - 16), size))
294 reg = (insn >> 14) & 0x1f;
296 if (!access_ok(check, WINREG_ADDR(reg - 16), size))
299 if (!(insn & 0x2000)) {
302 if (!access_ok(check, WINREG_ADDR(reg - 16), size))
310 static void user_mna_trap_fault(struct pt_regs *regs, unsigned int insn)
314 info.si_signo = SIGBUS;
316 info.si_code = BUS_ADRALN;
317 info.si_addr = (void __user *)safe_compute_effective_address(regs, insn);
319 send_sig_info(SIGBUS, &info, current);
322 asmlinkage void user_unaligned_trap(struct pt_regs *regs, unsigned int insn)
326 if(!(current->thread.flags & SPARC_FLAG_UNALIGNED) ||
327 (((insn >> 30) & 3) != 3))
329 dir = decode_direction(insn);
330 if(!ok_for_user(regs, insn, dir)) {
333 int err, size = decode_access_size(insn);
336 if(floating_point_load_or_store_p(insn)) {
337 printk("User FPU load/store unaligned unsupported.\n");
341 addr = compute_effective_address(regs, insn);
342 perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, addr);
345 err = do_int_load(fetch_reg_addr(((insn>>25)&0x1f),
347 size, (unsigned long *) addr,
348 decode_signedness(insn));
352 err = do_int_store(((insn>>25)&0x1f), size,
353 (unsigned long *) addr, regs);
358 * This was supported in 2.4. However, we question
359 * the value of SWAP instruction across word boundaries.
361 printk("Unaligned SWAP unsupported.\n");
366 unaligned_panic("Impossible user unaligned trap.");
377 user_mna_trap_fault(regs, insn);