1 /* 2 * Handle unaligned accesses by emulation. 3 * 4 * This file is subject to the terms and conditions of the GNU General Public 5 * License. See the file "COPYING" in the main directory of this archive 6 * for more details. 7 * 8 * Copyright (C) 1996, 1998, 1999, 2002 by Ralf Baechle 9 * Copyright (C) 1999 Silicon Graphics, Inc. 10 * Copyright (C) 2014 Imagination Technologies Ltd. 11 * 12 * This file contains exception handler for address error exception with the 13 * special capability to execute faulting instructions in software. The 14 * handler does not try to handle the case when the program counter points 15 * to an address not aligned to a word boundary. 16 * 17 * Putting data to unaligned addresses is a bad practice even on Intel where 18 * only the performance is affected. Much worse is that such code is non- 19 * portable. Due to several programs that die on MIPS due to alignment 20 * problems I decided to implement this handler anyway though I originally 21 * didn't intend to do this at all for user code. 22 * 23 * For now I enable fixing of address errors by default to make life easier. 24 * I however intend to disable this somewhen in the future when the alignment 25 * problems with user programs have been fixed. For programmers this is the 26 * right way to go. 27 * 28 * Fixing address errors is a per process option. The option is inherited 29 * across fork(2) and execve(2) calls. If you really want to use the 30 * option in your user programs - I discourage the use of the software 31 * emulation strongly - use the following code in your userland stuff: 32 * 33 * #include <sys/sysmips.h> 34 * 35 * ... 36 * sysmips(MIPS_FIXADE, x); 37 * ... 38 * 39 * The argument x is 0 for disabling software emulation, enabled otherwise. 40 * 41 * Below a little program to play around with this feature. 42 * 43 * #include <stdio.h> 44 * #include <sys/sysmips.h> 45 * 46 * struct foo { 47 * unsigned char bar[8]; 48 * }; 49 * 50 * main(int argc, char *argv[]) 51 * { 52 * struct foo x = {0, 1, 2, 3, 4, 5, 6, 7}; 53 * unsigned int *p = (unsigned int *) (x.bar + 3); 54 * int i; 55 * 56 * if (argc > 1) 57 * sysmips(MIPS_FIXADE, atoi(argv[1])); 58 * 59 * printf("*p = %08lx\n", *p); 60 * 61 * *p = 0xdeadface; 62 * 63 * for(i = 0; i <= 7; i++) 64 * printf("%02x ", x.bar[i]); 65 * printf("\n"); 66 * } 67 * 68 * Coprocessor loads are not supported; I think this case is unimportant 69 * in the practice. 70 * 71 * TODO: Handle ndc (attempted store to doubleword in uncached memory) 72 * exception for the R6000. 73 * A store crossing a page boundary might be executed only partially. 74 * Undo the partial store in this case. 75 */ 76 #include <linux/context_tracking.h> 77 #include <linux/mm.h> 78 #include <linux/signal.h> 79 #include <linux/smp.h> 80 #include <linux/sched.h> 81 #include <linux/debugfs.h> 82 #include <linux/perf_event.h> 83 84 #include <asm/asm.h> 85 #include <asm/branch.h> 86 #include <asm/byteorder.h> 87 #include <asm/cop2.h> 88 #include <asm/debug.h> 89 #include <asm/fpu.h> 90 #include <asm/fpu_emulator.h> 91 #include <asm/inst.h> 92 #include <asm/mmu_context.h> 93 #include <linux/uaccess.h> 94 95 #define STR(x) __STR(x) 96 #define __STR(x) #x 97 98 enum { 99 UNALIGNED_ACTION_QUIET, 100 UNALIGNED_ACTION_SIGNAL, 101 UNALIGNED_ACTION_SHOW, 102 }; 103 #ifdef CONFIG_DEBUG_FS 104 static u32 unaligned_instructions; 105 static u32 unaligned_action; 106 #else 107 #define unaligned_action UNALIGNED_ACTION_QUIET 108 #endif 109 extern void show_registers(struct pt_regs *regs); 110 111 #ifdef __BIG_ENDIAN 112 #define _LoadHW(addr, value, res, type) \ 113 do { \ 114 __asm__ __volatile__ (".set\tnoat\n" \ 115 "1:\t"type##_lb("%0", "0(%2)")"\n" \ 116 "2:\t"type##_lbu("$1", "1(%2)")"\n\t"\ 117 "sll\t%0, 0x8\n\t" \ 118 "or\t%0, $1\n\t" \ 119 "li\t%1, 0\n" \ 120 "3:\t.set\tat\n\t" \ 121 ".insn\n\t" \ 122 ".section\t.fixup,\"ax\"\n\t" \ 123 "4:\tli\t%1, %3\n\t" \ 124 "j\t3b\n\t" \ 125 ".previous\n\t" \ 126 ".section\t__ex_table,\"a\"\n\t" \ 127 STR(PTR)"\t1b, 4b\n\t" \ 128 STR(PTR)"\t2b, 4b\n\t" \ 129 ".previous" \ 130 : "=&r" (value), "=r" (res) \ 131 : "r" (addr), "i" (-EFAULT)); \ 132 } while(0) 133 134 #ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 135 #define _LoadW(addr, value, res, type) \ 136 do { \ 137 __asm__ __volatile__ ( \ 138 "1:\t"type##_lwl("%0", "(%2)")"\n" \ 139 "2:\t"type##_lwr("%0", "3(%2)")"\n\t"\ 140 "li\t%1, 0\n" \ 141 "3:\n\t" \ 142 ".insn\n\t" \ 143 ".section\t.fixup,\"ax\"\n\t" \ 144 "4:\tli\t%1, %3\n\t" \ 145 "j\t3b\n\t" \ 146 ".previous\n\t" \ 147 ".section\t__ex_table,\"a\"\n\t" \ 148 STR(PTR)"\t1b, 4b\n\t" \ 149 STR(PTR)"\t2b, 4b\n\t" \ 150 ".previous" \ 151 : "=&r" (value), "=r" (res) \ 152 : "r" (addr), "i" (-EFAULT)); \ 153 } while(0) 154 155 #else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 156 /* For CPUs without lwl instruction */ 157 #define _LoadW(addr, value, res, type) \ 158 do { \ 159 __asm__ __volatile__ ( \ 160 ".set\tpush\n" \ 161 ".set\tnoat\n\t" \ 162 "1:"type##_lb("%0", "0(%2)")"\n\t" \ 163 "2:"type##_lbu("$1", "1(%2)")"\n\t" \ 164 "sll\t%0, 0x8\n\t" \ 165 "or\t%0, $1\n\t" \ 166 "3:"type##_lbu("$1", "2(%2)")"\n\t" \ 167 "sll\t%0, 0x8\n\t" \ 168 "or\t%0, $1\n\t" \ 169 "4:"type##_lbu("$1", "3(%2)")"\n\t" \ 170 "sll\t%0, 0x8\n\t" \ 171 "or\t%0, $1\n\t" \ 172 "li\t%1, 0\n" \ 173 ".set\tpop\n" \ 174 "10:\n\t" \ 175 ".insn\n\t" \ 176 ".section\t.fixup,\"ax\"\n\t" \ 177 "11:\tli\t%1, %3\n\t" \ 178 "j\t10b\n\t" \ 179 ".previous\n\t" \ 180 ".section\t__ex_table,\"a\"\n\t" \ 181 STR(PTR)"\t1b, 11b\n\t" \ 182 STR(PTR)"\t2b, 11b\n\t" \ 183 STR(PTR)"\t3b, 11b\n\t" \ 184 STR(PTR)"\t4b, 11b\n\t" \ 185 ".previous" \ 186 : "=&r" (value), "=r" (res) \ 187 : "r" (addr), "i" (-EFAULT)); \ 188 } while(0) 189 190 #endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 191 192 #define _LoadHWU(addr, value, res, type) \ 193 do { \ 194 __asm__ __volatile__ ( \ 195 ".set\tnoat\n" \ 196 "1:\t"type##_lbu("%0", "0(%2)")"\n" \ 197 "2:\t"type##_lbu("$1", "1(%2)")"\n\t"\ 198 "sll\t%0, 0x8\n\t" \ 199 "or\t%0, $1\n\t" \ 200 "li\t%1, 0\n" \ 201 "3:\n\t" \ 202 ".insn\n\t" \ 203 ".set\tat\n\t" \ 204 ".section\t.fixup,\"ax\"\n\t" \ 205 "4:\tli\t%1, %3\n\t" \ 206 "j\t3b\n\t" \ 207 ".previous\n\t" \ 208 ".section\t__ex_table,\"a\"\n\t" \ 209 STR(PTR)"\t1b, 4b\n\t" \ 210 STR(PTR)"\t2b, 4b\n\t" \ 211 ".previous" \ 212 : "=&r" (value), "=r" (res) \ 213 : "r" (addr), "i" (-EFAULT)); \ 214 } while(0) 215 216 #ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 217 #define _LoadWU(addr, value, res, type) \ 218 do { \ 219 __asm__ __volatile__ ( \ 220 "1:\t"type##_lwl("%0", "(%2)")"\n" \ 221 "2:\t"type##_lwr("%0", "3(%2)")"\n\t"\ 222 "dsll\t%0, %0, 32\n\t" \ 223 "dsrl\t%0, %0, 32\n\t" \ 224 "li\t%1, 0\n" \ 225 "3:\n\t" \ 226 ".insn\n\t" \ 227 "\t.section\t.fixup,\"ax\"\n\t" \ 228 "4:\tli\t%1, %3\n\t" \ 229 "j\t3b\n\t" \ 230 ".previous\n\t" \ 231 ".section\t__ex_table,\"a\"\n\t" \ 232 STR(PTR)"\t1b, 4b\n\t" \ 233 STR(PTR)"\t2b, 4b\n\t" \ 234 ".previous" \ 235 : "=&r" (value), "=r" (res) \ 236 : "r" (addr), "i" (-EFAULT)); \ 237 } while(0) 238 239 #define _LoadDW(addr, value, res) \ 240 do { \ 241 __asm__ __volatile__ ( \ 242 "1:\tldl\t%0, (%2)\n" \ 243 "2:\tldr\t%0, 7(%2)\n\t" \ 244 "li\t%1, 0\n" \ 245 "3:\n\t" \ 246 ".insn\n\t" \ 247 "\t.section\t.fixup,\"ax\"\n\t" \ 248 "4:\tli\t%1, %3\n\t" \ 249 "j\t3b\n\t" \ 250 ".previous\n\t" \ 251 ".section\t__ex_table,\"a\"\n\t" \ 252 STR(PTR)"\t1b, 4b\n\t" \ 253 STR(PTR)"\t2b, 4b\n\t" \ 254 ".previous" \ 255 : "=&r" (value), "=r" (res) \ 256 : "r" (addr), "i" (-EFAULT)); \ 257 } while(0) 258 259 #else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 260 /* For CPUs without lwl and ldl instructions */ 261 #define _LoadWU(addr, value, res, type) \ 262 do { \ 263 __asm__ __volatile__ ( \ 264 ".set\tpush\n\t" \ 265 ".set\tnoat\n\t" \ 266 "1:"type##_lbu("%0", "0(%2)")"\n\t" \ 267 "2:"type##_lbu("$1", "1(%2)")"\n\t" \ 268 "sll\t%0, 0x8\n\t" \ 269 "or\t%0, $1\n\t" \ 270 "3:"type##_lbu("$1", "2(%2)")"\n\t" \ 271 "sll\t%0, 0x8\n\t" \ 272 "or\t%0, $1\n\t" \ 273 "4:"type##_lbu("$1", "3(%2)")"\n\t" \ 274 "sll\t%0, 0x8\n\t" \ 275 "or\t%0, $1\n\t" \ 276 "li\t%1, 0\n" \ 277 ".set\tpop\n" \ 278 "10:\n\t" \ 279 ".insn\n\t" \ 280 ".section\t.fixup,\"ax\"\n\t" \ 281 "11:\tli\t%1, %3\n\t" \ 282 "j\t10b\n\t" \ 283 ".previous\n\t" \ 284 ".section\t__ex_table,\"a\"\n\t" \ 285 STR(PTR)"\t1b, 11b\n\t" \ 286 STR(PTR)"\t2b, 11b\n\t" \ 287 STR(PTR)"\t3b, 11b\n\t" \ 288 STR(PTR)"\t4b, 11b\n\t" \ 289 ".previous" \ 290 : "=&r" (value), "=r" (res) \ 291 : "r" (addr), "i" (-EFAULT)); \ 292 } while(0) 293 294 #define _LoadDW(addr, value, res) \ 295 do { \ 296 __asm__ __volatile__ ( \ 297 ".set\tpush\n\t" \ 298 ".set\tnoat\n\t" \ 299 "1:lb\t%0, 0(%2)\n\t" \ 300 "2:lbu\t $1, 1(%2)\n\t" \ 301 "dsll\t%0, 0x8\n\t" \ 302 "or\t%0, $1\n\t" \ 303 "3:lbu\t$1, 2(%2)\n\t" \ 304 "dsll\t%0, 0x8\n\t" \ 305 "or\t%0, $1\n\t" \ 306 "4:lbu\t$1, 3(%2)\n\t" \ 307 "dsll\t%0, 0x8\n\t" \ 308 "or\t%0, $1\n\t" \ 309 "5:lbu\t$1, 4(%2)\n\t" \ 310 "dsll\t%0, 0x8\n\t" \ 311 "or\t%0, $1\n\t" \ 312 "6:lbu\t$1, 5(%2)\n\t" \ 313 "dsll\t%0, 0x8\n\t" \ 314 "or\t%0, $1\n\t" \ 315 "7:lbu\t$1, 6(%2)\n\t" \ 316 "dsll\t%0, 0x8\n\t" \ 317 "or\t%0, $1\n\t" \ 318 "8:lbu\t$1, 7(%2)\n\t" \ 319 "dsll\t%0, 0x8\n\t" \ 320 "or\t%0, $1\n\t" \ 321 "li\t%1, 0\n" \ 322 ".set\tpop\n\t" \ 323 "10:\n\t" \ 324 ".insn\n\t" \ 325 ".section\t.fixup,\"ax\"\n\t" \ 326 "11:\tli\t%1, %3\n\t" \ 327 "j\t10b\n\t" \ 328 ".previous\n\t" \ 329 ".section\t__ex_table,\"a\"\n\t" \ 330 STR(PTR)"\t1b, 11b\n\t" \ 331 STR(PTR)"\t2b, 11b\n\t" \ 332 STR(PTR)"\t3b, 11b\n\t" \ 333 STR(PTR)"\t4b, 11b\n\t" \ 334 STR(PTR)"\t5b, 11b\n\t" \ 335 STR(PTR)"\t6b, 11b\n\t" \ 336 STR(PTR)"\t7b, 11b\n\t" \ 337 STR(PTR)"\t8b, 11b\n\t" \ 338 ".previous" \ 339 : "=&r" (value), "=r" (res) \ 340 : "r" (addr), "i" (-EFAULT)); \ 341 } while(0) 342 343 #endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 344 345 346 #define _StoreHW(addr, value, res, type) \ 347 do { \ 348 __asm__ __volatile__ ( \ 349 ".set\tnoat\n" \ 350 "1:\t"type##_sb("%1", "1(%2)")"\n" \ 351 "srl\t$1, %1, 0x8\n" \ 352 "2:\t"type##_sb("$1", "0(%2)")"\n" \ 353 ".set\tat\n\t" \ 354 "li\t%0, 0\n" \ 355 "3:\n\t" \ 356 ".insn\n\t" \ 357 ".section\t.fixup,\"ax\"\n\t" \ 358 "4:\tli\t%0, %3\n\t" \ 359 "j\t3b\n\t" \ 360 ".previous\n\t" \ 361 ".section\t__ex_table,\"a\"\n\t" \ 362 STR(PTR)"\t1b, 4b\n\t" \ 363 STR(PTR)"\t2b, 4b\n\t" \ 364 ".previous" \ 365 : "=r" (res) \ 366 : "r" (value), "r" (addr), "i" (-EFAULT));\ 367 } while(0) 368 369 #ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 370 #define _StoreW(addr, value, res, type) \ 371 do { \ 372 __asm__ __volatile__ ( \ 373 "1:\t"type##_swl("%1", "(%2)")"\n" \ 374 "2:\t"type##_swr("%1", "3(%2)")"\n\t"\ 375 "li\t%0, 0\n" \ 376 "3:\n\t" \ 377 ".insn\n\t" \ 378 ".section\t.fixup,\"ax\"\n\t" \ 379 "4:\tli\t%0, %3\n\t" \ 380 "j\t3b\n\t" \ 381 ".previous\n\t" \ 382 ".section\t__ex_table,\"a\"\n\t" \ 383 STR(PTR)"\t1b, 4b\n\t" \ 384 STR(PTR)"\t2b, 4b\n\t" \ 385 ".previous" \ 386 : "=r" (res) \ 387 : "r" (value), "r" (addr), "i" (-EFAULT)); \ 388 } while(0) 389 390 #define _StoreDW(addr, value, res) \ 391 do { \ 392 __asm__ __volatile__ ( \ 393 "1:\tsdl\t%1,(%2)\n" \ 394 "2:\tsdr\t%1, 7(%2)\n\t" \ 395 "li\t%0, 0\n" \ 396 "3:\n\t" \ 397 ".insn\n\t" \ 398 ".section\t.fixup,\"ax\"\n\t" \ 399 "4:\tli\t%0, %3\n\t" \ 400 "j\t3b\n\t" \ 401 ".previous\n\t" \ 402 ".section\t__ex_table,\"a\"\n\t" \ 403 STR(PTR)"\t1b, 4b\n\t" \ 404 STR(PTR)"\t2b, 4b\n\t" \ 405 ".previous" \ 406 : "=r" (res) \ 407 : "r" (value), "r" (addr), "i" (-EFAULT)); \ 408 } while(0) 409 410 #else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 411 #define _StoreW(addr, value, res, type) \ 412 do { \ 413 __asm__ __volatile__ ( \ 414 ".set\tpush\n\t" \ 415 ".set\tnoat\n\t" \ 416 "1:"type##_sb("%1", "3(%2)")"\n\t" \ 417 "srl\t$1, %1, 0x8\n\t" \ 418 "2:"type##_sb("$1", "2(%2)")"\n\t" \ 419 "srl\t$1, $1, 0x8\n\t" \ 420 "3:"type##_sb("$1", "1(%2)")"\n\t" \ 421 "srl\t$1, $1, 0x8\n\t" \ 422 "4:"type##_sb("$1", "0(%2)")"\n\t" \ 423 ".set\tpop\n\t" \ 424 "li\t%0, 0\n" \ 425 "10:\n\t" \ 426 ".insn\n\t" \ 427 ".section\t.fixup,\"ax\"\n\t" \ 428 "11:\tli\t%0, %3\n\t" \ 429 "j\t10b\n\t" \ 430 ".previous\n\t" \ 431 ".section\t__ex_table,\"a\"\n\t" \ 432 STR(PTR)"\t1b, 11b\n\t" \ 433 STR(PTR)"\t2b, 11b\n\t" \ 434 STR(PTR)"\t3b, 11b\n\t" \ 435 STR(PTR)"\t4b, 11b\n\t" \ 436 ".previous" \ 437 : "=&r" (res) \ 438 : "r" (value), "r" (addr), "i" (-EFAULT) \ 439 : "memory"); \ 440 } while(0) 441 442 #define _StoreDW(addr, value, res) \ 443 do { \ 444 __asm__ __volatile__ ( \ 445 ".set\tpush\n\t" \ 446 ".set\tnoat\n\t" \ 447 "1:sb\t%1, 7(%2)\n\t" \ 448 "dsrl\t$1, %1, 0x8\n\t" \ 449 "2:sb\t$1, 6(%2)\n\t" \ 450 "dsrl\t$1, $1, 0x8\n\t" \ 451 "3:sb\t$1, 5(%2)\n\t" \ 452 "dsrl\t$1, $1, 0x8\n\t" \ 453 "4:sb\t$1, 4(%2)\n\t" \ 454 "dsrl\t$1, $1, 0x8\n\t" \ 455 "5:sb\t$1, 3(%2)\n\t" \ 456 "dsrl\t$1, $1, 0x8\n\t" \ 457 "6:sb\t$1, 2(%2)\n\t" \ 458 "dsrl\t$1, $1, 0x8\n\t" \ 459 "7:sb\t$1, 1(%2)\n\t" \ 460 "dsrl\t$1, $1, 0x8\n\t" \ 461 "8:sb\t$1, 0(%2)\n\t" \ 462 "dsrl\t$1, $1, 0x8\n\t" \ 463 ".set\tpop\n\t" \ 464 "li\t%0, 0\n" \ 465 "10:\n\t" \ 466 ".insn\n\t" \ 467 ".section\t.fixup,\"ax\"\n\t" \ 468 "11:\tli\t%0, %3\n\t" \ 469 "j\t10b\n\t" \ 470 ".previous\n\t" \ 471 ".section\t__ex_table,\"a\"\n\t" \ 472 STR(PTR)"\t1b, 11b\n\t" \ 473 STR(PTR)"\t2b, 11b\n\t" \ 474 STR(PTR)"\t3b, 11b\n\t" \ 475 STR(PTR)"\t4b, 11b\n\t" \ 476 STR(PTR)"\t5b, 11b\n\t" \ 477 STR(PTR)"\t6b, 11b\n\t" \ 478 STR(PTR)"\t7b, 11b\n\t" \ 479 STR(PTR)"\t8b, 11b\n\t" \ 480 ".previous" \ 481 : "=&r" (res) \ 482 : "r" (value), "r" (addr), "i" (-EFAULT) \ 483 : "memory"); \ 484 } while(0) 485 486 #endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 487 488 #else /* __BIG_ENDIAN */ 489 490 #define _LoadHW(addr, value, res, type) \ 491 do { \ 492 __asm__ __volatile__ (".set\tnoat\n" \ 493 "1:\t"type##_lb("%0", "1(%2)")"\n" \ 494 "2:\t"type##_lbu("$1", "0(%2)")"\n\t"\ 495 "sll\t%0, 0x8\n\t" \ 496 "or\t%0, $1\n\t" \ 497 "li\t%1, 0\n" \ 498 "3:\t.set\tat\n\t" \ 499 ".insn\n\t" \ 500 ".section\t.fixup,\"ax\"\n\t" \ 501 "4:\tli\t%1, %3\n\t" \ 502 "j\t3b\n\t" \ 503 ".previous\n\t" \ 504 ".section\t__ex_table,\"a\"\n\t" \ 505 STR(PTR)"\t1b, 4b\n\t" \ 506 STR(PTR)"\t2b, 4b\n\t" \ 507 ".previous" \ 508 : "=&r" (value), "=r" (res) \ 509 : "r" (addr), "i" (-EFAULT)); \ 510 } while(0) 511 512 #ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 513 #define _LoadW(addr, value, res, type) \ 514 do { \ 515 __asm__ __volatile__ ( \ 516 "1:\t"type##_lwl("%0", "3(%2)")"\n" \ 517 "2:\t"type##_lwr("%0", "(%2)")"\n\t"\ 518 "li\t%1, 0\n" \ 519 "3:\n\t" \ 520 ".insn\n\t" \ 521 ".section\t.fixup,\"ax\"\n\t" \ 522 "4:\tli\t%1, %3\n\t" \ 523 "j\t3b\n\t" \ 524 ".previous\n\t" \ 525 ".section\t__ex_table,\"a\"\n\t" \ 526 STR(PTR)"\t1b, 4b\n\t" \ 527 STR(PTR)"\t2b, 4b\n\t" \ 528 ".previous" \ 529 : "=&r" (value), "=r" (res) \ 530 : "r" (addr), "i" (-EFAULT)); \ 531 } while(0) 532 533 #else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 534 /* For CPUs without lwl instruction */ 535 #define _LoadW(addr, value, res, type) \ 536 do { \ 537 __asm__ __volatile__ ( \ 538 ".set\tpush\n" \ 539 ".set\tnoat\n\t" \ 540 "1:"type##_lb("%0", "3(%2)")"\n\t" \ 541 "2:"type##_lbu("$1", "2(%2)")"\n\t" \ 542 "sll\t%0, 0x8\n\t" \ 543 "or\t%0, $1\n\t" \ 544 "3:"type##_lbu("$1", "1(%2)")"\n\t" \ 545 "sll\t%0, 0x8\n\t" \ 546 "or\t%0, $1\n\t" \ 547 "4:"type##_lbu("$1", "0(%2)")"\n\t" \ 548 "sll\t%0, 0x8\n\t" \ 549 "or\t%0, $1\n\t" \ 550 "li\t%1, 0\n" \ 551 ".set\tpop\n" \ 552 "10:\n\t" \ 553 ".insn\n\t" \ 554 ".section\t.fixup,\"ax\"\n\t" \ 555 "11:\tli\t%1, %3\n\t" \ 556 "j\t10b\n\t" \ 557 ".previous\n\t" \ 558 ".section\t__ex_table,\"a\"\n\t" \ 559 STR(PTR)"\t1b, 11b\n\t" \ 560 STR(PTR)"\t2b, 11b\n\t" \ 561 STR(PTR)"\t3b, 11b\n\t" \ 562 STR(PTR)"\t4b, 11b\n\t" \ 563 ".previous" \ 564 : "=&r" (value), "=r" (res) \ 565 : "r" (addr), "i" (-EFAULT)); \ 566 } while(0) 567 568 #endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 569 570 571 #define _LoadHWU(addr, value, res, type) \ 572 do { \ 573 __asm__ __volatile__ ( \ 574 ".set\tnoat\n" \ 575 "1:\t"type##_lbu("%0", "1(%2)")"\n" \ 576 "2:\t"type##_lbu("$1", "0(%2)")"\n\t"\ 577 "sll\t%0, 0x8\n\t" \ 578 "or\t%0, $1\n\t" \ 579 "li\t%1, 0\n" \ 580 "3:\n\t" \ 581 ".insn\n\t" \ 582 ".set\tat\n\t" \ 583 ".section\t.fixup,\"ax\"\n\t" \ 584 "4:\tli\t%1, %3\n\t" \ 585 "j\t3b\n\t" \ 586 ".previous\n\t" \ 587 ".section\t__ex_table,\"a\"\n\t" \ 588 STR(PTR)"\t1b, 4b\n\t" \ 589 STR(PTR)"\t2b, 4b\n\t" \ 590 ".previous" \ 591 : "=&r" (value), "=r" (res) \ 592 : "r" (addr), "i" (-EFAULT)); \ 593 } while(0) 594 595 #ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 596 #define _LoadWU(addr, value, res, type) \ 597 do { \ 598 __asm__ __volatile__ ( \ 599 "1:\t"type##_lwl("%0", "3(%2)")"\n" \ 600 "2:\t"type##_lwr("%0", "(%2)")"\n\t"\ 601 "dsll\t%0, %0, 32\n\t" \ 602 "dsrl\t%0, %0, 32\n\t" \ 603 "li\t%1, 0\n" \ 604 "3:\n\t" \ 605 ".insn\n\t" \ 606 "\t.section\t.fixup,\"ax\"\n\t" \ 607 "4:\tli\t%1, %3\n\t" \ 608 "j\t3b\n\t" \ 609 ".previous\n\t" \ 610 ".section\t__ex_table,\"a\"\n\t" \ 611 STR(PTR)"\t1b, 4b\n\t" \ 612 STR(PTR)"\t2b, 4b\n\t" \ 613 ".previous" \ 614 : "=&r" (value), "=r" (res) \ 615 : "r" (addr), "i" (-EFAULT)); \ 616 } while(0) 617 618 #define _LoadDW(addr, value, res) \ 619 do { \ 620 __asm__ __volatile__ ( \ 621 "1:\tldl\t%0, 7(%2)\n" \ 622 "2:\tldr\t%0, (%2)\n\t" \ 623 "li\t%1, 0\n" \ 624 "3:\n\t" \ 625 ".insn\n\t" \ 626 "\t.section\t.fixup,\"ax\"\n\t" \ 627 "4:\tli\t%1, %3\n\t" \ 628 "j\t3b\n\t" \ 629 ".previous\n\t" \ 630 ".section\t__ex_table,\"a\"\n\t" \ 631 STR(PTR)"\t1b, 4b\n\t" \ 632 STR(PTR)"\t2b, 4b\n\t" \ 633 ".previous" \ 634 : "=&r" (value), "=r" (res) \ 635 : "r" (addr), "i" (-EFAULT)); \ 636 } while(0) 637 638 #else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 639 /* For CPUs without lwl and ldl instructions */ 640 #define _LoadWU(addr, value, res, type) \ 641 do { \ 642 __asm__ __volatile__ ( \ 643 ".set\tpush\n\t" \ 644 ".set\tnoat\n\t" \ 645 "1:"type##_lbu("%0", "3(%2)")"\n\t" \ 646 "2:"type##_lbu("$1", "2(%2)")"\n\t" \ 647 "sll\t%0, 0x8\n\t" \ 648 "or\t%0, $1\n\t" \ 649 "3:"type##_lbu("$1", "1(%2)")"\n\t" \ 650 "sll\t%0, 0x8\n\t" \ 651 "or\t%0, $1\n\t" \ 652 "4:"type##_lbu("$1", "0(%2)")"\n\t" \ 653 "sll\t%0, 0x8\n\t" \ 654 "or\t%0, $1\n\t" \ 655 "li\t%1, 0\n" \ 656 ".set\tpop\n" \ 657 "10:\n\t" \ 658 ".insn\n\t" \ 659 ".section\t.fixup,\"ax\"\n\t" \ 660 "11:\tli\t%1, %3\n\t" \ 661 "j\t10b\n\t" \ 662 ".previous\n\t" \ 663 ".section\t__ex_table,\"a\"\n\t" \ 664 STR(PTR)"\t1b, 11b\n\t" \ 665 STR(PTR)"\t2b, 11b\n\t" \ 666 STR(PTR)"\t3b, 11b\n\t" \ 667 STR(PTR)"\t4b, 11b\n\t" \ 668 ".previous" \ 669 : "=&r" (value), "=r" (res) \ 670 : "r" (addr), "i" (-EFAULT)); \ 671 } while(0) 672 673 #define _LoadDW(addr, value, res) \ 674 do { \ 675 __asm__ __volatile__ ( \ 676 ".set\tpush\n\t" \ 677 ".set\tnoat\n\t" \ 678 "1:lb\t%0, 7(%2)\n\t" \ 679 "2:lbu\t$1, 6(%2)\n\t" \ 680 "dsll\t%0, 0x8\n\t" \ 681 "or\t%0, $1\n\t" \ 682 "3:lbu\t$1, 5(%2)\n\t" \ 683 "dsll\t%0, 0x8\n\t" \ 684 "or\t%0, $1\n\t" \ 685 "4:lbu\t$1, 4(%2)\n\t" \ 686 "dsll\t%0, 0x8\n\t" \ 687 "or\t%0, $1\n\t" \ 688 "5:lbu\t$1, 3(%2)\n\t" \ 689 "dsll\t%0, 0x8\n\t" \ 690 "or\t%0, $1\n\t" \ 691 "6:lbu\t$1, 2(%2)\n\t" \ 692 "dsll\t%0, 0x8\n\t" \ 693 "or\t%0, $1\n\t" \ 694 "7:lbu\t$1, 1(%2)\n\t" \ 695 "dsll\t%0, 0x8\n\t" \ 696 "or\t%0, $1\n\t" \ 697 "8:lbu\t$1, 0(%2)\n\t" \ 698 "dsll\t%0, 0x8\n\t" \ 699 "or\t%0, $1\n\t" \ 700 "li\t%1, 0\n" \ 701 ".set\tpop\n\t" \ 702 "10:\n\t" \ 703 ".insn\n\t" \ 704 ".section\t.fixup,\"ax\"\n\t" \ 705 "11:\tli\t%1, %3\n\t" \ 706 "j\t10b\n\t" \ 707 ".previous\n\t" \ 708 ".section\t__ex_table,\"a\"\n\t" \ 709 STR(PTR)"\t1b, 11b\n\t" \ 710 STR(PTR)"\t2b, 11b\n\t" \ 711 STR(PTR)"\t3b, 11b\n\t" \ 712 STR(PTR)"\t4b, 11b\n\t" \ 713 STR(PTR)"\t5b, 11b\n\t" \ 714 STR(PTR)"\t6b, 11b\n\t" \ 715 STR(PTR)"\t7b, 11b\n\t" \ 716 STR(PTR)"\t8b, 11b\n\t" \ 717 ".previous" \ 718 : "=&r" (value), "=r" (res) \ 719 : "r" (addr), "i" (-EFAULT)); \ 720 } while(0) 721 #endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 722 723 #define _StoreHW(addr, value, res, type) \ 724 do { \ 725 __asm__ __volatile__ ( \ 726 ".set\tnoat\n" \ 727 "1:\t"type##_sb("%1", "0(%2)")"\n" \ 728 "srl\t$1,%1, 0x8\n" \ 729 "2:\t"type##_sb("$1", "1(%2)")"\n" \ 730 ".set\tat\n\t" \ 731 "li\t%0, 0\n" \ 732 "3:\n\t" \ 733 ".insn\n\t" \ 734 ".section\t.fixup,\"ax\"\n\t" \ 735 "4:\tli\t%0, %3\n\t" \ 736 "j\t3b\n\t" \ 737 ".previous\n\t" \ 738 ".section\t__ex_table,\"a\"\n\t" \ 739 STR(PTR)"\t1b, 4b\n\t" \ 740 STR(PTR)"\t2b, 4b\n\t" \ 741 ".previous" \ 742 : "=r" (res) \ 743 : "r" (value), "r" (addr), "i" (-EFAULT));\ 744 } while(0) 745 746 #ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 747 #define _StoreW(addr, value, res, type) \ 748 do { \ 749 __asm__ __volatile__ ( \ 750 "1:\t"type##_swl("%1", "3(%2)")"\n" \ 751 "2:\t"type##_swr("%1", "(%2)")"\n\t"\ 752 "li\t%0, 0\n" \ 753 "3:\n\t" \ 754 ".insn\n\t" \ 755 ".section\t.fixup,\"ax\"\n\t" \ 756 "4:\tli\t%0, %3\n\t" \ 757 "j\t3b\n\t" \ 758 ".previous\n\t" \ 759 ".section\t__ex_table,\"a\"\n\t" \ 760 STR(PTR)"\t1b, 4b\n\t" \ 761 STR(PTR)"\t2b, 4b\n\t" \ 762 ".previous" \ 763 : "=r" (res) \ 764 : "r" (value), "r" (addr), "i" (-EFAULT)); \ 765 } while(0) 766 767 #define _StoreDW(addr, value, res) \ 768 do { \ 769 __asm__ __volatile__ ( \ 770 "1:\tsdl\t%1, 7(%2)\n" \ 771 "2:\tsdr\t%1, (%2)\n\t" \ 772 "li\t%0, 0\n" \ 773 "3:\n\t" \ 774 ".insn\n\t" \ 775 ".section\t.fixup,\"ax\"\n\t" \ 776 "4:\tli\t%0, %3\n\t" \ 777 "j\t3b\n\t" \ 778 ".previous\n\t" \ 779 ".section\t__ex_table,\"a\"\n\t" \ 780 STR(PTR)"\t1b, 4b\n\t" \ 781 STR(PTR)"\t2b, 4b\n\t" \ 782 ".previous" \ 783 : "=r" (res) \ 784 : "r" (value), "r" (addr), "i" (-EFAULT)); \ 785 } while(0) 786 787 #else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 788 /* For CPUs without swl and sdl instructions */ 789 #define _StoreW(addr, value, res, type) \ 790 do { \ 791 __asm__ __volatile__ ( \ 792 ".set\tpush\n\t" \ 793 ".set\tnoat\n\t" \ 794 "1:"type##_sb("%1", "0(%2)")"\n\t" \ 795 "srl\t$1, %1, 0x8\n\t" \ 796 "2:"type##_sb("$1", "1(%2)")"\n\t" \ 797 "srl\t$1, $1, 0x8\n\t" \ 798 "3:"type##_sb("$1", "2(%2)")"\n\t" \ 799 "srl\t$1, $1, 0x8\n\t" \ 800 "4:"type##_sb("$1", "3(%2)")"\n\t" \ 801 ".set\tpop\n\t" \ 802 "li\t%0, 0\n" \ 803 "10:\n\t" \ 804 ".insn\n\t" \ 805 ".section\t.fixup,\"ax\"\n\t" \ 806 "11:\tli\t%0, %3\n\t" \ 807 "j\t10b\n\t" \ 808 ".previous\n\t" \ 809 ".section\t__ex_table,\"a\"\n\t" \ 810 STR(PTR)"\t1b, 11b\n\t" \ 811 STR(PTR)"\t2b, 11b\n\t" \ 812 STR(PTR)"\t3b, 11b\n\t" \ 813 STR(PTR)"\t4b, 11b\n\t" \ 814 ".previous" \ 815 : "=&r" (res) \ 816 : "r" (value), "r" (addr), "i" (-EFAULT) \ 817 : "memory"); \ 818 } while(0) 819 820 #define _StoreDW(addr, value, res) \ 821 do { \ 822 __asm__ __volatile__ ( \ 823 ".set\tpush\n\t" \ 824 ".set\tnoat\n\t" \ 825 "1:sb\t%1, 0(%2)\n\t" \ 826 "dsrl\t$1, %1, 0x8\n\t" \ 827 "2:sb\t$1, 1(%2)\n\t" \ 828 "dsrl\t$1, $1, 0x8\n\t" \ 829 "3:sb\t$1, 2(%2)\n\t" \ 830 "dsrl\t$1, $1, 0x8\n\t" \ 831 "4:sb\t$1, 3(%2)\n\t" \ 832 "dsrl\t$1, $1, 0x8\n\t" \ 833 "5:sb\t$1, 4(%2)\n\t" \ 834 "dsrl\t$1, $1, 0x8\n\t" \ 835 "6:sb\t$1, 5(%2)\n\t" \ 836 "dsrl\t$1, $1, 0x8\n\t" \ 837 "7:sb\t$1, 6(%2)\n\t" \ 838 "dsrl\t$1, $1, 0x8\n\t" \ 839 "8:sb\t$1, 7(%2)\n\t" \ 840 "dsrl\t$1, $1, 0x8\n\t" \ 841 ".set\tpop\n\t" \ 842 "li\t%0, 0\n" \ 843 "10:\n\t" \ 844 ".insn\n\t" \ 845 ".section\t.fixup,\"ax\"\n\t" \ 846 "11:\tli\t%0, %3\n\t" \ 847 "j\t10b\n\t" \ 848 ".previous\n\t" \ 849 ".section\t__ex_table,\"a\"\n\t" \ 850 STR(PTR)"\t1b, 11b\n\t" \ 851 STR(PTR)"\t2b, 11b\n\t" \ 852 STR(PTR)"\t3b, 11b\n\t" \ 853 STR(PTR)"\t4b, 11b\n\t" \ 854 STR(PTR)"\t5b, 11b\n\t" \ 855 STR(PTR)"\t6b, 11b\n\t" \ 856 STR(PTR)"\t7b, 11b\n\t" \ 857 STR(PTR)"\t8b, 11b\n\t" \ 858 ".previous" \ 859 : "=&r" (res) \ 860 : "r" (value), "r" (addr), "i" (-EFAULT) \ 861 : "memory"); \ 862 } while(0) 863 864 #endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 865 #endif 866 867 #define LoadHWU(addr, value, res) _LoadHWU(addr, value, res, kernel) 868 #define LoadHWUE(addr, value, res) _LoadHWU(addr, value, res, user) 869 #define LoadWU(addr, value, res) _LoadWU(addr, value, res, kernel) 870 #define LoadWUE(addr, value, res) _LoadWU(addr, value, res, user) 871 #define LoadHW(addr, value, res) _LoadHW(addr, value, res, kernel) 872 #define LoadHWE(addr, value, res) _LoadHW(addr, value, res, user) 873 #define LoadW(addr, value, res) _LoadW(addr, value, res, kernel) 874 #define LoadWE(addr, value, res) _LoadW(addr, value, res, user) 875 #define LoadDW(addr, value, res) _LoadDW(addr, value, res) 876 877 #define StoreHW(addr, value, res) _StoreHW(addr, value, res, kernel) 878 #define StoreHWE(addr, value, res) _StoreHW(addr, value, res, user) 879 #define StoreW(addr, value, res) _StoreW(addr, value, res, kernel) 880 #define StoreWE(addr, value, res) _StoreW(addr, value, res, user) 881 #define StoreDW(addr, value, res) _StoreDW(addr, value, res) 882 883 static void emulate_load_store_insn(struct pt_regs *regs, 884 void __user *addr, unsigned int __user *pc) 885 { 886 unsigned long origpc, orig31, value; 887 union mips_instruction insn; 888 unsigned int res; 889 #ifdef CONFIG_EVA 890 mm_segment_t seg; 891 #endif 892 origpc = (unsigned long)pc; 893 orig31 = regs->regs[31]; 894 895 perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); 896 897 /* 898 * This load never faults. 899 */ 900 __get_user(insn.word, pc); 901 902 switch (insn.i_format.opcode) { 903 /* 904 * These are instructions that a compiler doesn't generate. We 905 * can assume therefore that the code is MIPS-aware and 906 * really buggy. Emulating these instructions would break the 907 * semantics anyway. 908 */ 909 case ll_op: 910 case lld_op: 911 case sc_op: 912 case scd_op: 913 914 /* 915 * For these instructions the only way to create an address 916 * error is an attempted access to kernel/supervisor address 917 * space. 918 */ 919 case ldl_op: 920 case ldr_op: 921 case lwl_op: 922 case lwr_op: 923 case sdl_op: 924 case sdr_op: 925 case swl_op: 926 case swr_op: 927 case lb_op: 928 case lbu_op: 929 case sb_op: 930 goto sigbus; 931 932 /* 933 * The remaining opcodes are the ones that are really of 934 * interest. 935 */ 936 case spec3_op: 937 if (insn.dsp_format.func == lx_op) { 938 switch (insn.dsp_format.op) { 939 case lwx_op: 940 if (!access_ok(addr, 4)) 941 goto sigbus; 942 LoadW(addr, value, res); 943 if (res) 944 goto fault; 945 compute_return_epc(regs); 946 regs->regs[insn.dsp_format.rd] = value; 947 break; 948 case lhx_op: 949 if (!access_ok(addr, 2)) 950 goto sigbus; 951 LoadHW(addr, value, res); 952 if (res) 953 goto fault; 954 compute_return_epc(regs); 955 regs->regs[insn.dsp_format.rd] = value; 956 break; 957 default: 958 goto sigill; 959 } 960 } 961 #ifdef CONFIG_EVA 962 else { 963 /* 964 * we can land here only from kernel accessing user 965 * memory, so we need to "switch" the address limit to 966 * user space, so that address check can work properly. 967 */ 968 seg = get_fs(); 969 set_fs(USER_DS); 970 switch (insn.spec3_format.func) { 971 case lhe_op: 972 if (!access_ok(addr, 2)) { 973 set_fs(seg); 974 goto sigbus; 975 } 976 LoadHWE(addr, value, res); 977 if (res) { 978 set_fs(seg); 979 goto fault; 980 } 981 compute_return_epc(regs); 982 regs->regs[insn.spec3_format.rt] = value; 983 break; 984 case lwe_op: 985 if (!access_ok(addr, 4)) { 986 set_fs(seg); 987 goto sigbus; 988 } 989 LoadWE(addr, value, res); 990 if (res) { 991 set_fs(seg); 992 goto fault; 993 } 994 compute_return_epc(regs); 995 regs->regs[insn.spec3_format.rt] = value; 996 break; 997 case lhue_op: 998 if (!access_ok(addr, 2)) { 999 set_fs(seg); 1000 goto sigbus; 1001 } 1002 LoadHWUE(addr, value, res); 1003 if (res) { 1004 set_fs(seg); 1005 goto fault; 1006 } 1007 compute_return_epc(regs); 1008 regs->regs[insn.spec3_format.rt] = value; 1009 break; 1010 case she_op: 1011 if (!access_ok(addr, 2)) { 1012 set_fs(seg); 1013 goto sigbus; 1014 } 1015 compute_return_epc(regs); 1016 value = regs->regs[insn.spec3_format.rt]; 1017 StoreHWE(addr, value, res); 1018 if (res) { 1019 set_fs(seg); 1020 goto fault; 1021 } 1022 break; 1023 case swe_op: 1024 if (!access_ok(addr, 4)) { 1025 set_fs(seg); 1026 goto sigbus; 1027 } 1028 compute_return_epc(regs); 1029 value = regs->regs[insn.spec3_format.rt]; 1030 StoreWE(addr, value, res); 1031 if (res) { 1032 set_fs(seg); 1033 goto fault; 1034 } 1035 break; 1036 default: 1037 set_fs(seg); 1038 goto sigill; 1039 } 1040 set_fs(seg); 1041 } 1042 #endif 1043 break; 1044 case lh_op: 1045 if (!access_ok(addr, 2)) 1046 goto sigbus; 1047 1048 if (IS_ENABLED(CONFIG_EVA)) { 1049 if (uaccess_kernel()) 1050 LoadHW(addr, value, res); 1051 else 1052 LoadHWE(addr, value, res); 1053 } else { 1054 LoadHW(addr, value, res); 1055 } 1056 1057 if (res) 1058 goto fault; 1059 compute_return_epc(regs); 1060 regs->regs[insn.i_format.rt] = value; 1061 break; 1062 1063 case lw_op: 1064 if (!access_ok(addr, 4)) 1065 goto sigbus; 1066 1067 if (IS_ENABLED(CONFIG_EVA)) { 1068 if (uaccess_kernel()) 1069 LoadW(addr, value, res); 1070 else 1071 LoadWE(addr, value, res); 1072 } else { 1073 LoadW(addr, value, res); 1074 } 1075 1076 if (res) 1077 goto fault; 1078 compute_return_epc(regs); 1079 regs->regs[insn.i_format.rt] = value; 1080 break; 1081 1082 case lhu_op: 1083 if (!access_ok(addr, 2)) 1084 goto sigbus; 1085 1086 if (IS_ENABLED(CONFIG_EVA)) { 1087 if (uaccess_kernel()) 1088 LoadHWU(addr, value, res); 1089 else 1090 LoadHWUE(addr, value, res); 1091 } else { 1092 LoadHWU(addr, value, res); 1093 } 1094 1095 if (res) 1096 goto fault; 1097 compute_return_epc(regs); 1098 regs->regs[insn.i_format.rt] = value; 1099 break; 1100 1101 case lwu_op: 1102 #ifdef CONFIG_64BIT 1103 /* 1104 * A 32-bit kernel might be running on a 64-bit processor. But 1105 * if we're on a 32-bit processor and an i-cache incoherency 1106 * or race makes us see a 64-bit instruction here the sdl/sdr 1107 * would blow up, so for now we don't handle unaligned 64-bit 1108 * instructions on 32-bit kernels. 1109 */ 1110 if (!access_ok(addr, 4)) 1111 goto sigbus; 1112 1113 LoadWU(addr, value, res); 1114 if (res) 1115 goto fault; 1116 compute_return_epc(regs); 1117 regs->regs[insn.i_format.rt] = value; 1118 break; 1119 #endif /* CONFIG_64BIT */ 1120 1121 /* Cannot handle 64-bit instructions in 32-bit kernel */ 1122 goto sigill; 1123 1124 case ld_op: 1125 #ifdef CONFIG_64BIT 1126 /* 1127 * A 32-bit kernel might be running on a 64-bit processor. But 1128 * if we're on a 32-bit processor and an i-cache incoherency 1129 * or race makes us see a 64-bit instruction here the sdl/sdr 1130 * would blow up, so for now we don't handle unaligned 64-bit 1131 * instructions on 32-bit kernels. 1132 */ 1133 if (!access_ok(addr, 8)) 1134 goto sigbus; 1135 1136 LoadDW(addr, value, res); 1137 if (res) 1138 goto fault; 1139 compute_return_epc(regs); 1140 regs->regs[insn.i_format.rt] = value; 1141 break; 1142 #endif /* CONFIG_64BIT */ 1143 1144 /* Cannot handle 64-bit instructions in 32-bit kernel */ 1145 goto sigill; 1146 1147 case sh_op: 1148 if (!access_ok(addr, 2)) 1149 goto sigbus; 1150 1151 compute_return_epc(regs); 1152 value = regs->regs[insn.i_format.rt]; 1153 1154 if (IS_ENABLED(CONFIG_EVA)) { 1155 if (uaccess_kernel()) 1156 StoreHW(addr, value, res); 1157 else 1158 StoreHWE(addr, value, res); 1159 } else { 1160 StoreHW(addr, value, res); 1161 } 1162 1163 if (res) 1164 goto fault; 1165 break; 1166 1167 case sw_op: 1168 if (!access_ok(addr, 4)) 1169 goto sigbus; 1170 1171 compute_return_epc(regs); 1172 value = regs->regs[insn.i_format.rt]; 1173 1174 if (IS_ENABLED(CONFIG_EVA)) { 1175 if (uaccess_kernel()) 1176 StoreW(addr, value, res); 1177 else 1178 StoreWE(addr, value, res); 1179 } else { 1180 StoreW(addr, value, res); 1181 } 1182 1183 if (res) 1184 goto fault; 1185 break; 1186 1187 case sd_op: 1188 #ifdef CONFIG_64BIT 1189 /* 1190 * A 32-bit kernel might be running on a 64-bit processor. But 1191 * if we're on a 32-bit processor and an i-cache incoherency 1192 * or race makes us see a 64-bit instruction here the sdl/sdr 1193 * would blow up, so for now we don't handle unaligned 64-bit 1194 * instructions on 32-bit kernels. 1195 */ 1196 if (!access_ok(addr, 8)) 1197 goto sigbus; 1198 1199 compute_return_epc(regs); 1200 value = regs->regs[insn.i_format.rt]; 1201 StoreDW(addr, value, res); 1202 if (res) 1203 goto fault; 1204 break; 1205 #endif /* CONFIG_64BIT */ 1206 1207 /* Cannot handle 64-bit instructions in 32-bit kernel */ 1208 goto sigill; 1209 1210 #ifdef CONFIG_MIPS_FP_SUPPORT 1211 1212 case lwc1_op: 1213 case ldc1_op: 1214 case swc1_op: 1215 case sdc1_op: 1216 case cop1x_op: { 1217 void __user *fault_addr = NULL; 1218 1219 die_if_kernel("Unaligned FP access in kernel code", regs); 1220 BUG_ON(!used_math()); 1221 1222 res = fpu_emulator_cop1Handler(regs, ¤t->thread.fpu, 1, 1223 &fault_addr); 1224 own_fpu(1); /* Restore FPU state. */ 1225 1226 /* Signal if something went wrong. */ 1227 process_fpemu_return(res, fault_addr, 0); 1228 1229 if (res == 0) 1230 break; 1231 return; 1232 } 1233 #endif /* CONFIG_MIPS_FP_SUPPORT */ 1234 1235 #ifdef CONFIG_CPU_HAS_MSA 1236 1237 case msa_op: { 1238 unsigned int wd, preempted; 1239 enum msa_2b_fmt df; 1240 union fpureg *fpr; 1241 1242 if (!cpu_has_msa) 1243 goto sigill; 1244 1245 /* 1246 * If we've reached this point then userland should have taken 1247 * the MSA disabled exception & initialised vector context at 1248 * some point in the past. 1249 */ 1250 BUG_ON(!thread_msa_context_live()); 1251 1252 df = insn.msa_mi10_format.df; 1253 wd = insn.msa_mi10_format.wd; 1254 fpr = ¤t->thread.fpu.fpr[wd]; 1255 1256 switch (insn.msa_mi10_format.func) { 1257 case msa_ld_op: 1258 if (!access_ok(addr, sizeof(*fpr))) 1259 goto sigbus; 1260 1261 do { 1262 /* 1263 * If we have live MSA context keep track of 1264 * whether we get preempted in order to avoid 1265 * the register context we load being clobbered 1266 * by the live context as it's saved during 1267 * preemption. If we don't have live context 1268 * then it can't be saved to clobber the value 1269 * we load. 1270 */ 1271 preempted = test_thread_flag(TIF_USEDMSA); 1272 1273 res = __copy_from_user_inatomic(fpr, addr, 1274 sizeof(*fpr)); 1275 if (res) 1276 goto fault; 1277 1278 /* 1279 * Update the hardware register if it is in use 1280 * by the task in this quantum, in order to 1281 * avoid having to save & restore the whole 1282 * vector context. 1283 */ 1284 preempt_disable(); 1285 if (test_thread_flag(TIF_USEDMSA)) { 1286 write_msa_wr(wd, fpr, df); 1287 preempted = 0; 1288 } 1289 preempt_enable(); 1290 } while (preempted); 1291 break; 1292 1293 case msa_st_op: 1294 if (!access_ok(addr, sizeof(*fpr))) 1295 goto sigbus; 1296 1297 /* 1298 * Update from the hardware register if it is in use by 1299 * the task in this quantum, in order to avoid having to 1300 * save & restore the whole vector context. 1301 */ 1302 preempt_disable(); 1303 if (test_thread_flag(TIF_USEDMSA)) 1304 read_msa_wr(wd, fpr, df); 1305 preempt_enable(); 1306 1307 res = __copy_to_user_inatomic(addr, fpr, sizeof(*fpr)); 1308 if (res) 1309 goto fault; 1310 break; 1311 1312 default: 1313 goto sigbus; 1314 } 1315 1316 compute_return_epc(regs); 1317 break; 1318 } 1319 #endif /* CONFIG_CPU_HAS_MSA */ 1320 1321 #ifndef CONFIG_CPU_MIPSR6 1322 /* 1323 * COP2 is available to implementor for application specific use. 1324 * It's up to applications to register a notifier chain and do 1325 * whatever they have to do, including possible sending of signals. 1326 * 1327 * This instruction has been reallocated in Release 6 1328 */ 1329 case lwc2_op: 1330 cu2_notifier_call_chain(CU2_LWC2_OP, regs); 1331 break; 1332 1333 case ldc2_op: 1334 cu2_notifier_call_chain(CU2_LDC2_OP, regs); 1335 break; 1336 1337 case swc2_op: 1338 cu2_notifier_call_chain(CU2_SWC2_OP, regs); 1339 break; 1340 1341 case sdc2_op: 1342 cu2_notifier_call_chain(CU2_SDC2_OP, regs); 1343 break; 1344 #endif 1345 default: 1346 /* 1347 * Pheeee... We encountered an yet unknown instruction or 1348 * cache coherence problem. Die sucker, die ... 1349 */ 1350 goto sigill; 1351 } 1352 1353 #ifdef CONFIG_DEBUG_FS 1354 unaligned_instructions++; 1355 #endif 1356 1357 return; 1358 1359 fault: 1360 /* roll back jump/branch */ 1361 regs->cp0_epc = origpc; 1362 regs->regs[31] = orig31; 1363 /* Did we have an exception handler installed? */ 1364 if (fixup_exception(regs)) 1365 return; 1366 1367 die_if_kernel("Unhandled kernel unaligned access", regs); 1368 force_sig(SIGSEGV, current); 1369 1370 return; 1371 1372 sigbus: 1373 die_if_kernel("Unhandled kernel unaligned access", regs); 1374 force_sig(SIGBUS, current); 1375 1376 return; 1377 1378 sigill: 1379 die_if_kernel 1380 ("Unhandled kernel unaligned access or invalid instruction", regs); 1381 force_sig(SIGILL, current); 1382 } 1383 1384 /* Recode table from 16-bit register notation to 32-bit GPR. */ 1385 const int reg16to32[] = { 16, 17, 2, 3, 4, 5, 6, 7 }; 1386 1387 /* Recode table from 16-bit STORE register notation to 32-bit GPR. */ 1388 static const int reg16to32st[] = { 0, 17, 2, 3, 4, 5, 6, 7 }; 1389 1390 static void emulate_load_store_microMIPS(struct pt_regs *regs, 1391 void __user *addr) 1392 { 1393 unsigned long value; 1394 unsigned int res; 1395 int i; 1396 unsigned int reg = 0, rvar; 1397 unsigned long orig31; 1398 u16 __user *pc16; 1399 u16 halfword; 1400 unsigned int word; 1401 unsigned long origpc, contpc; 1402 union mips_instruction insn; 1403 struct mm_decoded_insn mminsn; 1404 1405 origpc = regs->cp0_epc; 1406 orig31 = regs->regs[31]; 1407 1408 mminsn.micro_mips_mode = 1; 1409 1410 /* 1411 * This load never faults. 1412 */ 1413 pc16 = (unsigned short __user *)msk_isa16_mode(regs->cp0_epc); 1414 __get_user(halfword, pc16); 1415 pc16++; 1416 contpc = regs->cp0_epc + 2; 1417 word = ((unsigned int)halfword << 16); 1418 mminsn.pc_inc = 2; 1419 1420 if (!mm_insn_16bit(halfword)) { 1421 __get_user(halfword, pc16); 1422 pc16++; 1423 contpc = regs->cp0_epc + 4; 1424 mminsn.pc_inc = 4; 1425 word |= halfword; 1426 } 1427 mminsn.insn = word; 1428 1429 if (get_user(halfword, pc16)) 1430 goto fault; 1431 mminsn.next_pc_inc = 2; 1432 word = ((unsigned int)halfword << 16); 1433 1434 if (!mm_insn_16bit(halfword)) { 1435 pc16++; 1436 if (get_user(halfword, pc16)) 1437 goto fault; 1438 mminsn.next_pc_inc = 4; 1439 word |= halfword; 1440 } 1441 mminsn.next_insn = word; 1442 1443 insn = (union mips_instruction)(mminsn.insn); 1444 if (mm_isBranchInstr(regs, mminsn, &contpc)) 1445 insn = (union mips_instruction)(mminsn.next_insn); 1446 1447 /* Parse instruction to find what to do */ 1448 1449 switch (insn.mm_i_format.opcode) { 1450 1451 case mm_pool32a_op: 1452 switch (insn.mm_x_format.func) { 1453 case mm_lwxs_op: 1454 reg = insn.mm_x_format.rd; 1455 goto loadW; 1456 } 1457 1458 goto sigbus; 1459 1460 case mm_pool32b_op: 1461 switch (insn.mm_m_format.func) { 1462 case mm_lwp_func: 1463 reg = insn.mm_m_format.rd; 1464 if (reg == 31) 1465 goto sigbus; 1466 1467 if (!access_ok(addr, 8)) 1468 goto sigbus; 1469 1470 LoadW(addr, value, res); 1471 if (res) 1472 goto fault; 1473 regs->regs[reg] = value; 1474 addr += 4; 1475 LoadW(addr, value, res); 1476 if (res) 1477 goto fault; 1478 regs->regs[reg + 1] = value; 1479 goto success; 1480 1481 case mm_swp_func: 1482 reg = insn.mm_m_format.rd; 1483 if (reg == 31) 1484 goto sigbus; 1485 1486 if (!access_ok(addr, 8)) 1487 goto sigbus; 1488 1489 value = regs->regs[reg]; 1490 StoreW(addr, value, res); 1491 if (res) 1492 goto fault; 1493 addr += 4; 1494 value = regs->regs[reg + 1]; 1495 StoreW(addr, value, res); 1496 if (res) 1497 goto fault; 1498 goto success; 1499 1500 case mm_ldp_func: 1501 #ifdef CONFIG_64BIT 1502 reg = insn.mm_m_format.rd; 1503 if (reg == 31) 1504 goto sigbus; 1505 1506 if (!access_ok(addr, 16)) 1507 goto sigbus; 1508 1509 LoadDW(addr, value, res); 1510 if (res) 1511 goto fault; 1512 regs->regs[reg] = value; 1513 addr += 8; 1514 LoadDW(addr, value, res); 1515 if (res) 1516 goto fault; 1517 regs->regs[reg + 1] = value; 1518 goto success; 1519 #endif /* CONFIG_64BIT */ 1520 1521 goto sigill; 1522 1523 case mm_sdp_func: 1524 #ifdef CONFIG_64BIT 1525 reg = insn.mm_m_format.rd; 1526 if (reg == 31) 1527 goto sigbus; 1528 1529 if (!access_ok(addr, 16)) 1530 goto sigbus; 1531 1532 value = regs->regs[reg]; 1533 StoreDW(addr, value, res); 1534 if (res) 1535 goto fault; 1536 addr += 8; 1537 value = regs->regs[reg + 1]; 1538 StoreDW(addr, value, res); 1539 if (res) 1540 goto fault; 1541 goto success; 1542 #endif /* CONFIG_64BIT */ 1543 1544 goto sigill; 1545 1546 case mm_lwm32_func: 1547 reg = insn.mm_m_format.rd; 1548 rvar = reg & 0xf; 1549 if ((rvar > 9) || !reg) 1550 goto sigill; 1551 if (reg & 0x10) { 1552 if (!access_ok(addr, 4 * (rvar + 1))) 1553 goto sigbus; 1554 } else { 1555 if (!access_ok(addr, 4 * rvar)) 1556 goto sigbus; 1557 } 1558 if (rvar == 9) 1559 rvar = 8; 1560 for (i = 16; rvar; rvar--, i++) { 1561 LoadW(addr, value, res); 1562 if (res) 1563 goto fault; 1564 addr += 4; 1565 regs->regs[i] = value; 1566 } 1567 if ((reg & 0xf) == 9) { 1568 LoadW(addr, value, res); 1569 if (res) 1570 goto fault; 1571 addr += 4; 1572 regs->regs[30] = value; 1573 } 1574 if (reg & 0x10) { 1575 LoadW(addr, value, res); 1576 if (res) 1577 goto fault; 1578 regs->regs[31] = value; 1579 } 1580 goto success; 1581 1582 case mm_swm32_func: 1583 reg = insn.mm_m_format.rd; 1584 rvar = reg & 0xf; 1585 if ((rvar > 9) || !reg) 1586 goto sigill; 1587 if (reg & 0x10) { 1588 if (!access_ok(addr, 4 * (rvar + 1))) 1589 goto sigbus; 1590 } else { 1591 if (!access_ok(addr, 4 * rvar)) 1592 goto sigbus; 1593 } 1594 if (rvar == 9) 1595 rvar = 8; 1596 for (i = 16; rvar; rvar--, i++) { 1597 value = regs->regs[i]; 1598 StoreW(addr, value, res); 1599 if (res) 1600 goto fault; 1601 addr += 4; 1602 } 1603 if ((reg & 0xf) == 9) { 1604 value = regs->regs[30]; 1605 StoreW(addr, value, res); 1606 if (res) 1607 goto fault; 1608 addr += 4; 1609 } 1610 if (reg & 0x10) { 1611 value = regs->regs[31]; 1612 StoreW(addr, value, res); 1613 if (res) 1614 goto fault; 1615 } 1616 goto success; 1617 1618 case mm_ldm_func: 1619 #ifdef CONFIG_64BIT 1620 reg = insn.mm_m_format.rd; 1621 rvar = reg & 0xf; 1622 if ((rvar > 9) || !reg) 1623 goto sigill; 1624 if (reg & 0x10) { 1625 if (!access_ok(addr, 8 * (rvar + 1))) 1626 goto sigbus; 1627 } else { 1628 if (!access_ok(addr, 8 * rvar)) 1629 goto sigbus; 1630 } 1631 if (rvar == 9) 1632 rvar = 8; 1633 1634 for (i = 16; rvar; rvar--, i++) { 1635 LoadDW(addr, value, res); 1636 if (res) 1637 goto fault; 1638 addr += 4; 1639 regs->regs[i] = value; 1640 } 1641 if ((reg & 0xf) == 9) { 1642 LoadDW(addr, value, res); 1643 if (res) 1644 goto fault; 1645 addr += 8; 1646 regs->regs[30] = value; 1647 } 1648 if (reg & 0x10) { 1649 LoadDW(addr, value, res); 1650 if (res) 1651 goto fault; 1652 regs->regs[31] = value; 1653 } 1654 goto success; 1655 #endif /* CONFIG_64BIT */ 1656 1657 goto sigill; 1658 1659 case mm_sdm_func: 1660 #ifdef CONFIG_64BIT 1661 reg = insn.mm_m_format.rd; 1662 rvar = reg & 0xf; 1663 if ((rvar > 9) || !reg) 1664 goto sigill; 1665 if (reg & 0x10) { 1666 if (!access_ok(addr, 8 * (rvar + 1))) 1667 goto sigbus; 1668 } else { 1669 if (!access_ok(addr, 8 * rvar)) 1670 goto sigbus; 1671 } 1672 if (rvar == 9) 1673 rvar = 8; 1674 1675 for (i = 16; rvar; rvar--, i++) { 1676 value = regs->regs[i]; 1677 StoreDW(addr, value, res); 1678 if (res) 1679 goto fault; 1680 addr += 8; 1681 } 1682 if ((reg & 0xf) == 9) { 1683 value = regs->regs[30]; 1684 StoreDW(addr, value, res); 1685 if (res) 1686 goto fault; 1687 addr += 8; 1688 } 1689 if (reg & 0x10) { 1690 value = regs->regs[31]; 1691 StoreDW(addr, value, res); 1692 if (res) 1693 goto fault; 1694 } 1695 goto success; 1696 #endif /* CONFIG_64BIT */ 1697 1698 goto sigill; 1699 1700 /* LWC2, SWC2, LDC2, SDC2 are not serviced */ 1701 } 1702 1703 goto sigbus; 1704 1705 case mm_pool32c_op: 1706 switch (insn.mm_m_format.func) { 1707 case mm_lwu_func: 1708 reg = insn.mm_m_format.rd; 1709 goto loadWU; 1710 } 1711 1712 /* LL,SC,LLD,SCD are not serviced */ 1713 goto sigbus; 1714 1715 #ifdef CONFIG_MIPS_FP_SUPPORT 1716 case mm_pool32f_op: 1717 switch (insn.mm_x_format.func) { 1718 case mm_lwxc1_func: 1719 case mm_swxc1_func: 1720 case mm_ldxc1_func: 1721 case mm_sdxc1_func: 1722 goto fpu_emul; 1723 } 1724 1725 goto sigbus; 1726 1727 case mm_ldc132_op: 1728 case mm_sdc132_op: 1729 case mm_lwc132_op: 1730 case mm_swc132_op: { 1731 void __user *fault_addr = NULL; 1732 1733 fpu_emul: 1734 /* roll back jump/branch */ 1735 regs->cp0_epc = origpc; 1736 regs->regs[31] = orig31; 1737 1738 die_if_kernel("Unaligned FP access in kernel code", regs); 1739 BUG_ON(!used_math()); 1740 BUG_ON(!is_fpu_owner()); 1741 1742 res = fpu_emulator_cop1Handler(regs, ¤t->thread.fpu, 1, 1743 &fault_addr); 1744 own_fpu(1); /* restore FPU state */ 1745 1746 /* If something went wrong, signal */ 1747 process_fpemu_return(res, fault_addr, 0); 1748 1749 if (res == 0) 1750 goto success; 1751 return; 1752 } 1753 #endif /* CONFIG_MIPS_FP_SUPPORT */ 1754 1755 case mm_lh32_op: 1756 reg = insn.mm_i_format.rt; 1757 goto loadHW; 1758 1759 case mm_lhu32_op: 1760 reg = insn.mm_i_format.rt; 1761 goto loadHWU; 1762 1763 case mm_lw32_op: 1764 reg = insn.mm_i_format.rt; 1765 goto loadW; 1766 1767 case mm_sh32_op: 1768 reg = insn.mm_i_format.rt; 1769 goto storeHW; 1770 1771 case mm_sw32_op: 1772 reg = insn.mm_i_format.rt; 1773 goto storeW; 1774 1775 case mm_ld32_op: 1776 reg = insn.mm_i_format.rt; 1777 goto loadDW; 1778 1779 case mm_sd32_op: 1780 reg = insn.mm_i_format.rt; 1781 goto storeDW; 1782 1783 case mm_pool16c_op: 1784 switch (insn.mm16_m_format.func) { 1785 case mm_lwm16_op: 1786 reg = insn.mm16_m_format.rlist; 1787 rvar = reg + 1; 1788 if (!access_ok(addr, 4 * rvar)) 1789 goto sigbus; 1790 1791 for (i = 16; rvar; rvar--, i++) { 1792 LoadW(addr, value, res); 1793 if (res) 1794 goto fault; 1795 addr += 4; 1796 regs->regs[i] = value; 1797 } 1798 LoadW(addr, value, res); 1799 if (res) 1800 goto fault; 1801 regs->regs[31] = value; 1802 1803 goto success; 1804 1805 case mm_swm16_op: 1806 reg = insn.mm16_m_format.rlist; 1807 rvar = reg + 1; 1808 if (!access_ok(addr, 4 * rvar)) 1809 goto sigbus; 1810 1811 for (i = 16; rvar; rvar--, i++) { 1812 value = regs->regs[i]; 1813 StoreW(addr, value, res); 1814 if (res) 1815 goto fault; 1816 addr += 4; 1817 } 1818 value = regs->regs[31]; 1819 StoreW(addr, value, res); 1820 if (res) 1821 goto fault; 1822 1823 goto success; 1824 1825 } 1826 1827 goto sigbus; 1828 1829 case mm_lhu16_op: 1830 reg = reg16to32[insn.mm16_rb_format.rt]; 1831 goto loadHWU; 1832 1833 case mm_lw16_op: 1834 reg = reg16to32[insn.mm16_rb_format.rt]; 1835 goto loadW; 1836 1837 case mm_sh16_op: 1838 reg = reg16to32st[insn.mm16_rb_format.rt]; 1839 goto storeHW; 1840 1841 case mm_sw16_op: 1842 reg = reg16to32st[insn.mm16_rb_format.rt]; 1843 goto storeW; 1844 1845 case mm_lwsp16_op: 1846 reg = insn.mm16_r5_format.rt; 1847 goto loadW; 1848 1849 case mm_swsp16_op: 1850 reg = insn.mm16_r5_format.rt; 1851 goto storeW; 1852 1853 case mm_lwgp16_op: 1854 reg = reg16to32[insn.mm16_r3_format.rt]; 1855 goto loadW; 1856 1857 default: 1858 goto sigill; 1859 } 1860 1861 loadHW: 1862 if (!access_ok(addr, 2)) 1863 goto sigbus; 1864 1865 LoadHW(addr, value, res); 1866 if (res) 1867 goto fault; 1868 regs->regs[reg] = value; 1869 goto success; 1870 1871 loadHWU: 1872 if (!access_ok(addr, 2)) 1873 goto sigbus; 1874 1875 LoadHWU(addr, value, res); 1876 if (res) 1877 goto fault; 1878 regs->regs[reg] = value; 1879 goto success; 1880 1881 loadW: 1882 if (!access_ok(addr, 4)) 1883 goto sigbus; 1884 1885 LoadW(addr, value, res); 1886 if (res) 1887 goto fault; 1888 regs->regs[reg] = value; 1889 goto success; 1890 1891 loadWU: 1892 #ifdef CONFIG_64BIT 1893 /* 1894 * A 32-bit kernel might be running on a 64-bit processor. But 1895 * if we're on a 32-bit processor and an i-cache incoherency 1896 * or race makes us see a 64-bit instruction here the sdl/sdr 1897 * would blow up, so for now we don't handle unaligned 64-bit 1898 * instructions on 32-bit kernels. 1899 */ 1900 if (!access_ok(addr, 4)) 1901 goto sigbus; 1902 1903 LoadWU(addr, value, res); 1904 if (res) 1905 goto fault; 1906 regs->regs[reg] = value; 1907 goto success; 1908 #endif /* CONFIG_64BIT */ 1909 1910 /* Cannot handle 64-bit instructions in 32-bit kernel */ 1911 goto sigill; 1912 1913 loadDW: 1914 #ifdef CONFIG_64BIT 1915 /* 1916 * A 32-bit kernel might be running on a 64-bit processor. But 1917 * if we're on a 32-bit processor and an i-cache incoherency 1918 * or race makes us see a 64-bit instruction here the sdl/sdr 1919 * would blow up, so for now we don't handle unaligned 64-bit 1920 * instructions on 32-bit kernels. 1921 */ 1922 if (!access_ok(addr, 8)) 1923 goto sigbus; 1924 1925 LoadDW(addr, value, res); 1926 if (res) 1927 goto fault; 1928 regs->regs[reg] = value; 1929 goto success; 1930 #endif /* CONFIG_64BIT */ 1931 1932 /* Cannot handle 64-bit instructions in 32-bit kernel */ 1933 goto sigill; 1934 1935 storeHW: 1936 if (!access_ok(addr, 2)) 1937 goto sigbus; 1938 1939 value = regs->regs[reg]; 1940 StoreHW(addr, value, res); 1941 if (res) 1942 goto fault; 1943 goto success; 1944 1945 storeW: 1946 if (!access_ok(addr, 4)) 1947 goto sigbus; 1948 1949 value = regs->regs[reg]; 1950 StoreW(addr, value, res); 1951 if (res) 1952 goto fault; 1953 goto success; 1954 1955 storeDW: 1956 #ifdef CONFIG_64BIT 1957 /* 1958 * A 32-bit kernel might be running on a 64-bit processor. But 1959 * if we're on a 32-bit processor and an i-cache incoherency 1960 * or race makes us see a 64-bit instruction here the sdl/sdr 1961 * would blow up, so for now we don't handle unaligned 64-bit 1962 * instructions on 32-bit kernels. 1963 */ 1964 if (!access_ok(addr, 8)) 1965 goto sigbus; 1966 1967 value = regs->regs[reg]; 1968 StoreDW(addr, value, res); 1969 if (res) 1970 goto fault; 1971 goto success; 1972 #endif /* CONFIG_64BIT */ 1973 1974 /* Cannot handle 64-bit instructions in 32-bit kernel */ 1975 goto sigill; 1976 1977 success: 1978 regs->cp0_epc = contpc; /* advance or branch */ 1979 1980 #ifdef CONFIG_DEBUG_FS 1981 unaligned_instructions++; 1982 #endif 1983 return; 1984 1985 fault: 1986 /* roll back jump/branch */ 1987 regs->cp0_epc = origpc; 1988 regs->regs[31] = orig31; 1989 /* Did we have an exception handler installed? */ 1990 if (fixup_exception(regs)) 1991 return; 1992 1993 die_if_kernel("Unhandled kernel unaligned access", regs); 1994 force_sig(SIGSEGV, current); 1995 1996 return; 1997 1998 sigbus: 1999 die_if_kernel("Unhandled kernel unaligned access", regs); 2000 force_sig(SIGBUS, current); 2001 2002 return; 2003 2004 sigill: 2005 die_if_kernel 2006 ("Unhandled kernel unaligned access or invalid instruction", regs); 2007 force_sig(SIGILL, current); 2008 } 2009 2010 static void emulate_load_store_MIPS16e(struct pt_regs *regs, void __user * addr) 2011 { 2012 unsigned long value; 2013 unsigned int res; 2014 int reg; 2015 unsigned long orig31; 2016 u16 __user *pc16; 2017 unsigned long origpc; 2018 union mips16e_instruction mips16inst, oldinst; 2019 unsigned int opcode; 2020 int extended = 0; 2021 2022 origpc = regs->cp0_epc; 2023 orig31 = regs->regs[31]; 2024 pc16 = (unsigned short __user *)msk_isa16_mode(origpc); 2025 /* 2026 * This load never faults. 2027 */ 2028 __get_user(mips16inst.full, pc16); 2029 oldinst = mips16inst; 2030 2031 /* skip EXTEND instruction */ 2032 if (mips16inst.ri.opcode == MIPS16e_extend_op) { 2033 extended = 1; 2034 pc16++; 2035 __get_user(mips16inst.full, pc16); 2036 } else if (delay_slot(regs)) { 2037 /* skip jump instructions */ 2038 /* JAL/JALX are 32 bits but have OPCODE in first short int */ 2039 if (mips16inst.ri.opcode == MIPS16e_jal_op) 2040 pc16++; 2041 pc16++; 2042 if (get_user(mips16inst.full, pc16)) 2043 goto sigbus; 2044 } 2045 2046 opcode = mips16inst.ri.opcode; 2047 switch (opcode) { 2048 case MIPS16e_i64_op: /* I64 or RI64 instruction */ 2049 switch (mips16inst.i64.func) { /* I64/RI64 func field check */ 2050 case MIPS16e_ldpc_func: 2051 case MIPS16e_ldsp_func: 2052 reg = reg16to32[mips16inst.ri64.ry]; 2053 goto loadDW; 2054 2055 case MIPS16e_sdsp_func: 2056 reg = reg16to32[mips16inst.ri64.ry]; 2057 goto writeDW; 2058 2059 case MIPS16e_sdrasp_func: 2060 reg = 29; /* GPRSP */ 2061 goto writeDW; 2062 } 2063 2064 goto sigbus; 2065 2066 case MIPS16e_swsp_op: 2067 reg = reg16to32[mips16inst.ri.rx]; 2068 if (extended && cpu_has_mips16e2) 2069 switch (mips16inst.ri.imm >> 5) { 2070 case 0: /* SWSP */ 2071 case 1: /* SWGP */ 2072 break; 2073 case 2: /* SHGP */ 2074 opcode = MIPS16e_sh_op; 2075 break; 2076 default: 2077 goto sigbus; 2078 } 2079 break; 2080 2081 case MIPS16e_lwpc_op: 2082 reg = reg16to32[mips16inst.ri.rx]; 2083 break; 2084 2085 case MIPS16e_lwsp_op: 2086 reg = reg16to32[mips16inst.ri.rx]; 2087 if (extended && cpu_has_mips16e2) 2088 switch (mips16inst.ri.imm >> 5) { 2089 case 0: /* LWSP */ 2090 case 1: /* LWGP */ 2091 break; 2092 case 2: /* LHGP */ 2093 opcode = MIPS16e_lh_op; 2094 break; 2095 case 4: /* LHUGP */ 2096 opcode = MIPS16e_lhu_op; 2097 break; 2098 default: 2099 goto sigbus; 2100 } 2101 break; 2102 2103 case MIPS16e_i8_op: 2104 if (mips16inst.i8.func != MIPS16e_swrasp_func) 2105 goto sigbus; 2106 reg = 29; /* GPRSP */ 2107 break; 2108 2109 default: 2110 reg = reg16to32[mips16inst.rri.ry]; 2111 break; 2112 } 2113 2114 switch (opcode) { 2115 2116 case MIPS16e_lb_op: 2117 case MIPS16e_lbu_op: 2118 case MIPS16e_sb_op: 2119 goto sigbus; 2120 2121 case MIPS16e_lh_op: 2122 if (!access_ok(addr, 2)) 2123 goto sigbus; 2124 2125 LoadHW(addr, value, res); 2126 if (res) 2127 goto fault; 2128 MIPS16e_compute_return_epc(regs, &oldinst); 2129 regs->regs[reg] = value; 2130 break; 2131 2132 case MIPS16e_lhu_op: 2133 if (!access_ok(addr, 2)) 2134 goto sigbus; 2135 2136 LoadHWU(addr, value, res); 2137 if (res) 2138 goto fault; 2139 MIPS16e_compute_return_epc(regs, &oldinst); 2140 regs->regs[reg] = value; 2141 break; 2142 2143 case MIPS16e_lw_op: 2144 case MIPS16e_lwpc_op: 2145 case MIPS16e_lwsp_op: 2146 if (!access_ok(addr, 4)) 2147 goto sigbus; 2148 2149 LoadW(addr, value, res); 2150 if (res) 2151 goto fault; 2152 MIPS16e_compute_return_epc(regs, &oldinst); 2153 regs->regs[reg] = value; 2154 break; 2155 2156 case MIPS16e_lwu_op: 2157 #ifdef CONFIG_64BIT 2158 /* 2159 * A 32-bit kernel might be running on a 64-bit processor. But 2160 * if we're on a 32-bit processor and an i-cache incoherency 2161 * or race makes us see a 64-bit instruction here the sdl/sdr 2162 * would blow up, so for now we don't handle unaligned 64-bit 2163 * instructions on 32-bit kernels. 2164 */ 2165 if (!access_ok(addr, 4)) 2166 goto sigbus; 2167 2168 LoadWU(addr, value, res); 2169 if (res) 2170 goto fault; 2171 MIPS16e_compute_return_epc(regs, &oldinst); 2172 regs->regs[reg] = value; 2173 break; 2174 #endif /* CONFIG_64BIT */ 2175 2176 /* Cannot handle 64-bit instructions in 32-bit kernel */ 2177 goto sigill; 2178 2179 case MIPS16e_ld_op: 2180 loadDW: 2181 #ifdef CONFIG_64BIT 2182 /* 2183 * A 32-bit kernel might be running on a 64-bit processor. But 2184 * if we're on a 32-bit processor and an i-cache incoherency 2185 * or race makes us see a 64-bit instruction here the sdl/sdr 2186 * would blow up, so for now we don't handle unaligned 64-bit 2187 * instructions on 32-bit kernels. 2188 */ 2189 if (!access_ok(addr, 8)) 2190 goto sigbus; 2191 2192 LoadDW(addr, value, res); 2193 if (res) 2194 goto fault; 2195 MIPS16e_compute_return_epc(regs, &oldinst); 2196 regs->regs[reg] = value; 2197 break; 2198 #endif /* CONFIG_64BIT */ 2199 2200 /* Cannot handle 64-bit instructions in 32-bit kernel */ 2201 goto sigill; 2202 2203 case MIPS16e_sh_op: 2204 if (!access_ok(addr, 2)) 2205 goto sigbus; 2206 2207 MIPS16e_compute_return_epc(regs, &oldinst); 2208 value = regs->regs[reg]; 2209 StoreHW(addr, value, res); 2210 if (res) 2211 goto fault; 2212 break; 2213 2214 case MIPS16e_sw_op: 2215 case MIPS16e_swsp_op: 2216 case MIPS16e_i8_op: /* actually - MIPS16e_swrasp_func */ 2217 if (!access_ok(addr, 4)) 2218 goto sigbus; 2219 2220 MIPS16e_compute_return_epc(regs, &oldinst); 2221 value = regs->regs[reg]; 2222 StoreW(addr, value, res); 2223 if (res) 2224 goto fault; 2225 break; 2226 2227 case MIPS16e_sd_op: 2228 writeDW: 2229 #ifdef CONFIG_64BIT 2230 /* 2231 * A 32-bit kernel might be running on a 64-bit processor. But 2232 * if we're on a 32-bit processor and an i-cache incoherency 2233 * or race makes us see a 64-bit instruction here the sdl/sdr 2234 * would blow up, so for now we don't handle unaligned 64-bit 2235 * instructions on 32-bit kernels. 2236 */ 2237 if (!access_ok(addr, 8)) 2238 goto sigbus; 2239 2240 MIPS16e_compute_return_epc(regs, &oldinst); 2241 value = regs->regs[reg]; 2242 StoreDW(addr, value, res); 2243 if (res) 2244 goto fault; 2245 break; 2246 #endif /* CONFIG_64BIT */ 2247 2248 /* Cannot handle 64-bit instructions in 32-bit kernel */ 2249 goto sigill; 2250 2251 default: 2252 /* 2253 * Pheeee... We encountered an yet unknown instruction or 2254 * cache coherence problem. Die sucker, die ... 2255 */ 2256 goto sigill; 2257 } 2258 2259 #ifdef CONFIG_DEBUG_FS 2260 unaligned_instructions++; 2261 #endif 2262 2263 return; 2264 2265 fault: 2266 /* roll back jump/branch */ 2267 regs->cp0_epc = origpc; 2268 regs->regs[31] = orig31; 2269 /* Did we have an exception handler installed? */ 2270 if (fixup_exception(regs)) 2271 return; 2272 2273 die_if_kernel("Unhandled kernel unaligned access", regs); 2274 force_sig(SIGSEGV, current); 2275 2276 return; 2277 2278 sigbus: 2279 die_if_kernel("Unhandled kernel unaligned access", regs); 2280 force_sig(SIGBUS, current); 2281 2282 return; 2283 2284 sigill: 2285 die_if_kernel 2286 ("Unhandled kernel unaligned access or invalid instruction", regs); 2287 force_sig(SIGILL, current); 2288 } 2289 2290 asmlinkage void do_ade(struct pt_regs *regs) 2291 { 2292 enum ctx_state prev_state; 2293 unsigned int __user *pc; 2294 mm_segment_t seg; 2295 2296 prev_state = exception_enter(); 2297 perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 2298 1, regs, regs->cp0_badvaddr); 2299 /* 2300 * Did we catch a fault trying to load an instruction? 2301 */ 2302 if (regs->cp0_badvaddr == regs->cp0_epc) 2303 goto sigbus; 2304 2305 if (user_mode(regs) && !test_thread_flag(TIF_FIXADE)) 2306 goto sigbus; 2307 if (unaligned_action == UNALIGNED_ACTION_SIGNAL) 2308 goto sigbus; 2309 2310 /* 2311 * Do branch emulation only if we didn't forward the exception. 2312 * This is all so but ugly ... 2313 */ 2314 2315 /* 2316 * Are we running in microMIPS mode? 2317 */ 2318 if (get_isa16_mode(regs->cp0_epc)) { 2319 /* 2320 * Did we catch a fault trying to load an instruction in 2321 * 16-bit mode? 2322 */ 2323 if (regs->cp0_badvaddr == msk_isa16_mode(regs->cp0_epc)) 2324 goto sigbus; 2325 if (unaligned_action == UNALIGNED_ACTION_SHOW) 2326 show_registers(regs); 2327 2328 if (cpu_has_mmips) { 2329 seg = get_fs(); 2330 if (!user_mode(regs)) 2331 set_fs(KERNEL_DS); 2332 emulate_load_store_microMIPS(regs, 2333 (void __user *)regs->cp0_badvaddr); 2334 set_fs(seg); 2335 2336 return; 2337 } 2338 2339 if (cpu_has_mips16) { 2340 seg = get_fs(); 2341 if (!user_mode(regs)) 2342 set_fs(KERNEL_DS); 2343 emulate_load_store_MIPS16e(regs, 2344 (void __user *)regs->cp0_badvaddr); 2345 set_fs(seg); 2346 2347 return; 2348 } 2349 2350 goto sigbus; 2351 } 2352 2353 if (unaligned_action == UNALIGNED_ACTION_SHOW) 2354 show_registers(regs); 2355 pc = (unsigned int __user *)exception_epc(regs); 2356 2357 seg = get_fs(); 2358 if (!user_mode(regs)) 2359 set_fs(KERNEL_DS); 2360 emulate_load_store_insn(regs, (void __user *)regs->cp0_badvaddr, pc); 2361 set_fs(seg); 2362 2363 return; 2364 2365 sigbus: 2366 die_if_kernel("Kernel unaligned instruction access", regs); 2367 force_sig(SIGBUS, current); 2368 2369 /* 2370 * XXX On return from the signal handler we should advance the epc 2371 */ 2372 exception_exit(prev_state); 2373 } 2374 2375 #ifdef CONFIG_DEBUG_FS 2376 static int __init debugfs_unaligned(void) 2377 { 2378 debugfs_create_u32("unaligned_instructions", S_IRUGO, mips_debugfs_dir, 2379 &unaligned_instructions); 2380 debugfs_create_u32("unaligned_action", S_IRUGO | S_IWUSR, 2381 mips_debugfs_dir, &unaligned_action); 2382 return 0; 2383 } 2384 arch_initcall(debugfs_unaligned); 2385 #endif 2386