ptrace_64.c 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580
  1. /*
  2. * arch/sh/kernel/ptrace_64.c
  3. *
  4. * Copyright (C) 2000, 2001 Paolo Alberelli
  5. * Copyright (C) 2003 - 2008 Paul Mundt
  6. *
  7. * Started from SH3/4 version:
  8. * SuperH version: Copyright (C) 1999, 2000 Kaz Kojima & Niibe Yutaka
  9. *
  10. * Original x86 implementation:
  11. * By Ross Biro 1/23/92
  12. * edited by Linus Torvalds
  13. *
  14. * This file is subject to the terms and conditions of the GNU General Public
  15. * License. See the file "COPYING" in the main directory of this archive
  16. * for more details.
  17. */
  18. #include <linux/kernel.h>
  19. #include <linux/rwsem.h>
  20. #include <linux/sched.h>
  21. #include <linux/sched/task_stack.h>
  22. #include <linux/mm.h>
  23. #include <linux/smp.h>
  24. #include <linux/bitops.h>
  25. #include <linux/errno.h>
  26. #include <linux/ptrace.h>
  27. #include <linux/user.h>
  28. #include <linux/signal.h>
  29. #include <linux/syscalls.h>
  30. #include <linux/audit.h>
  31. #include <linux/seccomp.h>
  32. #include <linux/tracehook.h>
  33. #include <linux/elf.h>
  34. #include <linux/regset.h>
  35. #include <asm/io.h>
  36. #include <linux/uaccess.h>
  37. #include <asm/pgtable.h>
  38. #include <asm/processor.h>
  39. #include <asm/mmu_context.h>
  40. #include <asm/syscalls.h>
  41. #include <asm/fpu.h>
  42. #include <asm/traps.h>
  43. #define CREATE_TRACE_POINTS
  44. #include <trace/events/syscalls.h>
  45. /* This mask defines the bits of the SR which the user is not allowed to
  46. change, which are everything except S, Q, M, PR, SZ, FR. */
  47. #define SR_MASK (0xffff8cfd)
  48. /*
  49. * does not yet catch signals sent when the child dies.
  50. * in exit.c or in signal.c.
  51. */
  52. /*
  53. * This routine will get a word from the user area in the process kernel stack.
  54. */
  55. static inline int get_stack_long(struct task_struct *task, int offset)
  56. {
  57. unsigned char *stack;
  58. stack = (unsigned char *)(task->thread.uregs);
  59. stack += offset;
  60. return (*((int *)stack));
  61. }
  62. static inline unsigned long
  63. get_fpu_long(struct task_struct *task, unsigned long addr)
  64. {
  65. unsigned long tmp;
  66. struct pt_regs *regs;
  67. regs = (struct pt_regs*)((unsigned char *)task + THREAD_SIZE) - 1;
  68. if (!tsk_used_math(task)) {
  69. if (addr == offsetof(struct user_fpu_struct, fpscr)) {
  70. tmp = FPSCR_INIT;
  71. } else {
  72. tmp = 0xffffffffUL; /* matches initial value in fpu.c */
  73. }
  74. return tmp;
  75. }
  76. if (last_task_used_math == task) {
  77. enable_fpu();
  78. save_fpu(task);
  79. disable_fpu();
  80. last_task_used_math = 0;
  81. regs->sr |= SR_FD;
  82. }
  83. tmp = ((long *)task->thread.xstate)[addr / sizeof(unsigned long)];
  84. return tmp;
  85. }
  86. /*
  87. * This routine will put a word into the user area in the process kernel stack.
  88. */
  89. static inline int put_stack_long(struct task_struct *task, int offset,
  90. unsigned long data)
  91. {
  92. unsigned char *stack;
  93. stack = (unsigned char *)(task->thread.uregs);
  94. stack += offset;
  95. *(unsigned long *) stack = data;
  96. return 0;
  97. }
  98. static inline int
  99. put_fpu_long(struct task_struct *task, unsigned long addr, unsigned long data)
  100. {
  101. struct pt_regs *regs;
  102. regs = (struct pt_regs*)((unsigned char *)task + THREAD_SIZE) - 1;
  103. if (!tsk_used_math(task)) {
  104. init_fpu(task);
  105. } else if (last_task_used_math == task) {
  106. enable_fpu();
  107. save_fpu(task);
  108. disable_fpu();
  109. last_task_used_math = 0;
  110. regs->sr |= SR_FD;
  111. }
  112. ((long *)task->thread.xstate)[addr / sizeof(unsigned long)] = data;
  113. return 0;
  114. }
  115. void user_enable_single_step(struct task_struct *child)
  116. {
  117. struct pt_regs *regs = child->thread.uregs;
  118. regs->sr |= SR_SSTEP; /* auto-resetting upon exception */
  119. set_tsk_thread_flag(child, TIF_SINGLESTEP);
  120. }
  121. void user_disable_single_step(struct task_struct *child)
  122. {
  123. struct pt_regs *regs = child->thread.uregs;
  124. regs->sr &= ~SR_SSTEP;
  125. clear_tsk_thread_flag(child, TIF_SINGLESTEP);
  126. }
  127. static int genregs_get(struct task_struct *target,
  128. const struct user_regset *regset,
  129. unsigned int pos, unsigned int count,
  130. void *kbuf, void __user *ubuf)
  131. {
  132. const struct pt_regs *regs = task_pt_regs(target);
  133. int ret;
  134. /* PC, SR, SYSCALL */
  135. ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
  136. &regs->pc,
  137. 0, 3 * sizeof(unsigned long long));
  138. /* R1 -> R63 */
  139. if (!ret)
  140. ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
  141. regs->regs,
  142. offsetof(struct pt_regs, regs[0]),
  143. 63 * sizeof(unsigned long long));
  144. /* TR0 -> TR7 */
  145. if (!ret)
  146. ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
  147. regs->tregs,
  148. offsetof(struct pt_regs, tregs[0]),
  149. 8 * sizeof(unsigned long long));
  150. if (!ret)
  151. ret = user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
  152. sizeof(struct pt_regs), -1);
  153. return ret;
  154. }
  155. static int genregs_set(struct task_struct *target,
  156. const struct user_regset *regset,
  157. unsigned int pos, unsigned int count,
  158. const void *kbuf, const void __user *ubuf)
  159. {
  160. struct pt_regs *regs = task_pt_regs(target);
  161. int ret;
  162. /* PC, SR, SYSCALL */
  163. ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
  164. &regs->pc,
  165. 0, 3 * sizeof(unsigned long long));
  166. /* R1 -> R63 */
  167. if (!ret && count > 0)
  168. ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
  169. regs->regs,
  170. offsetof(struct pt_regs, regs[0]),
  171. 63 * sizeof(unsigned long long));
  172. /* TR0 -> TR7 */
  173. if (!ret && count > 0)
  174. ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
  175. regs->tregs,
  176. offsetof(struct pt_regs, tregs[0]),
  177. 8 * sizeof(unsigned long long));
  178. if (!ret)
  179. ret = user_regset_copyin_ignore(&pos, &count, &kbuf, &ubuf,
  180. sizeof(struct pt_regs), -1);
  181. return ret;
  182. }
  183. #ifdef CONFIG_SH_FPU
  184. int fpregs_get(struct task_struct *target,
  185. const struct user_regset *regset,
  186. unsigned int pos, unsigned int count,
  187. void *kbuf, void __user *ubuf)
  188. {
  189. int ret;
  190. ret = init_fpu(target);
  191. if (ret)
  192. return ret;
  193. return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
  194. &target->thread.xstate->hardfpu, 0, -1);
  195. }
  196. static int fpregs_set(struct task_struct *target,
  197. const struct user_regset *regset,
  198. unsigned int pos, unsigned int count,
  199. const void *kbuf, const void __user *ubuf)
  200. {
  201. int ret;
  202. ret = init_fpu(target);
  203. if (ret)
  204. return ret;
  205. set_stopped_child_used_math(target);
  206. return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
  207. &target->thread.xstate->hardfpu, 0, -1);
  208. }
  209. static int fpregs_active(struct task_struct *target,
  210. const struct user_regset *regset)
  211. {
  212. return tsk_used_math(target) ? regset->n : 0;
  213. }
  214. #endif
  215. const struct pt_regs_offset regoffset_table[] = {
  216. REG_OFFSET_NAME(pc),
  217. REG_OFFSET_NAME(sr),
  218. REG_OFFSET_NAME(syscall_nr),
  219. REGS_OFFSET_NAME(0),
  220. REGS_OFFSET_NAME(1),
  221. REGS_OFFSET_NAME(2),
  222. REGS_OFFSET_NAME(3),
  223. REGS_OFFSET_NAME(4),
  224. REGS_OFFSET_NAME(5),
  225. REGS_OFFSET_NAME(6),
  226. REGS_OFFSET_NAME(7),
  227. REGS_OFFSET_NAME(8),
  228. REGS_OFFSET_NAME(9),
  229. REGS_OFFSET_NAME(10),
  230. REGS_OFFSET_NAME(11),
  231. REGS_OFFSET_NAME(12),
  232. REGS_OFFSET_NAME(13),
  233. REGS_OFFSET_NAME(14),
  234. REGS_OFFSET_NAME(15),
  235. REGS_OFFSET_NAME(16),
  236. REGS_OFFSET_NAME(17),
  237. REGS_OFFSET_NAME(18),
  238. REGS_OFFSET_NAME(19),
  239. REGS_OFFSET_NAME(20),
  240. REGS_OFFSET_NAME(21),
  241. REGS_OFFSET_NAME(22),
  242. REGS_OFFSET_NAME(23),
  243. REGS_OFFSET_NAME(24),
  244. REGS_OFFSET_NAME(25),
  245. REGS_OFFSET_NAME(26),
  246. REGS_OFFSET_NAME(27),
  247. REGS_OFFSET_NAME(28),
  248. REGS_OFFSET_NAME(29),
  249. REGS_OFFSET_NAME(30),
  250. REGS_OFFSET_NAME(31),
  251. REGS_OFFSET_NAME(32),
  252. REGS_OFFSET_NAME(33),
  253. REGS_OFFSET_NAME(34),
  254. REGS_OFFSET_NAME(35),
  255. REGS_OFFSET_NAME(36),
  256. REGS_OFFSET_NAME(37),
  257. REGS_OFFSET_NAME(38),
  258. REGS_OFFSET_NAME(39),
  259. REGS_OFFSET_NAME(40),
  260. REGS_OFFSET_NAME(41),
  261. REGS_OFFSET_NAME(42),
  262. REGS_OFFSET_NAME(43),
  263. REGS_OFFSET_NAME(44),
  264. REGS_OFFSET_NAME(45),
  265. REGS_OFFSET_NAME(46),
  266. REGS_OFFSET_NAME(47),
  267. REGS_OFFSET_NAME(48),
  268. REGS_OFFSET_NAME(49),
  269. REGS_OFFSET_NAME(50),
  270. REGS_OFFSET_NAME(51),
  271. REGS_OFFSET_NAME(52),
  272. REGS_OFFSET_NAME(53),
  273. REGS_OFFSET_NAME(54),
  274. REGS_OFFSET_NAME(55),
  275. REGS_OFFSET_NAME(56),
  276. REGS_OFFSET_NAME(57),
  277. REGS_OFFSET_NAME(58),
  278. REGS_OFFSET_NAME(59),
  279. REGS_OFFSET_NAME(60),
  280. REGS_OFFSET_NAME(61),
  281. REGS_OFFSET_NAME(62),
  282. REGS_OFFSET_NAME(63),
  283. TREGS_OFFSET_NAME(0),
  284. TREGS_OFFSET_NAME(1),
  285. TREGS_OFFSET_NAME(2),
  286. TREGS_OFFSET_NAME(3),
  287. TREGS_OFFSET_NAME(4),
  288. TREGS_OFFSET_NAME(5),
  289. TREGS_OFFSET_NAME(6),
  290. TREGS_OFFSET_NAME(7),
  291. REG_OFFSET_END,
  292. };
  293. /*
  294. * These are our native regset flavours.
  295. */
  296. enum sh_regset {
  297. REGSET_GENERAL,
  298. #ifdef CONFIG_SH_FPU
  299. REGSET_FPU,
  300. #endif
  301. };
  302. static const struct user_regset sh_regsets[] = {
  303. /*
  304. * Format is:
  305. * PC, SR, SYSCALL,
  306. * R1 --> R63,
  307. * TR0 --> TR7,
  308. */
  309. [REGSET_GENERAL] = {
  310. .core_note_type = NT_PRSTATUS,
  311. .n = ELF_NGREG,
  312. .size = sizeof(long long),
  313. .align = sizeof(long long),
  314. .get = genregs_get,
  315. .set = genregs_set,
  316. },
  317. #ifdef CONFIG_SH_FPU
  318. [REGSET_FPU] = {
  319. .core_note_type = NT_PRFPREG,
  320. .n = sizeof(struct user_fpu_struct) /
  321. sizeof(long long),
  322. .size = sizeof(long long),
  323. .align = sizeof(long long),
  324. .get = fpregs_get,
  325. .set = fpregs_set,
  326. .active = fpregs_active,
  327. },
  328. #endif
  329. };
  330. static const struct user_regset_view user_sh64_native_view = {
  331. .name = "sh64",
  332. .e_machine = EM_SH,
  333. .regsets = sh_regsets,
  334. .n = ARRAY_SIZE(sh_regsets),
  335. };
  336. const struct user_regset_view *task_user_regset_view(struct task_struct *task)
  337. {
  338. return &user_sh64_native_view;
  339. }
  340. long arch_ptrace(struct task_struct *child, long request,
  341. unsigned long addr, unsigned long data)
  342. {
  343. int ret;
  344. unsigned long __user *datap = (unsigned long __user *) data;
  345. switch (request) {
  346. /* read the word at location addr in the USER area. */
  347. case PTRACE_PEEKUSR: {
  348. unsigned long tmp;
  349. ret = -EIO;
  350. if ((addr & 3) || addr < 0)
  351. break;
  352. if (addr < sizeof(struct pt_regs))
  353. tmp = get_stack_long(child, addr);
  354. else if ((addr >= offsetof(struct user, fpu)) &&
  355. (addr < offsetof(struct user, u_fpvalid))) {
  356. unsigned long index;
  357. ret = init_fpu(child);
  358. if (ret)
  359. break;
  360. index = addr - offsetof(struct user, fpu);
  361. tmp = get_fpu_long(child, index);
  362. } else if (addr == offsetof(struct user, u_fpvalid)) {
  363. tmp = !!tsk_used_math(child);
  364. } else {
  365. break;
  366. }
  367. ret = put_user(tmp, datap);
  368. break;
  369. }
  370. case PTRACE_POKEUSR:
  371. /* write the word at location addr in the USER area. We must
  372. disallow any changes to certain SR bits or u_fpvalid, since
  373. this could crash the kernel or result in a security
  374. loophole. */
  375. ret = -EIO;
  376. if ((addr & 3) || addr < 0)
  377. break;
  378. if (addr < sizeof(struct pt_regs)) {
  379. /* Ignore change of top 32 bits of SR */
  380. if (addr == offsetof (struct pt_regs, sr)+4)
  381. {
  382. ret = 0;
  383. break;
  384. }
  385. /* If lower 32 bits of SR, ignore non-user bits */
  386. if (addr == offsetof (struct pt_regs, sr))
  387. {
  388. long cursr = get_stack_long(child, addr);
  389. data &= ~(SR_MASK);
  390. data |= (cursr & SR_MASK);
  391. }
  392. ret = put_stack_long(child, addr, data);
  393. }
  394. else if ((addr >= offsetof(struct user, fpu)) &&
  395. (addr < offsetof(struct user, u_fpvalid))) {
  396. unsigned long index;
  397. ret = init_fpu(child);
  398. if (ret)
  399. break;
  400. index = addr - offsetof(struct user, fpu);
  401. ret = put_fpu_long(child, index, data);
  402. }
  403. break;
  404. case PTRACE_GETREGS:
  405. return copy_regset_to_user(child, &user_sh64_native_view,
  406. REGSET_GENERAL,
  407. 0, sizeof(struct pt_regs),
  408. datap);
  409. case PTRACE_SETREGS:
  410. return copy_regset_from_user(child, &user_sh64_native_view,
  411. REGSET_GENERAL,
  412. 0, sizeof(struct pt_regs),
  413. datap);
  414. #ifdef CONFIG_SH_FPU
  415. case PTRACE_GETFPREGS:
  416. return copy_regset_to_user(child, &user_sh64_native_view,
  417. REGSET_FPU,
  418. 0, sizeof(struct user_fpu_struct),
  419. datap);
  420. case PTRACE_SETFPREGS:
  421. return copy_regset_from_user(child, &user_sh64_native_view,
  422. REGSET_FPU,
  423. 0, sizeof(struct user_fpu_struct),
  424. datap);
  425. #endif
  426. default:
  427. ret = ptrace_request(child, request, addr, data);
  428. break;
  429. }
  430. return ret;
  431. }
  432. asmlinkage int sh64_ptrace(long request, long pid,
  433. unsigned long addr, unsigned long data)
  434. {
  435. #define WPC_DBRMODE 0x0d104008
  436. static unsigned long first_call;
  437. if (!test_and_set_bit(0, &first_call)) {
  438. /* Set WPC.DBRMODE to 0. This makes all debug events get
  439. * delivered through RESVEC, i.e. into the handlers in entry.S.
  440. * (If the kernel was downloaded using a remote gdb, WPC.DBRMODE
  441. * would normally be left set to 1, which makes debug events get
  442. * delivered through DBRVEC, i.e. into the remote gdb's
  443. * handlers. This prevents ptrace getting them, and confuses
  444. * the remote gdb.) */
  445. printk("DBRMODE set to 0 to permit native debugging\n");
  446. poke_real_address_q(WPC_DBRMODE, 0);
  447. }
  448. return sys_ptrace(request, pid, addr, data);
  449. }
  450. asmlinkage long long do_syscall_trace_enter(struct pt_regs *regs)
  451. {
  452. long long ret = 0;
  453. secure_computing_strict(regs->regs[9]);
  454. if (test_thread_flag(TIF_SYSCALL_TRACE) &&
  455. tracehook_report_syscall_entry(regs))
  456. /*
  457. * Tracing decided this syscall should not happen.
  458. * We'll return a bogus call number to get an ENOSYS
  459. * error, but leave the original number in regs->regs[0].
  460. */
  461. ret = -1LL;
  462. if (unlikely(test_thread_flag(TIF_SYSCALL_TRACEPOINT)))
  463. trace_sys_enter(regs, regs->regs[9]);
  464. audit_syscall_entry(regs->regs[1], regs->regs[2], regs->regs[3],
  465. regs->regs[4], regs->regs[5]);
  466. return ret ?: regs->regs[9];
  467. }
  468. asmlinkage void do_syscall_trace_leave(struct pt_regs *regs)
  469. {
  470. int step;
  471. audit_syscall_exit(regs);
  472. if (unlikely(test_thread_flag(TIF_SYSCALL_TRACEPOINT)))
  473. trace_sys_exit(regs, regs->regs[9]);
  474. step = test_thread_flag(TIF_SINGLESTEP);
  475. if (step || test_thread_flag(TIF_SYSCALL_TRACE))
  476. tracehook_report_syscall_exit(regs, step);
  477. }
  478. /* Called with interrupts disabled */
  479. asmlinkage void do_single_step(unsigned long long vec, struct pt_regs *regs)
  480. {
  481. /* This is called after a single step exception (DEBUGSS).
  482. There is no need to change the PC, as it is a post-execution
  483. exception, as entry.S does not do anything to the PC for DEBUGSS.
  484. We need to clear the Single Step setting in SR to avoid
  485. continually stepping. */
  486. local_irq_enable();
  487. regs->sr &= ~SR_SSTEP;
  488. force_sig(SIGTRAP, current);
  489. }
  490. /* Called with interrupts disabled */
  491. BUILD_TRAP_HANDLER(breakpoint)
  492. {
  493. TRAP_HANDLER_DECL;
  494. /* We need to forward step the PC, to counteract the backstep done
  495. in signal.c. */
  496. local_irq_enable();
  497. force_sig(SIGTRAP, current);
  498. regs->pc += 4;
  499. }
  500. /*
  501. * Called by kernel/ptrace.c when detaching..
  502. *
  503. * Make sure single step bits etc are not set.
  504. */
  505. void ptrace_disable(struct task_struct *child)
  506. {
  507. user_disable_single_step(child);
  508. }