123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304 |
- #include <linux/err.h>
- #include <linux/slab.h>
- #include <asm/branch.h>
- #include <asm/cacheflush.h>
- #include <asm/fpu_emulator.h>
- #include <asm/inst.h>
- #include <asm/mipsregs.h>
- #include <asm/uaccess.h>
- /**
- * struct emuframe - The 'emulation' frame structure
- * @emul: The instruction to 'emulate'.
- * @badinst: A break instruction to cause a return to the kernel.
- *
- * This structure defines the frames placed within the delay slot emulation
- * page in response to a call to mips_dsemul(). Each thread may be allocated
- * only one frame at any given time. The kernel stores within it the
- * instruction to be 'emulated' followed by a break instruction, then
- * executes the frame in user mode. The break causes a trap to the kernel
- * which leads to do_dsemulret() being called unless the instruction in
- * @emul causes a trap itself, is a branch, or a signal is delivered to
- * the thread. In these cases the allocated frame will either be reused by
- * a subsequent delay slot 'emulation', or be freed during signal delivery or
- * upon thread exit.
- *
- * This approach is used because:
- *
- * - Actually emulating all instructions isn't feasible. We would need to
- * be able to handle instructions from all revisions of the MIPS ISA,
- * all ASEs & all vendor instruction set extensions. This would be a
- * whole lot of work & continual maintenance burden as new instructions
- * are introduced, and in the case of some vendor extensions may not
- * even be possible. Thus we need to take the approach of actually
- * executing the instruction.
- *
- * - We must execute the instruction within user context. If we were to
- * execute the instruction in kernel mode then it would have access to
- * kernel resources without very careful checks, leaving us with a
- * high potential for security or stability issues to arise.
- *
- * - We used to place the frame on the users stack, but this requires
- * that the stack be executable. This is bad for security so the
- * per-process page is now used instead.
- *
- * - The instruction in @emul may be something entirely invalid for a
- * delay slot. The user may (intentionally or otherwise) place a branch
- * in a delay slot, or a kernel mode instruction, or something else
- * which generates an exception. Thus we can't rely upon the break in
- * @badinst always being hit. For this reason we track the index of the
- * frame allocated to each thread, allowing us to clean it up at later
- * points such as signal delivery or thread exit.
- *
- * - The user may generate a fake struct emuframe if they wish, invoking
- * the BRK_MEMU break instruction themselves. We must therefore not
- * trust that BRK_MEMU means there's actually a valid frame allocated
- * to the thread, and must not allow the user to do anything they
- * couldn't already.
- */
- struct emuframe {
- mips_instruction emul;
- mips_instruction badinst;
- };
- static const int emupage_frame_count = PAGE_SIZE / sizeof(struct emuframe);
- static inline __user struct emuframe *dsemul_page(void)
- {
- return (__user struct emuframe *)STACK_TOP;
- }
- static int alloc_emuframe(void)
- {
- mm_context_t *mm_ctx = ¤t->mm->context;
- int idx;
- retry:
- spin_lock(&mm_ctx->bd_emupage_lock);
- /* Ensure we have an allocation bitmap */
- if (!mm_ctx->bd_emupage_allocmap) {
- mm_ctx->bd_emupage_allocmap =
- kcalloc(BITS_TO_LONGS(emupage_frame_count),
- sizeof(unsigned long),
- GFP_ATOMIC);
- if (!mm_ctx->bd_emupage_allocmap) {
- idx = BD_EMUFRAME_NONE;
- goto out_unlock;
- }
- }
- /* Attempt to allocate a single bit/frame */
- idx = bitmap_find_free_region(mm_ctx->bd_emupage_allocmap,
- emupage_frame_count, 0);
- if (idx < 0) {
- /*
- * Failed to allocate a frame. We'll wait until one becomes
- * available. We unlock the page so that other threads actually
- * get the opportunity to free their frames, which means
- * technically the result of bitmap_full may be incorrect.
- * However the worst case is that we repeat all this and end up
- * back here again.
- */
- spin_unlock(&mm_ctx->bd_emupage_lock);
- if (!wait_event_killable(mm_ctx->bd_emupage_queue,
- !bitmap_full(mm_ctx->bd_emupage_allocmap,
- emupage_frame_count)))
- goto retry;
- /* Received a fatal signal - just give in */
- return BD_EMUFRAME_NONE;
- }
- /* Success! */
- pr_debug("allocate emuframe %d to %d\n", idx, current->pid);
- out_unlock:
- spin_unlock(&mm_ctx->bd_emupage_lock);
- return idx;
- }
- static void free_emuframe(int idx, struct mm_struct *mm)
- {
- mm_context_t *mm_ctx = &mm->context;
- spin_lock(&mm_ctx->bd_emupage_lock);
- pr_debug("free emuframe %d from %d\n", idx, current->pid);
- bitmap_clear(mm_ctx->bd_emupage_allocmap, idx, 1);
- /* If some thread is waiting for a frame, now's its chance */
- wake_up(&mm_ctx->bd_emupage_queue);
- spin_unlock(&mm_ctx->bd_emupage_lock);
- }
- static bool within_emuframe(struct pt_regs *regs)
- {
- unsigned long base = (unsigned long)dsemul_page();
- if (regs->cp0_epc < base)
- return false;
- if (regs->cp0_epc >= (base + PAGE_SIZE))
- return false;
- return true;
- }
- bool dsemul_thread_cleanup(struct task_struct *tsk)
- {
- int fr_idx;
- /* Clear any allocated frame, retrieving its index */
- fr_idx = atomic_xchg(&tsk->thread.bd_emu_frame, BD_EMUFRAME_NONE);
- /* If no frame was allocated, we're done */
- if (fr_idx == BD_EMUFRAME_NONE)
- return false;
- task_lock(tsk);
- /* Free the frame that this thread had allocated */
- if (tsk->mm)
- free_emuframe(fr_idx, tsk->mm);
- task_unlock(tsk);
- return true;
- }
- bool dsemul_thread_rollback(struct pt_regs *regs)
- {
- struct emuframe __user *fr;
- int fr_idx;
- /* Do nothing if we're not executing from a frame */
- if (!within_emuframe(regs))
- return false;
- /* Find the frame being executed */
- fr_idx = atomic_read(¤t->thread.bd_emu_frame);
- if (fr_idx == BD_EMUFRAME_NONE)
- return false;
- fr = &dsemul_page()[fr_idx];
- /*
- * If the PC is at the emul instruction, roll back to the branch. If
- * PC is at the badinst (break) instruction, we've already emulated the
- * instruction so progress to the continue PC. If it's anything else
- * then something is amiss & the user has branched into some other area
- * of the emupage - we'll free the allocated frame anyway.
- */
- if (msk_isa16_mode(regs->cp0_epc) == (unsigned long)&fr->emul)
- regs->cp0_epc = current->thread.bd_emu_branch_pc;
- else if (msk_isa16_mode(regs->cp0_epc) == (unsigned long)&fr->badinst)
- regs->cp0_epc = current->thread.bd_emu_cont_pc;
- atomic_set(¤t->thread.bd_emu_frame, BD_EMUFRAME_NONE);
- free_emuframe(fr_idx, current->mm);
- return true;
- }
- void dsemul_mm_cleanup(struct mm_struct *mm)
- {
- mm_context_t *mm_ctx = &mm->context;
- kfree(mm_ctx->bd_emupage_allocmap);
- }
- int mips_dsemul(struct pt_regs *regs, mips_instruction ir,
- unsigned long branch_pc, unsigned long cont_pc)
- {
- int isa16 = get_isa16_mode(regs->cp0_epc);
- mips_instruction break_math;
- struct emuframe __user *fr;
- int err, fr_idx;
- /* NOP is easy */
- if (ir == 0)
- return -1;
- /* microMIPS instructions */
- if (isa16) {
- union mips_instruction insn = { .word = ir };
- /* NOP16 aka MOVE16 $0, $0 */
- if ((ir >> 16) == MM_NOP16)
- return -1;
- /* ADDIUPC */
- if (insn.mm_a_format.opcode == mm_addiupc_op) {
- unsigned int rs;
- s32 v;
- rs = (((insn.mm_a_format.rs + 0xe) & 0xf) + 2);
- v = regs->cp0_epc & ~3;
- v += insn.mm_a_format.simmediate << 2;
- regs->regs[rs] = (long)v;
- return -1;
- }
- }
- pr_debug("dsemul 0x%08lx cont at 0x%08lx\n", regs->cp0_epc, cont_pc);
- /* Allocate a frame if we don't already have one */
- fr_idx = atomic_read(¤t->thread.bd_emu_frame);
- if (fr_idx == BD_EMUFRAME_NONE)
- fr_idx = alloc_emuframe();
- if (fr_idx == BD_EMUFRAME_NONE)
- return SIGBUS;
- fr = &dsemul_page()[fr_idx];
- /* Retrieve the appropriately encoded break instruction */
- break_math = BREAK_MATH(isa16);
- /* Write the instructions to the frame */
- if (isa16) {
- err = __put_user(ir >> 16,
- (u16 __user *)(&fr->emul));
- err |= __put_user(ir & 0xffff,
- (u16 __user *)((long)(&fr->emul) + 2));
- err |= __put_user(break_math >> 16,
- (u16 __user *)(&fr->badinst));
- err |= __put_user(break_math & 0xffff,
- (u16 __user *)((long)(&fr->badinst) + 2));
- } else {
- err = __put_user(ir, &fr->emul);
- err |= __put_user(break_math, &fr->badinst);
- }
- if (unlikely(err)) {
- MIPS_FPU_EMU_INC_STATS(errors);
- free_emuframe(fr_idx, current->mm);
- return SIGBUS;
- }
- /* Record the PC of the branch, PC to continue from & frame index */
- current->thread.bd_emu_branch_pc = branch_pc;
- current->thread.bd_emu_cont_pc = cont_pc;
- atomic_set(¤t->thread.bd_emu_frame, fr_idx);
- /* Change user register context to execute the frame */
- regs->cp0_epc = (unsigned long)&fr->emul | isa16;
- /* Ensure the icache observes our newly written frame */
- flush_cache_sigtramp((unsigned long)&fr->emul);
- return 0;
- }
- bool do_dsemulret(struct pt_regs *xcp)
- {
- /* Cleanup the allocated frame, returning if there wasn't one */
- if (!dsemul_thread_cleanup(current)) {
- MIPS_FPU_EMU_INC_STATS(errors);
- return false;
- }
- /* Set EPC to return to post-branch instruction */
- xcp->cp0_epc = current->thread.bd_emu_cont_pc;
- pr_debug("dsemulret to 0x%08lx\n", xcp->cp0_epc);
- MIPS_FPU_EMU_INC_STATS(ds_emul);
- return true;
- }
|