123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946 |
- /*
- * Linux/PA-RISC Project (http://www.parisc-linux.org/)
- *
- * System call entry code / Linux gateway page
- * Copyright (c) Matthew Wilcox 1999 <willy@bofh.ai>
- * Licensed under the GNU GPL.
- * thanks to Philipp Rumpf, Mike Shaver and various others
- * sorry about the wall, puffin..
- */
- /*
- How does the Linux gateway page on PA-RISC work?
- ------------------------------------------------
- The Linux gateway page on PA-RISC is "special".
- It actually has PAGE_GATEWAY bits set (this is linux terminology; in parisc
- terminology it's Execute, promote to PL0) in the page map. So anything
- executing on this page executes with kernel level privilege (there's more to it
- than that: to have this happen, you also have to use a branch with a ,gate
- completer to activate the privilege promotion). The upshot is that everything
- that runs on the gateway page runs at kernel privilege but with the current
- user process address space (although you have access to kernel space via %sr2).
- For the 0x100 syscall entry, we redo the space registers to point to the kernel
- address space (preserving the user address space in %sr3), move to wide mode if
- required, save the user registers and branch into the kernel syscall entry
- point. For all the other functions, we execute at kernel privilege but don't
- flip address spaces. The basic upshot of this is that these code snippets are
- executed atomically (because the kernel can't be pre-empted) and they may
- perform architecturally forbidden (to PL3) operations (like setting control
- registers).
- */
- #include <asm/asm-offsets.h>
- #include <asm/unistd.h>
- #include <asm/errno.h>
- #include <asm/page.h>
- #include <asm/psw.h>
- #include <asm/thread_info.h>
- #include <asm/assembly.h>
- #include <asm/processor.h>
- #include <asm/cache.h>
- #include <linux/linkage.h>
- /* We fill the empty parts of the gateway page with
- * something that will kill the kernel or a
- * userspace application.
- */
- #define KILL_INSN break 0,0
- .level LEVEL
- .text
- .import syscall_exit,code
- .import syscall_exit_rfi,code
- /* Linux gateway page is aliased to virtual page 0 in the kernel
- * address space. Since it is a gateway page it cannot be
- * dereferenced, so null pointers will still fault. We start
- * the actual entry point at 0x100. We put break instructions
- * at the beginning of the page to trap null indirect function
- * pointers.
- */
- .align PAGE_SIZE
- ENTRY(linux_gateway_page)
- /* ADDRESS 0x00 to 0xb0 = 176 bytes / 4 bytes per insn = 44 insns */
- .rept 44
- KILL_INSN
- .endr
- /* ADDRESS 0xb0 to 0xb8, lws uses two insns for entry */
- /* Light-weight-syscall entry must always be located at 0xb0 */
- /* WARNING: Keep this number updated with table size changes */
- #define __NR_lws_entries (3)
- lws_entry:
- gate lws_start, %r0 /* increase privilege */
- depi 3, 31, 2, %r31 /* Ensure we return into user mode. */
- /* Fill from 0xb8 to 0xe0 */
- .rept 10
- KILL_INSN
- .endr
- /* This function MUST be located at 0xe0 for glibc's threading
- mechanism to work. DO NOT MOVE THIS CODE EVER! */
- set_thread_pointer:
- gate .+8, %r0 /* increase privilege */
- depi 3, 31, 2, %r31 /* Ensure we return into user mode. */
- be 0(%sr7,%r31) /* return to user space */
- mtctl %r26, %cr27 /* move arg0 to the control register */
- /* Increase the chance of trapping if random jumps occur to this
- address, fill from 0xf0 to 0x100 */
- .rept 4
- KILL_INSN
- .endr
- /* This address must remain fixed at 0x100 for glibc's syscalls to work */
- .align 256
- linux_gateway_entry:
- gate .+8, %r0 /* become privileged */
- mtsp %r0,%sr4 /* get kernel space into sr4 */
- mtsp %r0,%sr5 /* get kernel space into sr5 */
- mtsp %r0,%sr6 /* get kernel space into sr6 */
- mfsp %sr7,%r1 /* save user sr7 */
- mtsp %r1,%sr3 /* and store it in sr3 */
- #ifdef CONFIG_64BIT
- /* for now we can *always* set the W bit on entry to the syscall
- * since we don't support wide userland processes. We could
- * also save the current SM other than in r0 and restore it on
- * exit from the syscall, and also use that value to know
- * whether to do narrow or wide syscalls. -PB
- */
- ssm PSW_SM_W, %r1
- extrd,u %r1,PSW_W_BIT,1,%r1
- /* sp must be aligned on 4, so deposit the W bit setting into
- * the bottom of sp temporarily */
- or,ev %r1,%r30,%r30
- b,n 1f
- /* The top halves of argument registers must be cleared on syscall
- * entry from narrow executable.
- */
- depdi 0, 31, 32, %r26
- depdi 0, 31, 32, %r25
- depdi 0, 31, 32, %r24
- depdi 0, 31, 32, %r23
- depdi 0, 31, 32, %r22
- depdi 0, 31, 32, %r21
- 1:
- #endif
- mfctl %cr30,%r1
- xor %r1,%r30,%r30 /* ye olde xor trick */
- xor %r1,%r30,%r1
- xor %r1,%r30,%r30
-
- ldo THREAD_SZ_ALGN+FRAME_SIZE(%r30),%r30 /* set up kernel stack */
- /* N.B.: It is critical that we don't set sr7 to 0 until r30
- * contains a valid kernel stack pointer. It is also
- * critical that we don't start using the kernel stack
- * until after sr7 has been set to 0.
- */
- mtsp %r0,%sr7 /* get kernel space into sr7 */
- STREGM %r1,FRAME_SIZE(%r30) /* save r1 (usp) here for now */
- mfctl %cr30,%r1 /* get task ptr in %r1 */
- LDREG TI_TASK(%r1),%r1
- /* Save some registers for sigcontext and potential task
- switch (see entry.S for the details of which ones are
- saved/restored). TASK_PT_PSW is zeroed so we can see whether
- a process is on a syscall or not. For an interrupt the real
- PSW value is stored. This is needed for gdb and sys_ptrace. */
- STREG %r0, TASK_PT_PSW(%r1)
- STREG %r2, TASK_PT_GR2(%r1) /* preserve rp */
- STREG %r19, TASK_PT_GR19(%r1)
- LDREGM -FRAME_SIZE(%r30), %r2 /* get users sp back */
- #ifdef CONFIG_64BIT
- extrd,u %r2,63,1,%r19 /* W hidden in bottom bit */
- #if 0
- xor %r19,%r2,%r2 /* clear bottom bit */
- depd,z %r19,1,1,%r19
- std %r19,TASK_PT_PSW(%r1)
- #endif
- #endif
- STREG %r2, TASK_PT_GR30(%r1) /* ... and save it */
-
- STREG %r20, TASK_PT_GR20(%r1) /* Syscall number */
- STREG %r21, TASK_PT_GR21(%r1)
- STREG %r22, TASK_PT_GR22(%r1)
- STREG %r23, TASK_PT_GR23(%r1) /* 4th argument */
- STREG %r24, TASK_PT_GR24(%r1) /* 3rd argument */
- STREG %r25, TASK_PT_GR25(%r1) /* 2nd argument */
- STREG %r26, TASK_PT_GR26(%r1) /* 1st argument */
- STREG %r27, TASK_PT_GR27(%r1) /* user dp */
- STREG %r28, TASK_PT_GR28(%r1) /* return value 0 */
- STREG %r0, TASK_PT_ORIG_R28(%r1) /* don't prohibit restarts */
- STREG %r29, TASK_PT_GR29(%r1) /* return value 1 */
- STREG %r31, TASK_PT_GR31(%r1) /* preserve syscall return ptr */
-
- ldo TASK_PT_FR0(%r1), %r27 /* save fpregs from the kernel */
- save_fp %r27 /* or potential task switch */
- mfctl %cr11, %r27 /* i.e. SAR */
- STREG %r27, TASK_PT_SAR(%r1)
- loadgp
- #ifdef CONFIG_64BIT
- ldo -16(%r30),%r29 /* Reference param save area */
- copy %r19,%r2 /* W bit back to r2 */
- #else
- /* no need to save these on stack in wide mode because the first 8
- * args are passed in registers */
- stw %r22, -52(%r30) /* 5th argument */
- stw %r21, -56(%r30) /* 6th argument */
- #endif
- /* Are we being ptraced? */
- mfctl %cr30, %r1
- LDREG TI_FLAGS(%r1),%r1
- ldi _TIF_SYSCALL_TRACE_MASK, %r19
- and,COND(=) %r1, %r19, %r0
- b,n .Ltracesys
-
- /* Note! We cannot use the syscall table that is mapped
- nearby since the gateway page is mapped execute-only. */
- #ifdef CONFIG_64BIT
- ldil L%sys_call_table, %r1
- or,= %r2,%r2,%r2
- addil L%(sys_call_table64-sys_call_table), %r1
- ldo R%sys_call_table(%r1), %r19
- or,= %r2,%r2,%r2
- ldo R%sys_call_table64(%r1), %r19
- #else
- ldil L%sys_call_table, %r1
- ldo R%sys_call_table(%r1), %r19
- #endif
- comiclr,>> __NR_Linux_syscalls, %r20, %r0
- b,n .Lsyscall_nosys
-
- LDREGX %r20(%r19), %r19
- /* If this is a sys_rt_sigreturn call, and the signal was received
- * when not in_syscall, then we want to return via syscall_exit_rfi,
- * not syscall_exit. Signal no. in r20, in_syscall in r25 (see
- * trampoline code in signal.c).
- */
- ldi __NR_rt_sigreturn,%r2
- comb,= %r2,%r20,.Lrt_sigreturn
- .Lin_syscall:
- ldil L%syscall_exit,%r2
- be 0(%sr7,%r19)
- ldo R%syscall_exit(%r2),%r2
- .Lrt_sigreturn:
- comib,<> 0,%r25,.Lin_syscall
- ldil L%syscall_exit_rfi,%r2
- be 0(%sr7,%r19)
- ldo R%syscall_exit_rfi(%r2),%r2
- /* Note! Because we are not running where we were linked, any
- calls to functions external to this file must be indirect. To
- be safe, we apply the opposite rule to functions within this
- file, with local labels given to them to ensure correctness. */
-
- .Lsyscall_nosys:
- syscall_nosys:
- ldil L%syscall_exit,%r1
- be R%syscall_exit(%sr7,%r1)
- ldo -ENOSYS(%r0),%r28 /* set errno */
- /* Warning! This trace code is a virtual duplicate of the code above so be
- * sure to maintain both! */
- .Ltracesys:
- tracesys:
- /* Need to save more registers so the debugger can see where we
- * are. This saves only the lower 8 bits of PSW, so that the C
- * bit is still clear on syscalls, and the D bit is set if this
- * full register save path has been executed. We check the D
- * bit on syscall_return_rfi to determine which registers to
- * restore. An interrupt results in a full PSW saved with the
- * C bit set, a non-straced syscall entry results in C and D clear
- * in the saved PSW.
- */
- ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
- LDREG TI_TASK(%r1), %r1
- ssm 0,%r2
- STREG %r2,TASK_PT_PSW(%r1) /* Lower 8 bits only!! */
- mfsp %sr0,%r2
- STREG %r2,TASK_PT_SR0(%r1)
- mfsp %sr1,%r2
- STREG %r2,TASK_PT_SR1(%r1)
- mfsp %sr2,%r2
- STREG %r2,TASK_PT_SR2(%r1)
- mfsp %sr3,%r2
- STREG %r2,TASK_PT_SR3(%r1)
- STREG %r2,TASK_PT_SR4(%r1)
- STREG %r2,TASK_PT_SR5(%r1)
- STREG %r2,TASK_PT_SR6(%r1)
- STREG %r2,TASK_PT_SR7(%r1)
- STREG %r2,TASK_PT_IASQ0(%r1)
- STREG %r2,TASK_PT_IASQ1(%r1)
- LDREG TASK_PT_GR31(%r1),%r2
- STREG %r2,TASK_PT_IAOQ0(%r1)
- ldo 4(%r2),%r2
- STREG %r2,TASK_PT_IAOQ1(%r1)
- ldo TASK_REGS(%r1),%r2
- /* reg_save %r2 */
- STREG %r3,PT_GR3(%r2)
- STREG %r4,PT_GR4(%r2)
- STREG %r5,PT_GR5(%r2)
- STREG %r6,PT_GR6(%r2)
- STREG %r7,PT_GR7(%r2)
- STREG %r8,PT_GR8(%r2)
- STREG %r9,PT_GR9(%r2)
- STREG %r10,PT_GR10(%r2)
- STREG %r11,PT_GR11(%r2)
- STREG %r12,PT_GR12(%r2)
- STREG %r13,PT_GR13(%r2)
- STREG %r14,PT_GR14(%r2)
- STREG %r15,PT_GR15(%r2)
- STREG %r16,PT_GR16(%r2)
- STREG %r17,PT_GR17(%r2)
- STREG %r18,PT_GR18(%r2)
- /* Finished saving things for the debugger */
- copy %r2,%r26
- ldil L%do_syscall_trace_enter,%r1
- ldil L%tracesys_next,%r2
- be R%do_syscall_trace_enter(%sr7,%r1)
- ldo R%tracesys_next(%r2),%r2
-
- tracesys_next:
- /* do_syscall_trace_enter either returned the syscallno, or -1L,
- * so we skip restoring the PT_GR20 below, since we pulled it from
- * task->thread.regs.gr[20] above.
- */
- copy %ret0,%r20
- ldil L%sys_call_table,%r1
- ldo R%sys_call_table(%r1), %r19
- ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
- LDREG TI_TASK(%r1), %r1
- LDREG TASK_PT_GR26(%r1), %r26 /* Restore the users args */
- LDREG TASK_PT_GR25(%r1), %r25
- LDREG TASK_PT_GR24(%r1), %r24
- LDREG TASK_PT_GR23(%r1), %r23
- LDREG TASK_PT_GR22(%r1), %r22
- LDREG TASK_PT_GR21(%r1), %r21
- #ifdef CONFIG_64BIT
- ldo -16(%r30),%r29 /* Reference param save area */
- #else
- stw %r22, -52(%r30) /* 5th argument */
- stw %r21, -56(%r30) /* 6th argument */
- #endif
- comiclr,>>= __NR_Linux_syscalls, %r20, %r0
- b,n .Lsyscall_nosys
- LDREGX %r20(%r19), %r19
- /* If this is a sys_rt_sigreturn call, and the signal was received
- * when not in_syscall, then we want to return via syscall_exit_rfi,
- * not syscall_exit. Signal no. in r20, in_syscall in r25 (see
- * trampoline code in signal.c).
- */
- ldi __NR_rt_sigreturn,%r2
- comb,= %r2,%r20,.Ltrace_rt_sigreturn
- .Ltrace_in_syscall:
- ldil L%tracesys_exit,%r2
- be 0(%sr7,%r19)
- ldo R%tracesys_exit(%r2),%r2
- /* Do *not* call this function on the gateway page, because it
- makes a direct call to syscall_trace. */
-
- tracesys_exit:
- ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
- LDREG TI_TASK(%r1), %r1
- #ifdef CONFIG_64BIT
- ldo -16(%r30),%r29 /* Reference param save area */
- #endif
- ldo TASK_REGS(%r1),%r26
- bl do_syscall_trace_exit,%r2
- STREG %r28,TASK_PT_GR28(%r1) /* save return value now */
- ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
- LDREG TI_TASK(%r1), %r1
- LDREG TASK_PT_GR28(%r1), %r28 /* Restore return val. */
- ldil L%syscall_exit,%r1
- be,n R%syscall_exit(%sr7,%r1)
- .Ltrace_rt_sigreturn:
- comib,<> 0,%r25,.Ltrace_in_syscall
- ldil L%tracesys_sigexit,%r2
- be 0(%sr7,%r19)
- ldo R%tracesys_sigexit(%r2),%r2
- tracesys_sigexit:
- ldo -THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1 /* get task ptr */
- LDREG TI_TASK(%r1), %r1
- #ifdef CONFIG_64BIT
- ldo -16(%r30),%r29 /* Reference param save area */
- #endif
- bl do_syscall_trace_exit,%r2
- ldo TASK_REGS(%r1),%r26
- ldil L%syscall_exit_rfi,%r1
- be,n R%syscall_exit_rfi(%sr7,%r1)
- /*********************************************************
- 32/64-bit Light-Weight-Syscall ABI
- * - Indicates a hint for userspace inline asm
- implementations.
- Syscall number (caller-saves)
- - %r20
- * In asm clobber.
- Argument registers (caller-saves)
- - %r26, %r25, %r24, %r23, %r22
- * In asm input.
- Return registers (caller-saves)
- - %r28 (return), %r21 (errno)
- * In asm output.
- Caller-saves registers
- - %r1, %r27, %r29
- - %r2 (return pointer)
- - %r31 (ble link register)
- * In asm clobber.
- Callee-saves registers
- - %r3-%r18
- - %r30 (stack pointer)
- * Not in asm clobber.
- If userspace is 32-bit:
- Callee-saves registers
- - %r19 (32-bit PIC register)
- Differences from 32-bit calling convention:
- - Syscall number in %r20
- - Additional argument register %r22 (arg4)
- - Callee-saves %r19.
- If userspace is 64-bit:
- Callee-saves registers
- - %r27 (64-bit PIC register)
- Differences from 64-bit calling convention:
- - Syscall number in %r20
- - Additional argument register %r22 (arg4)
- - Callee-saves %r27.
- Error codes returned by entry path:
- ENOSYS - r20 was an invalid LWS number.
- *********************************************************/
- lws_start:
- #ifdef CONFIG_64BIT
- /* FIXME: If we are a 64-bit kernel just
- * turn this on unconditionally.
- */
- ssm PSW_SM_W, %r1
- extrd,u %r1,PSW_W_BIT,1,%r1
- /* sp must be aligned on 4, so deposit the W bit setting into
- * the bottom of sp temporarily */
- or,ev %r1,%r30,%r30
- /* Clip LWS number to a 32-bit value always */
- depdi 0, 31, 32, %r20
- #endif
- /* Is the lws entry number valid? */
- comiclr,>> __NR_lws_entries, %r20, %r0
- b,n lws_exit_nosys
- /* WARNING: Trashing sr2 and sr3 */
- mfsp %sr7,%r1 /* get userspace into sr3 */
- mtsp %r1,%sr3
- mtsp %r0,%sr2 /* get kernel space into sr2 */
- /* Load table start */
- ldil L%lws_table, %r1
- ldo R%lws_table(%r1), %r28 /* Scratch use of r28 */
- LDREGX %r20(%sr2,r28), %r21 /* Scratch use of r21 */
- /* Jump to lws, lws table pointers already relocated */
- be,n 0(%sr2,%r21)
- lws_exit_nosys:
- ldo -ENOSYS(%r0),%r21 /* set errno */
- /* Fall through: Return to userspace */
- lws_exit:
- #ifdef CONFIG_64BIT
- /* decide whether to reset the wide mode bit
- *
- * For a syscall, the W bit is stored in the lowest bit
- * of sp. Extract it and reset W if it is zero */
- extrd,u,*<> %r30,63,1,%r1
- rsm PSW_SM_W, %r0
- /* now reset the lowest bit of sp if it was set */
- xor %r30,%r1,%r30
- #endif
- be,n 0(%sr7, %r31)
-
- /***************************************************
- Implementing 32bit CAS as an atomic operation:
- %r26 - Address to examine
- %r25 - Old value to check (old)
- %r24 - New value to set (new)
- %r28 - Return prev through this register.
- %r21 - Kernel error code
- If debugging is DISabled:
- %r21 has the following meanings:
- EAGAIN - CAS is busy, ldcw failed, try again.
- EFAULT - Read or write failed.
- If debugging is enabled:
- EDEADLOCK - CAS called recursively.
- EAGAIN && r28 == 1 - CAS is busy. Lock contended.
- EAGAIN && r28 == 2 - CAS is busy. ldcw failed.
- EFAULT - Read or write failed.
- Scratch: r20, r28, r1
- ****************************************************/
- /* Do not enable LWS debugging */
- #define ENABLE_LWS_DEBUG 0
- /* ELF64 Process entry path */
- lws_compare_and_swap64:
- #ifdef CONFIG_64BIT
- b,n lws_compare_and_swap
- #else
- /* If we are not a 64-bit kernel, then we don't
- * have 64-bit input registers, and calling
- * the 64-bit LWS CAS returns ENOSYS.
- */
- b,n lws_exit_nosys
- #endif
- /* ELF32 Process entry path */
- lws_compare_and_swap32:
- #ifdef CONFIG_64BIT
- /* Clip all the input registers */
- depdi 0, 31, 32, %r26
- depdi 0, 31, 32, %r25
- depdi 0, 31, 32, %r24
- #endif
- lws_compare_and_swap:
- /* Load start of lock table */
- ldil L%lws_lock_start, %r20
- ldo R%lws_lock_start(%r20), %r28
- /* Extract four bits from r26 and hash lock (Bits 4-7) */
- extru %r26, 27, 4, %r20
- /* Find lock to use, the hash is either one of 0 to
- 15, multiplied by 16 (keep it 16-byte aligned)
- and add to the lock table offset. */
- shlw %r20, 4, %r20
- add %r20, %r28, %r20
- # if ENABLE_LWS_DEBUG
- /*
- DEBUG, check for deadlock!
- If the thread register values are the same
- then we were the one that locked it last and
- this is a recurisve call that will deadlock.
- We *must* giveup this call and fail.
- */
- ldw 4(%sr2,%r20), %r28 /* Load thread register */
- /* WARNING: If cr27 cycles to the same value we have problems */
- mfctl %cr27, %r21 /* Get current thread register */
- cmpb,<>,n %r21, %r28, cas_lock /* Called recursive? */
- b lws_exit /* Return error! */
- ldo -EDEADLOCK(%r0), %r21
- cas_lock:
- cmpb,=,n %r0, %r28, cas_nocontend /* Is nobody using it? */
- ldo 1(%r0), %r28 /* 1st case */
- b lws_exit /* Contended... */
- ldo -EAGAIN(%r0), %r21 /* Spin in userspace */
- cas_nocontend:
- # endif
- /* ENABLE_LWS_DEBUG */
- rsm PSW_SM_I, %r0 /* Disable interrupts */
- /* COW breaks can cause contention on UP systems */
- LDCW 0(%sr2,%r20), %r28 /* Try to acquire the lock */
- cmpb,<>,n %r0, %r28, cas_action /* Did we get it? */
- cas_wouldblock:
- ldo 2(%r0), %r28 /* 2nd case */
- ssm PSW_SM_I, %r0
- b lws_exit /* Contended... */
- ldo -EAGAIN(%r0), %r21 /* Spin in userspace */
- /*
- prev = *addr;
- if ( prev == old )
- *addr = new;
- return prev;
- */
- /* NOTES:
- This all works becuse intr_do_signal
- and schedule both check the return iasq
- and see that we are on the kernel page
- so this process is never scheduled off
- or is ever sent any signal of any sort,
- thus it is wholly atomic from usrspaces
- perspective
- */
- cas_action:
- #if defined CONFIG_SMP && ENABLE_LWS_DEBUG
- /* DEBUG */
- mfctl %cr27, %r1
- stw %r1, 4(%sr2,%r20)
- #endif
- /* The load and store could fail */
- 1: ldw,ma 0(%sr3,%r26), %r28
- sub,<> %r28, %r25, %r0
- 2: stw,ma %r24, 0(%sr3,%r26)
- /* Free lock */
- stw,ma %r20, 0(%sr2,%r20)
- #if ENABLE_LWS_DEBUG
- /* Clear thread register indicator */
- stw %r0, 4(%sr2,%r20)
- #endif
- /* Enable interrupts */
- ssm PSW_SM_I, %r0
- /* Return to userspace, set no error */
- b lws_exit
- copy %r0, %r21
- 3:
- /* Error occurred on load or store */
- /* Free lock */
- stw %r20, 0(%sr2,%r20)
- #if ENABLE_LWS_DEBUG
- stw %r0, 4(%sr2,%r20)
- #endif
- ssm PSW_SM_I, %r0
- b lws_exit
- ldo -EFAULT(%r0),%r21 /* set errno */
- nop
- nop
- nop
- nop
- /* Two exception table entries, one for the load,
- the other for the store. Either return -EFAULT.
- Each of the entries must be relocated. */
- ASM_EXCEPTIONTABLE_ENTRY(1b-linux_gateway_page, 3b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(2b-linux_gateway_page, 3b-linux_gateway_page)
- /***************************************************
- New CAS implementation which uses pointers and variable size
- information. The value pointed by old and new MUST NOT change
- while performing CAS. The lock only protect the value at %r26.
- %r26 - Address to examine
- %r25 - Pointer to the value to check (old)
- %r24 - Pointer to the value to set (new)
- %r23 - Size of the variable (0/1/2/3 for 8/16/32/64 bit)
- %r28 - Return non-zero on failure
- %r21 - Kernel error code
- %r21 has the following meanings:
- EAGAIN - CAS is busy, ldcw failed, try again.
- EFAULT - Read or write failed.
- Scratch: r20, r22, r28, r29, r1, fr4 (32bit for 64bit CAS only)
- ****************************************************/
- /* ELF32 Process entry path */
- lws_compare_and_swap_2:
- #ifdef CONFIG_64BIT
- /* Clip the input registers */
- depdi 0, 31, 32, %r26
- depdi 0, 31, 32, %r25
- depdi 0, 31, 32, %r24
- depdi 0, 31, 32, %r23
- #endif
- /* Check the validity of the size pointer */
- subi,>>= 4, %r23, %r0
- b,n lws_exit_nosys
- /* Jump to the functions which will load the old and new values into
- registers depending on the their size */
- shlw %r23, 2, %r29
- blr %r29, %r0
- nop
- /* 8bit load */
- 4: ldb 0(%sr3,%r25), %r25
- b cas2_lock_start
- 5: ldb 0(%sr3,%r24), %r24
- nop
- nop
- nop
- nop
- nop
- /* 16bit load */
- 6: ldh 0(%sr3,%r25), %r25
- b cas2_lock_start
- 7: ldh 0(%sr3,%r24), %r24
- nop
- nop
- nop
- nop
- nop
- /* 32bit load */
- 8: ldw 0(%sr3,%r25), %r25
- b cas2_lock_start
- 9: ldw 0(%sr3,%r24), %r24
- nop
- nop
- nop
- nop
- nop
- /* 64bit load */
- #ifdef CONFIG_64BIT
- 10: ldd 0(%sr3,%r25), %r25
- 11: ldd 0(%sr3,%r24), %r24
- #else
- /* Load new value into r22/r23 - high/low */
- 10: ldw 0(%sr3,%r25), %r22
- 11: ldw 4(%sr3,%r25), %r23
- /* Load new value into fr4 for atomic store later */
- 12: flddx 0(%sr3,%r24), %fr4
- #endif
- cas2_lock_start:
- /* Load start of lock table */
- ldil L%lws_lock_start, %r20
- ldo R%lws_lock_start(%r20), %r28
- /* Extract four bits from r26 and hash lock (Bits 4-7) */
- extru %r26, 27, 4, %r20
- /* Find lock to use, the hash is either one of 0 to
- 15, multiplied by 16 (keep it 16-byte aligned)
- and add to the lock table offset. */
- shlw %r20, 4, %r20
- add %r20, %r28, %r20
- rsm PSW_SM_I, %r0 /* Disable interrupts */
- /* COW breaks can cause contention on UP systems */
- LDCW 0(%sr2,%r20), %r28 /* Try to acquire the lock */
- cmpb,<>,n %r0, %r28, cas2_action /* Did we get it? */
- cas2_wouldblock:
- ldo 2(%r0), %r28 /* 2nd case */
- ssm PSW_SM_I, %r0
- b lws_exit /* Contended... */
- ldo -EAGAIN(%r0), %r21 /* Spin in userspace */
- /*
- prev = *addr;
- if ( prev == old )
- *addr = new;
- return prev;
- */
- /* NOTES:
- This all works becuse intr_do_signal
- and schedule both check the return iasq
- and see that we are on the kernel page
- so this process is never scheduled off
- or is ever sent any signal of any sort,
- thus it is wholly atomic from usrspaces
- perspective
- */
- cas2_action:
- /* Jump to the correct function */
- blr %r29, %r0
- /* Set %r28 as non-zero for now */
- ldo 1(%r0),%r28
- /* 8bit CAS */
- 13: ldb,ma 0(%sr3,%r26), %r29
- sub,= %r29, %r25, %r0
- b,n cas2_end
- 14: stb,ma %r24, 0(%sr3,%r26)
- b cas2_end
- copy %r0, %r28
- nop
- nop
- /* 16bit CAS */
- 15: ldh,ma 0(%sr3,%r26), %r29
- sub,= %r29, %r25, %r0
- b,n cas2_end
- 16: sth,ma %r24, 0(%sr3,%r26)
- b cas2_end
- copy %r0, %r28
- nop
- nop
- /* 32bit CAS */
- 17: ldw,ma 0(%sr3,%r26), %r29
- sub,= %r29, %r25, %r0
- b,n cas2_end
- 18: stw,ma %r24, 0(%sr3,%r26)
- b cas2_end
- copy %r0, %r28
- nop
- nop
- /* 64bit CAS */
- #ifdef CONFIG_64BIT
- 19: ldd,ma 0(%sr3,%r26), %r29
- sub,= %r29, %r25, %r0
- b,n cas2_end
- 20: std,ma %r24, 0(%sr3,%r26)
- copy %r0, %r28
- #else
- /* Compare first word */
- 19: ldw,ma 0(%sr3,%r26), %r29
- sub,= %r29, %r22, %r0
- b,n cas2_end
- /* Compare second word */
- 20: ldw,ma 4(%sr3,%r26), %r29
- sub,= %r29, %r23, %r0
- b,n cas2_end
- /* Perform the store */
- 21: fstdx %fr4, 0(%sr3,%r26)
- copy %r0, %r28
- #endif
- cas2_end:
- /* Free lock */
- stw,ma %r20, 0(%sr2,%r20)
- /* Enable interrupts */
- ssm PSW_SM_I, %r0
- /* Return to userspace, set no error */
- b lws_exit
- copy %r0, %r21
- 22:
- /* Error occurred on load or store */
- /* Free lock */
- stw %r20, 0(%sr2,%r20)
- ssm PSW_SM_I, %r0
- ldo 1(%r0),%r28
- b lws_exit
- ldo -EFAULT(%r0),%r21 /* set errno */
- nop
- nop
- nop
- /* Exception table entries, for the load and store, return EFAULT.
- Each of the entries must be relocated. */
- ASM_EXCEPTIONTABLE_ENTRY(4b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(5b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(6b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(7b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(8b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(9b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(10b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(11b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(13b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(14b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(15b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(16b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(17b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(18b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(19b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(20b-linux_gateway_page, 22b-linux_gateway_page)
- #ifndef CONFIG_64BIT
- ASM_EXCEPTIONTABLE_ENTRY(12b-linux_gateway_page, 22b-linux_gateway_page)
- ASM_EXCEPTIONTABLE_ENTRY(21b-linux_gateway_page, 22b-linux_gateway_page)
- #endif
- /* Make sure nothing else is placed on this page */
- .align PAGE_SIZE
- END(linux_gateway_page)
- ENTRY(end_linux_gateway_page)
- /* Relocate symbols assuming linux_gateway_page is mapped
- to virtual address 0x0 */
- #define LWS_ENTRY(_name_) ASM_ULONG_INSN (lws_##_name_ - linux_gateway_page)
- .section .rodata,"a"
- .align 8
- /* Light-weight-syscall table */
- /* Start of lws table. */
- ENTRY(lws_table)
- LWS_ENTRY(compare_and_swap32) /* 0 - ELF32 Atomic 32bit CAS */
- LWS_ENTRY(compare_and_swap64) /* 1 - ELF64 Atomic 32bit CAS */
- LWS_ENTRY(compare_and_swap_2) /* 2 - ELF32 Atomic 64bit CAS */
- END(lws_table)
- /* End of lws table */
- .align 8
- ENTRY(sys_call_table)
- #include "syscall_table.S"
- END(sys_call_table)
- #ifdef CONFIG_64BIT
- .align 8
- ENTRY(sys_call_table64)
- #define SYSCALL_TABLE_64BIT
- #include "syscall_table.S"
- END(sys_call_table64)
- #endif
- /*
- All light-weight-syscall atomic operations
- will use this set of locks
- NOTE: The lws_lock_start symbol must be
- at least 16-byte aligned for safe use
- with ldcw.
- */
- .section .data
- .align L1_CACHE_BYTES
- ENTRY(lws_lock_start)
- /* lws locks */
- .rept 16
- /* Keep locks aligned at 16-bytes */
- .word 1
- .word 0
- .word 0
- .word 0
- .endr
- END(lws_lock_start)
- .previous
- .end
|