123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116 |
- #include <linux/kernel.h>
- #include <linux/sched.h>
- #include <linux/types.h>
- #include <asm/byteorder.h>
- #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
- __asm__ ("addcc %r4,%5,%1\n\t" \
- "addx %r2,%3,%0\n" \
- : "=r" (sh), \
- "=&r" (sl) \
- : "%rJ" ((USItype)(ah)), \
- "rI" ((USItype)(bh)), \
- "%rJ" ((USItype)(al)), \
- "rI" ((USItype)(bl)) \
- : "cc")
- #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
- __asm__ ("subcc %r4,%5,%1\n\t" \
- "subx %r2,%3,%0\n" \
- : "=r" (sh), \
- "=&r" (sl) \
- : "rJ" ((USItype)(ah)), \
- "rI" ((USItype)(bh)), \
- "rJ" ((USItype)(al)), \
- "rI" ((USItype)(bl)) \
- : "cc")
- #define umul_ppmm(w1, w0, u, v) \
- __asm__ ("! Inlined umul_ppmm\n\t" \
- "wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n\t" \
- "sra %3,31,%%g2 ! Don't move this insn\n\t" \
- "and %2,%%g2,%%g2 ! Don't move this insn\n\t" \
- "andcc %%g0,0,%%g1 ! Don't move this insn\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,%3,%%g1\n\t" \
- "mulscc %%g1,0,%%g1\n\t" \
- "add %%g1,%%g2,%0\n\t" \
- "rd %%y,%1\n" \
- : "=r" (w1), \
- "=r" (w0) \
- : "%rI" ((USItype)(u)), \
- "r" ((USItype)(v)) \
- : "%g1", "%g2", "cc")
- /* It's quite necessary to add this much assembler for the sparc.
- The default udiv_qrnnd (in C) is more than 10 times slower! */
- #define udiv_qrnnd(q, r, n1, n0, d) \
- __asm__ ("! Inlined udiv_qrnnd\n\t" \
- "mov 32,%%g1\n\t" \
- "subcc %1,%2,%%g0\n\t" \
- "1: bcs 5f\n\t" \
- "addxcc %0,%0,%0 ! shift n1n0 and a q-bit in lsb\n\t" \
- "sub %1,%2,%1 ! this kills msb of n\n\t" \
- "addx %1,%1,%1 ! so this can't give carry\n\t" \
- "subcc %%g1,1,%%g1\n\t" \
- "2: bne 1b\n\t" \
- "subcc %1,%2,%%g0\n\t" \
- "bcs 3f\n\t" \
- "addxcc %0,%0,%0 ! shift n1n0 and a q-bit in lsb\n\t" \
- "b 3f\n\t" \
- "sub %1,%2,%1 ! this kills msb of n\n\t" \
- "4: sub %1,%2,%1\n\t" \
- "5: addxcc %1,%1,%1\n\t" \
- "bcc 2b\n\t" \
- "subcc %%g1,1,%%g1\n\t" \
- "! Got carry from n. Subtract next step to cancel this carry.\n\t" \
- "bne 4b\n\t" \
- "addcc %0,%0,%0 ! shift n1n0 and a 0-bit in lsb\n\t" \
- "sub %1,%2,%1\n\t" \
- "3: xnor %0,0,%0\n\t" \
- "! End of inline udiv_qrnnd\n" \
- : "=&r" (q), \
- "=&r" (r) \
- : "r" ((USItype)(d)), \
- "1" ((USItype)(n1)), \
- "0" ((USItype)(n0)) : "%g1", "cc")
- #define UDIV_NEEDS_NORMALIZATION 0
- #define abort() \
- return 0
- #ifdef __BIG_ENDIAN
- #define __BYTE_ORDER __BIG_ENDIAN
- #else
- #define __BYTE_ORDER __LITTLE_ENDIAN
- #endif
|