774434bf33
Add _32 to filenames to make them 32 bit unique Signed-off-by: Sam Ravnborg <sam@ravnborg.org> Signed-off-by: David S. Miller <davem@davemloft.net>
116 lines
3.7 KiB
C
116 lines
3.7 KiB
C
#include <linux/kernel.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/types.h>
|
|
#include <asm/byteorder.h>
|
|
|
|
#define add_ssaaaa(sh, sl, ah, al, bh, bl) \
|
|
__asm__ ("addcc %r4,%5,%1\n\t" \
|
|
"addx %r2,%3,%0\n" \
|
|
: "=r" ((USItype)(sh)), \
|
|
"=&r" ((USItype)(sl)) \
|
|
: "%rJ" ((USItype)(ah)), \
|
|
"rI" ((USItype)(bh)), \
|
|
"%rJ" ((USItype)(al)), \
|
|
"rI" ((USItype)(bl)) \
|
|
: "cc")
|
|
#define sub_ddmmss(sh, sl, ah, al, bh, bl) \
|
|
__asm__ ("subcc %r4,%5,%1\n\t" \
|
|
"subx %r2,%3,%0\n" \
|
|
: "=r" ((USItype)(sh)), \
|
|
"=&r" ((USItype)(sl)) \
|
|
: "rJ" ((USItype)(ah)), \
|
|
"rI" ((USItype)(bh)), \
|
|
"rJ" ((USItype)(al)), \
|
|
"rI" ((USItype)(bl)) \
|
|
: "cc")
|
|
|
|
#define umul_ppmm(w1, w0, u, v) \
|
|
__asm__ ("! Inlined umul_ppmm\n\t" \
|
|
"wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n\t" \
|
|
"sra %3,31,%%g2 ! Don't move this insn\n\t" \
|
|
"and %2,%%g2,%%g2 ! Don't move this insn\n\t" \
|
|
"andcc %%g0,0,%%g1 ! Don't move this insn\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,%3,%%g1\n\t" \
|
|
"mulscc %%g1,0,%%g1\n\t" \
|
|
"add %%g1,%%g2,%0\n\t" \
|
|
"rd %%y,%1\n" \
|
|
: "=r" ((USItype)(w1)), \
|
|
"=r" ((USItype)(w0)) \
|
|
: "%rI" ((USItype)(u)), \
|
|
"r" ((USItype)(v)) \
|
|
: "%g1", "%g2", "cc")
|
|
|
|
/* It's quite necessary to add this much assembler for the sparc.
|
|
The default udiv_qrnnd (in C) is more than 10 times slower! */
|
|
#define udiv_qrnnd(q, r, n1, n0, d) \
|
|
__asm__ ("! Inlined udiv_qrnnd\n\t" \
|
|
"mov 32,%%g1\n\t" \
|
|
"subcc %1,%2,%%g0\n\t" \
|
|
"1: bcs 5f\n\t" \
|
|
"addxcc %0,%0,%0 ! shift n1n0 and a q-bit in lsb\n\t" \
|
|
"sub %1,%2,%1 ! this kills msb of n\n\t" \
|
|
"addx %1,%1,%1 ! so this can't give carry\n\t" \
|
|
"subcc %%g1,1,%%g1\n\t" \
|
|
"2: bne 1b\n\t" \
|
|
"subcc %1,%2,%%g0\n\t" \
|
|
"bcs 3f\n\t" \
|
|
"addxcc %0,%0,%0 ! shift n1n0 and a q-bit in lsb\n\t" \
|
|
"b 3f\n\t" \
|
|
"sub %1,%2,%1 ! this kills msb of n\n\t" \
|
|
"4: sub %1,%2,%1\n\t" \
|
|
"5: addxcc %1,%1,%1\n\t" \
|
|
"bcc 2b\n\t" \
|
|
"subcc %%g1,1,%%g1\n\t" \
|
|
"! Got carry from n. Subtract next step to cancel this carry.\n\t" \
|
|
"bne 4b\n\t" \
|
|
"addcc %0,%0,%0 ! shift n1n0 and a 0-bit in lsb\n\t" \
|
|
"sub %1,%2,%1\n\t" \
|
|
"3: xnor %0,0,%0\n\t" \
|
|
"! End of inline udiv_qrnnd\n" \
|
|
: "=&r" ((USItype)(q)), \
|
|
"=&r" ((USItype)(r)) \
|
|
: "r" ((USItype)(d)), \
|
|
"1" ((USItype)(n1)), \
|
|
"0" ((USItype)(n0)) : "%g1", "cc")
|
|
#define UDIV_NEEDS_NORMALIZATION 0
|
|
|
|
#define abort() \
|
|
return 0
|
|
|
|
#ifdef __BIG_ENDIAN
|
|
#define __BYTE_ORDER __BIG_ENDIAN
|
|
#else
|
|
#define __BYTE_ORDER __LITTLE_ENDIAN
|
|
#endif
|