8410904a77
* config/rl78/cmpsi2.S: Use function start and end macros. (__gcc_bcmp): New function. * config/rl78/lshrsi3.S: Use function start and end macros. * config/rl78/mulsi3.S: Add support for G10. (__mulqi3): New function for G10. * config/rl78/signbit.S: Use function start and end macros. * config/rl78/t-rl78 (LIB2ADD): Add bit-count.S, fpbit-sf.S and fpmath-sf.S. (LIB2FUNCS_EXCLUDE): Define. (LIB2FUNCS_ST): Define. * config/rl78/trampoline.S: Use function start and end macros. * config/rl78/vregs.h (START_FUNC): New macro. (START_ANOTHER_FUNC): New macro. (END_FUNC): New macro. (END_ANOTHER_FUNC): New macro. * config/rl78/bit-count.S: New file. Contains assembler implementations of the bit counting functions: ___clzhi2, __clzsi2, ctzhi2, ctzsi2, ffshi2, ffssi2, __partityhi2, __paritysi2, __popcounthi2 and __popcountsi2. * config/rl78/fpbit-sf.S: New file. Contains assembler implementationas of the math functions: __negsf2, __cmpsf2, __eqsf2, __nesf2, __lesf2, __ltsf2, __gesf2, gtsf2, __unordsf2, __fixsfsi, __fixunssfsi, __floatsisf and __floatunssisf. * config/rl78/fpmath-sf.S: New file. Contains assembler implementations of the math functions: __subsf3, __addsf3, __mulsf3 and __divsf3 From-SVN: r220162
324 lines
4.8 KiB
ArmAsm
324 lines
4.8 KiB
ArmAsm
; Copyright (C) 2011-2015 Free Software Foundation, Inc.
|
|
; Contributed by Red Hat.
|
|
;
|
|
; This file is free software; you can redistribute it and/or modify it
|
|
; under the terms of the GNU General Public License as published by the
|
|
; Free Software Foundation; either version 3, or (at your option) any
|
|
; later version.
|
|
;
|
|
; This file is distributed in the hope that it will be useful, but
|
|
; WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
; General Public License for more details.
|
|
;
|
|
; Under Section 7 of GPL version 3, you are granted additional
|
|
; permissions described in the GCC Runtime Library Exception, version
|
|
; 3.1, as published by the Free Software Foundation.
|
|
;
|
|
; You should have received a copy of the GNU General Public License and
|
|
; a copy of the GCC Runtime Library Exception along with this program;
|
|
; see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
; <http://www.gnu.org/licenses/>.
|
|
|
|
;; 32x32=32 multiply
|
|
|
|
#include "vregs.h"
|
|
|
|
;----------------------------------------------------------------------
|
|
|
|
; Register use:
|
|
; RB0 RB1 RB2
|
|
; AX op2L res32L res32H
|
|
; BC op2H (resH) op1
|
|
; DE count (resL-tmp)
|
|
; HL [sp+4]
|
|
|
|
; Register use (G10):
|
|
;
|
|
; AX op2L
|
|
; BC op2H
|
|
; DE count
|
|
; HL [sp+4]
|
|
; r8/r9 res32L
|
|
; r10/r11 (resH)
|
|
; r12/r13 (resL-tmp)
|
|
; r16/r17 res32H
|
|
; r18/r19 op1
|
|
|
|
START_FUNC ___mulsi3
|
|
;; A is at [sp+4]
|
|
;; B is at [sp+8]
|
|
;; result is in R8..R11
|
|
|
|
#ifdef __RL78_G10__
|
|
movw ax, r16
|
|
push ax
|
|
movw ax, r18
|
|
push ax
|
|
#else
|
|
sel rb2
|
|
push ax
|
|
push bc
|
|
sel rb0
|
|
#endif
|
|
clrw ax
|
|
movw r8, ax
|
|
movw r16, ax
|
|
|
|
movw ax, [sp+14]
|
|
cmpw ax, #0
|
|
bz $1f
|
|
cmpw ax, #0xffff
|
|
bnz $2f
|
|
movw ax, [sp+8]
|
|
#ifdef __RL78_G10__
|
|
push bc
|
|
movw bc, r8
|
|
xchw ax, bc
|
|
subw ax, bc
|
|
movw r8, ax
|
|
movw ax, bc
|
|
pop bc
|
|
#else
|
|
sel rb1
|
|
subw ax, r_0
|
|
sel rb0
|
|
#endif
|
|
br $1f
|
|
2:
|
|
movw bc, ax
|
|
movw ax, [sp+8]
|
|
cmpw ax, #0
|
|
skz
|
|
call !.Lmul_hi
|
|
1:
|
|
|
|
movw ax, [sp+10]
|
|
cmpw ax, #0
|
|
bz $1f
|
|
cmpw ax, #0xffff
|
|
bnz $2f
|
|
movw ax, [sp+12]
|
|
#ifdef __RL78_G10__
|
|
push bc
|
|
movw bc, r8
|
|
xchw ax, bc
|
|
subw ax, bc
|
|
movw r8, ax
|
|
movw ax, bc
|
|
pop bc
|
|
#else
|
|
sel rb1
|
|
subw ax, r_0
|
|
sel rb0
|
|
#endif
|
|
br $1f
|
|
2:
|
|
movw bc, ax
|
|
movw ax, [sp+12]
|
|
cmpw ax, #0
|
|
skz
|
|
call !.Lmul_hi
|
|
1:
|
|
|
|
movw ax, r8
|
|
movw r16, ax
|
|
clrw ax
|
|
movw r8, ax
|
|
|
|
;; now do R16:R8 += op1L * op2L
|
|
|
|
;; op1 is in AX.0 (needs to shrw)
|
|
;; op2 is in BC.2 and BC.1 (bc can shlw/rolcw)
|
|
;; res is in AX.2 and AX.1 (needs to addw)
|
|
|
|
movw ax, [sp+8]
|
|
movw r10, ax ; BC.1
|
|
movw ax, [sp+12]
|
|
|
|
cmpw ax, r10
|
|
bc $.Lmul_hisi_top
|
|
movw bc, r10
|
|
movw r10, ax
|
|
movw ax, bc
|
|
|
|
.Lmul_hisi_top:
|
|
movw bc, #0
|
|
|
|
.Lmul_hisi_loop:
|
|
shrw ax, 1
|
|
#ifdef __RL78_G10__
|
|
push ax
|
|
bnc $.Lmul_hisi_no_add_g10
|
|
movw ax, r8
|
|
addw ax, r10
|
|
movw r8, ax
|
|
sknc
|
|
incw r16
|
|
movw ax, r16
|
|
addw ax, r_2
|
|
movw r16, ax
|
|
.Lmul_hisi_no_add_g10:
|
|
movw ax, r10
|
|
shlw ax, 1
|
|
movw r10, ax
|
|
pop ax
|
|
#else
|
|
bnc $.Lmul_hisi_no_add
|
|
sel rb1
|
|
addw ax, bc
|
|
sel rb2
|
|
sknc
|
|
incw ax
|
|
addw ax, r_2
|
|
.Lmul_hisi_no_add:
|
|
sel rb1
|
|
shlw bc, 1
|
|
sel rb0
|
|
#endif
|
|
rolwc bc, 1
|
|
cmpw ax, #0
|
|
bz $.Lmul_hisi_done
|
|
|
|
shrw ax, 1
|
|
#ifdef __RL78_G10__
|
|
push ax
|
|
bnc $.Lmul_hisi_no_add2_g10
|
|
movw ax, r8
|
|
addw ax, r10
|
|
movw r8, ax
|
|
movw ax, r16
|
|
sknc
|
|
incw ax
|
|
addw ax, r_2
|
|
movw r16, ax
|
|
.Lmul_hisi_no_add2_g10:
|
|
movw ax, r10
|
|
shlw ax, 1
|
|
movw r10, ax
|
|
pop ax
|
|
#else
|
|
bnc $.Lmul_hisi_no_add2
|
|
sel rb1
|
|
addw ax, bc
|
|
sel rb2
|
|
sknc
|
|
incw ax
|
|
addw ax, r_2
|
|
.Lmul_hisi_no_add2:
|
|
sel rb1
|
|
shlw bc, 1
|
|
sel rb0
|
|
#endif
|
|
rolwc bc, 1
|
|
cmpw ax, #0
|
|
bnz $.Lmul_hisi_loop
|
|
|
|
.Lmul_hisi_done:
|
|
|
|
movw ax, r16
|
|
movw r10, ax
|
|
|
|
#ifdef __RL78_G10__
|
|
pop ax
|
|
movw r18, ax
|
|
pop ax
|
|
movw r16, ax
|
|
#else
|
|
sel rb2
|
|
pop bc
|
|
pop ax
|
|
sel rb0
|
|
#endif
|
|
|
|
ret
|
|
END_FUNC ___mulsi3
|
|
|
|
;----------------------------------------------------------------------
|
|
|
|
START_FUNC ___mulhi3
|
|
movw r8, #0
|
|
movw ax, [sp+6]
|
|
movw bc, ax
|
|
movw ax, [sp+4]
|
|
|
|
;; R8 += AX * BC
|
|
.Lmul_hi:
|
|
cmpw ax, bc
|
|
skc
|
|
xchw ax, bc
|
|
br $.Lmul_hi_loop
|
|
|
|
.Lmul_hi_top:
|
|
#ifdef __RL78_G10__
|
|
push ax
|
|
movw ax, r8
|
|
addw ax, r_2
|
|
movw r8, ax
|
|
pop ax
|
|
#else
|
|
sel rb1
|
|
addw ax, r_2
|
|
sel rb0
|
|
#endif
|
|
|
|
.Lmul_hi_no_add:
|
|
shlw bc, 1
|
|
.Lmul_hi_loop:
|
|
shrw ax, 1
|
|
bc $.Lmul_hi_top
|
|
cmpw ax, #0
|
|
bz $.Lmul_hi_done
|
|
|
|
shlw bc, 1
|
|
shrw ax, 1
|
|
bc $.Lmul_hi_top
|
|
cmpw ax, #0
|
|
bnz $.Lmul_hi_no_add
|
|
|
|
.Lmul_hi_done:
|
|
ret
|
|
END_FUNC ___mulhi3
|
|
|
|
;;; --------------------------------------
|
|
#ifdef __RL78_G10__
|
|
START_FUNC ___mulqi3
|
|
|
|
mov a, [sp+4]
|
|
mov r9, a
|
|
mov a, [sp+6]
|
|
mov r10, a
|
|
mov a, #9
|
|
mov r11, a
|
|
clrb a
|
|
mov r8, a
|
|
.L2:
|
|
cmp0 r10
|
|
skz
|
|
dec r11
|
|
sknz
|
|
ret
|
|
mov a, r10
|
|
and a, #1
|
|
mov r12, a
|
|
cmp0 r12
|
|
sknz
|
|
br !!.L3
|
|
mov a, r9
|
|
mov l, a
|
|
mov a, r8
|
|
add a, l
|
|
mov r8, a
|
|
.L3:
|
|
mov a, r9
|
|
add a, a
|
|
mov r9, a
|
|
mov a, r10
|
|
shr a, 1
|
|
mov r10, a
|
|
br !!.L2
|
|
|
|
END_FUNC ___mulqi3
|
|
#endif
|
|
|