8b33430b53
libgcc/config/libbid/ChangeLog: 2020-02-04 Martin Liska <mliska@suse.cz> PR libgcc/92565 * bid_internal.h (handle_UF_128_rem): Remove unused variable. (handle_UF_128): Likewise.
2604 lines
79 KiB
C
2604 lines
79 KiB
C
/* Copyright (C) 2007-2020 Free Software Foundation, Inc.
|
|
|
|
This file is part of GCC.
|
|
|
|
GCC is free software; you can redistribute it and/or modify it under
|
|
the terms of the GNU General Public License as published by the Free
|
|
Software Foundation; either version 3, or (at your option) any later
|
|
version.
|
|
|
|
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
for more details.
|
|
|
|
Under Section 7 of GPL version 3, you are granted additional
|
|
permissions described in the GCC Runtime Library Exception, version
|
|
3.1, as published by the Free Software Foundation.
|
|
|
|
You should have received a copy of the GNU General Public License and
|
|
a copy of the GCC Runtime Library Exception along with this program;
|
|
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#ifndef __BIDECIMAL_H
|
|
#define __BIDECIMAL_H
|
|
|
|
#include "bid_conf.h"
|
|
#include "bid_functions.h"
|
|
|
|
#define __BID_INLINE__ static __inline
|
|
|
|
/*********************************************************************
|
|
*
|
|
* Logical Shift Macros
|
|
*
|
|
*********************************************************************/
|
|
|
|
#define __shr_128(Q, A, k) \
|
|
{ \
|
|
(Q).w[0] = (A).w[0] >> k; \
|
|
(Q).w[0] |= (A).w[1] << (64-k); \
|
|
(Q).w[1] = (A).w[1] >> k; \
|
|
}
|
|
|
|
#define __shr_128_long(Q, A, k) \
|
|
{ \
|
|
if((k)<64) { \
|
|
(Q).w[0] = (A).w[0] >> k; \
|
|
(Q).w[0] |= (A).w[1] << (64-k); \
|
|
(Q).w[1] = (A).w[1] >> k; \
|
|
} \
|
|
else { \
|
|
(Q).w[0] = (A).w[1]>>((k)-64); \
|
|
(Q).w[1] = 0; \
|
|
} \
|
|
}
|
|
|
|
#define __shl_128_long(Q, A, k) \
|
|
{ \
|
|
if((k)<64) { \
|
|
(Q).w[1] = (A).w[1] << k; \
|
|
(Q).w[1] |= (A).w[0] >> (64-k); \
|
|
(Q).w[0] = (A).w[0] << k; \
|
|
} \
|
|
else { \
|
|
(Q).w[1] = (A).w[0]<<((k)-64); \
|
|
(Q).w[0] = 0; \
|
|
} \
|
|
}
|
|
|
|
#define __low_64(Q) (Q).w[0]
|
|
/*********************************************************************
|
|
*
|
|
* String Macros
|
|
*
|
|
*********************************************************************/
|
|
#define tolower_macro(x) (((unsigned char)((x)-'A')<=('Z'-'A'))?((x)-'A'+'a'):(x))
|
|
/*********************************************************************
|
|
*
|
|
* Compare Macros
|
|
*
|
|
*********************************************************************/
|
|
// greater than
|
|
// return 0 if A<=B
|
|
// non-zero if A>B
|
|
#define __unsigned_compare_gt_128(A, B) \
|
|
((A.w[1]>B.w[1]) || ((A.w[1]==B.w[1]) && (A.w[0]>B.w[0])))
|
|
// greater-or-equal
|
|
#define __unsigned_compare_ge_128(A, B) \
|
|
((A.w[1]>B.w[1]) || ((A.w[1]==B.w[1]) && (A.w[0]>=B.w[0])))
|
|
#define __test_equal_128(A, B) (((A).w[1]==(B).w[1]) && ((A).w[0]==(B).w[0]))
|
|
// tighten exponent range
|
|
#define __tight_bin_range_128(bp, P, bin_expon) \
|
|
{ \
|
|
UINT64 M; \
|
|
M = 1; \
|
|
(bp) = (bin_expon); \
|
|
if((bp)<63) { \
|
|
M <<= ((bp)+1); \
|
|
if((P).w[0] >= M) (bp)++; } \
|
|
else if((bp)>64) { \
|
|
M <<= ((bp)+1-64); \
|
|
if(((P).w[1]>M) ||((P).w[1]==M && (P).w[0]))\
|
|
(bp)++; } \
|
|
else if((P).w[1]) (bp)++; \
|
|
}
|
|
/*********************************************************************
|
|
*
|
|
* Add/Subtract Macros
|
|
*
|
|
*********************************************************************/
|
|
// add 64-bit value to 128-bit
|
|
#define __add_128_64(R128, A128, B64) \
|
|
{ \
|
|
UINT64 R64H; \
|
|
R64H = (A128).w[1]; \
|
|
(R128).w[0] = (B64) + (A128).w[0]; \
|
|
if((R128).w[0] < (B64)) \
|
|
R64H ++; \
|
|
(R128).w[1] = R64H; \
|
|
}
|
|
// subtract 64-bit value from 128-bit
|
|
#define __sub_128_64(R128, A128, B64) \
|
|
{ \
|
|
UINT64 R64H; \
|
|
R64H = (A128).w[1]; \
|
|
if((A128).w[0] < (B64)) \
|
|
R64H --; \
|
|
(R128).w[1] = R64H; \
|
|
(R128).w[0] = (A128).w[0] - (B64); \
|
|
}
|
|
// add 128-bit value to 128-bit
|
|
// assume no carry-out
|
|
#define __add_128_128(R128, A128, B128) \
|
|
{ \
|
|
UINT128 Q128; \
|
|
Q128.w[1] = (A128).w[1]+(B128).w[1]; \
|
|
Q128.w[0] = (B128).w[0] + (A128).w[0]; \
|
|
if(Q128.w[0] < (B128).w[0]) \
|
|
Q128.w[1] ++; \
|
|
(R128).w[1] = Q128.w[1]; \
|
|
(R128).w[0] = Q128.w[0]; \
|
|
}
|
|
#define __sub_128_128(R128, A128, B128) \
|
|
{ \
|
|
UINT128 Q128; \
|
|
Q128.w[1] = (A128).w[1]-(B128).w[1]; \
|
|
Q128.w[0] = (A128).w[0] - (B128).w[0]; \
|
|
if((A128).w[0] < (B128).w[0]) \
|
|
Q128.w[1] --; \
|
|
(R128).w[1] = Q128.w[1]; \
|
|
(R128).w[0] = Q128.w[0]; \
|
|
}
|
|
#define __add_carry_out(S, CY, X, Y) \
|
|
{ \
|
|
UINT64 X1=X; \
|
|
S = X + Y; \
|
|
CY = (S<X1) ? 1 : 0; \
|
|
}
|
|
#define __add_carry_in_out(S, CY, X, Y, CI) \
|
|
{ \
|
|
UINT64 X1; \
|
|
X1 = X + CI; \
|
|
S = X1 + Y; \
|
|
CY = ((S<X1) || (X1<CI)) ? 1 : 0; \
|
|
}
|
|
#define __sub_borrow_out(S, CY, X, Y) \
|
|
{ \
|
|
UINT64 X1=X; \
|
|
S = X - Y; \
|
|
CY = (S>X1) ? 1 : 0; \
|
|
}
|
|
#define __sub_borrow_in_out(S, CY, X, Y, CI) \
|
|
{ \
|
|
UINT64 X1, X0=X; \
|
|
X1 = X - CI; \
|
|
S = X1 - Y; \
|
|
CY = ((S>X1) || (X1>X0)) ? 1 : 0; \
|
|
}
|
|
// increment C128 and check for rounding overflow:
|
|
// if (C_128) = 10^34 then (C_128) = 10^33 and increment the exponent
|
|
#define INCREMENT(C_128, exp) \
|
|
{ \
|
|
C_128.w[0]++; \
|
|
if (C_128.w[0] == 0) C_128.w[1]++; \
|
|
if (C_128.w[1] == 0x0001ed09bead87c0ull && \
|
|
C_128.w[0] == 0x378d8e6400000000ull) { \
|
|
exp++; \
|
|
C_128.w[1] = 0x0000314dc6448d93ull; \
|
|
C_128.w[0] = 0x38c15b0a00000000ull; \
|
|
} \
|
|
}
|
|
// decrement C128 and check for rounding underflow, but only at the
|
|
// boundary: if C_128 = 10^33 - 1 and exp > 0 then C_128 = 10^34 - 1
|
|
// and decrement the exponent
|
|
#define DECREMENT(C_128, exp) \
|
|
{ \
|
|
C_128.w[0]--; \
|
|
if (C_128.w[0] == 0xffffffffffffffffull) C_128.w[1]--; \
|
|
if (C_128.w[1] == 0x0000314dc6448d93ull && \
|
|
C_128.w[0] == 0x38c15b09ffffffffull && exp > 0) { \
|
|
exp--; \
|
|
C_128.w[1] = 0x0001ed09bead87c0ull; \
|
|
C_128.w[0] = 0x378d8e63ffffffffull; \
|
|
} \
|
|
}
|
|
|
|
/*********************************************************************
|
|
*
|
|
* Multiply Macros
|
|
*
|
|
*********************************************************************/
|
|
#define __mul_64x64_to_64(P64, CX, CY) (P64) = (CX) * (CY)
|
|
/***************************************
|
|
* Signed, Full 64x64-bit Multiply
|
|
***************************************/
|
|
#define __imul_64x64_to_128(P, CX, CY) \
|
|
{ \
|
|
UINT64 SX, SY; \
|
|
__mul_64x64_to_128(P, CX, CY); \
|
|
\
|
|
SX = ((SINT64)(CX))>>63; \
|
|
SY = ((SINT64)(CY))>>63; \
|
|
SX &= CY; SY &= CX; \
|
|
\
|
|
(P).w[1] = (P).w[1] - SX - SY; \
|
|
}
|
|
/***************************************
|
|
* Signed, Full 64x128-bit Multiply
|
|
***************************************/
|
|
#define __imul_64x128_full(Ph, Ql, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, QM2, QM; \
|
|
\
|
|
__imul_64x64_to_128(ALBH, (A), (B).w[1]); \
|
|
__imul_64x64_to_128(ALBL, (A), (B).w[0]); \
|
|
\
|
|
(Ql).w[0] = ALBL.w[0]; \
|
|
QM.w[0] = ALBL.w[1]; \
|
|
QM.w[1] = ((SINT64)ALBL.w[1])>>63; \
|
|
__add_128_128(QM2, ALBH, QM); \
|
|
(Ql).w[1] = QM2.w[0]; \
|
|
Ph = QM2.w[1]; \
|
|
}
|
|
/*****************************************************
|
|
* Unsigned Multiply Macros
|
|
*****************************************************/
|
|
// get full 64x64bit product
|
|
//
|
|
#define __mul_64x64_to_128(P, CX, CY) \
|
|
{ \
|
|
UINT64 CXH, CXL, CYH,CYL,PL,PH,PM,PM2;\
|
|
CXH = (CX) >> 32; \
|
|
CXL = (UINT32)(CX); \
|
|
CYH = (CY) >> 32; \
|
|
CYL = (UINT32)(CY); \
|
|
\
|
|
PM = CXH*CYL; \
|
|
PH = CXH*CYH; \
|
|
PL = CXL*CYL; \
|
|
PM2 = CXL*CYH; \
|
|
PH += (PM>>32); \
|
|
PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \
|
|
\
|
|
(P).w[1] = PH + (PM>>32); \
|
|
(P).w[0] = (PM<<32)+(UINT32)PL; \
|
|
}
|
|
// get full 64x64bit product
|
|
// Note:
|
|
// This macro is used for CX < 2^61, CY < 2^61
|
|
//
|
|
#define __mul_64x64_to_128_fast(P, CX, CY) \
|
|
{ \
|
|
UINT64 CXH, CXL, CYH, CYL, PL, PH, PM; \
|
|
CXH = (CX) >> 32; \
|
|
CXL = (UINT32)(CX); \
|
|
CYH = (CY) >> 32; \
|
|
CYL = (UINT32)(CY); \
|
|
\
|
|
PM = CXH*CYL; \
|
|
PL = CXL*CYL; \
|
|
PH = CXH*CYH; \
|
|
PM += CXL*CYH; \
|
|
PM += (PL>>32); \
|
|
\
|
|
(P).w[1] = PH + (PM>>32); \
|
|
(P).w[0] = (PM<<32)+(UINT32)PL; \
|
|
}
|
|
// used for CX< 2^60
|
|
#define __sqr64_fast(P, CX) \
|
|
{ \
|
|
UINT64 CXH, CXL, PL, PH, PM; \
|
|
CXH = (CX) >> 32; \
|
|
CXL = (UINT32)(CX); \
|
|
\
|
|
PM = CXH*CXL; \
|
|
PL = CXL*CXL; \
|
|
PH = CXH*CXH; \
|
|
PM += PM; \
|
|
PM += (PL>>32); \
|
|
\
|
|
(P).w[1] = PH + (PM>>32); \
|
|
(P).w[0] = (PM<<32)+(UINT32)PL; \
|
|
}
|
|
// get full 64x64bit product
|
|
// Note:
|
|
// This implementation is used for CX < 2^61, CY < 2^61
|
|
//
|
|
#define __mul_64x64_to_64_high_fast(P, CX, CY) \
|
|
{ \
|
|
UINT64 CXH, CXL, CYH, CYL, PL, PH, PM; \
|
|
CXH = (CX) >> 32; \
|
|
CXL = (UINT32)(CX); \
|
|
CYH = (CY) >> 32; \
|
|
CYL = (UINT32)(CY); \
|
|
\
|
|
PM = CXH*CYL; \
|
|
PL = CXL*CYL; \
|
|
PH = CXH*CYH; \
|
|
PM += CXL*CYH; \
|
|
PM += (PL>>32); \
|
|
\
|
|
(P) = PH + (PM>>32); \
|
|
}
|
|
// get full 64x64bit product
|
|
//
|
|
#define __mul_64x64_to_128_full(P, CX, CY) \
|
|
{ \
|
|
UINT64 CXH, CXL, CYH,CYL,PL,PH,PM,PM2;\
|
|
CXH = (CX) >> 32; \
|
|
CXL = (UINT32)(CX); \
|
|
CYH = (CY) >> 32; \
|
|
CYL = (UINT32)(CY); \
|
|
\
|
|
PM = CXH*CYL; \
|
|
PH = CXH*CYH; \
|
|
PL = CXL*CYL; \
|
|
PM2 = CXL*CYH; \
|
|
PH += (PM>>32); \
|
|
PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \
|
|
\
|
|
(P).w[1] = PH + (PM>>32); \
|
|
(P).w[0] = (PM<<32)+(UINT32)PL; \
|
|
}
|
|
#define __mul_128x128_high(Q, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, AHBL, AHBH, QM, QM2; \
|
|
\
|
|
__mul_64x64_to_128(ALBH, (A).w[0], (B).w[1]); \
|
|
__mul_64x64_to_128(AHBL, (B).w[0], (A).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A).w[0], (B).w[0]); \
|
|
__mul_64x64_to_128(AHBH, (A).w[1],(B).w[1]); \
|
|
\
|
|
__add_128_128(QM, ALBH, AHBL); \
|
|
__add_128_64(QM2, QM, ALBL.w[1]); \
|
|
__add_128_64((Q), AHBH, QM2.w[1]); \
|
|
}
|
|
#define __mul_128x128_full(Qh, Ql, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, AHBL, AHBH, QM, QM2; \
|
|
\
|
|
__mul_64x64_to_128(ALBH, (A).w[0], (B).w[1]); \
|
|
__mul_64x64_to_128(AHBL, (B).w[0], (A).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A).w[0], (B).w[0]); \
|
|
__mul_64x64_to_128(AHBH, (A).w[1],(B).w[1]); \
|
|
\
|
|
__add_128_128(QM, ALBH, AHBL); \
|
|
(Ql).w[0] = ALBL.w[0]; \
|
|
__add_128_64(QM2, QM, ALBL.w[1]); \
|
|
__add_128_64((Qh), AHBH, QM2.w[1]); \
|
|
(Ql).w[1] = QM2.w[0]; \
|
|
}
|
|
#define __mul_128x128_low(Ql, A, B) \
|
|
{ \
|
|
UINT128 ALBL; \
|
|
UINT64 QM64; \
|
|
\
|
|
__mul_64x64_to_128(ALBL, (A).w[0], (B).w[0]); \
|
|
QM64 = (B).w[0]*(A).w[1] + (A).w[0]*(B).w[1]; \
|
|
\
|
|
(Ql).w[0] = ALBL.w[0]; \
|
|
(Ql).w[1] = QM64 + ALBL.w[1]; \
|
|
}
|
|
#define __mul_64x128_low(Ql, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, QM2; \
|
|
__mul_64x64_to_128(ALBH, (A), (B).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A), (B).w[0]); \
|
|
(Ql).w[0] = ALBL.w[0]; \
|
|
__add_128_64(QM2, ALBH, ALBL.w[1]); \
|
|
(Ql).w[1] = QM2.w[0]; \
|
|
}
|
|
#define __mul_64x128_full(Ph, Ql, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, QM2; \
|
|
\
|
|
__mul_64x64_to_128(ALBH, (A), (B).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A), (B).w[0]); \
|
|
\
|
|
(Ql).w[0] = ALBL.w[0]; \
|
|
__add_128_64(QM2, ALBH, ALBL.w[1]); \
|
|
(Ql).w[1] = QM2.w[0]; \
|
|
Ph = QM2.w[1]; \
|
|
}
|
|
#define __mul_64x128_to_192(Q, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, QM2; \
|
|
\
|
|
__mul_64x64_to_128(ALBH, (A), (B).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A), (B).w[0]); \
|
|
\
|
|
(Q).w[0] = ALBL.w[0]; \
|
|
__add_128_64(QM2, ALBH, ALBL.w[1]); \
|
|
(Q).w[1] = QM2.w[0]; \
|
|
(Q).w[2] = QM2.w[1]; \
|
|
}
|
|
#define __mul_64x128_to192(Q, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, QM2; \
|
|
\
|
|
__mul_64x64_to_128(ALBH, (A), (B).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A), (B).w[0]); \
|
|
\
|
|
(Q).w[0] = ALBL.w[0]; \
|
|
__add_128_64(QM2, ALBH, ALBL.w[1]); \
|
|
(Q).w[1] = QM2.w[0]; \
|
|
(Q).w[2] = QM2.w[1]; \
|
|
}
|
|
#define __mul_128x128_to_256(P256, A, B) \
|
|
{ \
|
|
UINT128 Qll, Qlh; \
|
|
UINT64 Phl, Phh, CY1, CY2; \
|
|
\
|
|
__mul_64x128_full(Phl, Qll, A.w[0], B); \
|
|
__mul_64x128_full(Phh, Qlh, A.w[1], B); \
|
|
(P256).w[0] = Qll.w[0]; \
|
|
__add_carry_out((P256).w[1],CY1, Qlh.w[0], Qll.w[1]); \
|
|
__add_carry_in_out((P256).w[2],CY2, Qlh.w[1], Phl, CY1); \
|
|
(P256).w[3] = Phh + CY2; \
|
|
}
|
|
//
|
|
// For better performance, will check A.w[1] against 0,
|
|
// but not B.w[1]
|
|
// Use this macro accordingly
|
|
#define __mul_128x128_to_256_check_A(P256, A, B) \
|
|
{ \
|
|
UINT128 Qll, Qlh; \
|
|
UINT64 Phl, Phh, CY1, CY2; \
|
|
\
|
|
__mul_64x128_full(Phl, Qll, A.w[0], B); \
|
|
(P256).w[0] = Qll.w[0]; \
|
|
if(A.w[1]) { \
|
|
__mul_64x128_full(Phh, Qlh, A.w[1], B); \
|
|
__add_carry_out((P256).w[1],CY1, Qlh.w[0], Qll.w[1]); \
|
|
__add_carry_in_out((P256).w[2],CY2, Qlh.w[1], Phl, CY1); \
|
|
(P256).w[3] = Phh + CY2; } \
|
|
else { \
|
|
(P256).w[1] = Qll.w[1]; \
|
|
(P256).w[2] = Phl; \
|
|
(P256).w[3] = 0; } \
|
|
}
|
|
#define __mul_64x192_to_256(lP, lA, lB) \
|
|
{ \
|
|
UINT128 lP0,lP1,lP2; \
|
|
UINT64 lC; \
|
|
__mul_64x64_to_128(lP0, lA, (lB).w[0]); \
|
|
__mul_64x64_to_128(lP1, lA, (lB).w[1]); \
|
|
__mul_64x64_to_128(lP2, lA, (lB).w[2]); \
|
|
(lP).w[0] = lP0.w[0]; \
|
|
__add_carry_out((lP).w[1],lC,lP1.w[0],lP0.w[1]); \
|
|
__add_carry_in_out((lP).w[2],lC,lP2.w[0],lP1.w[1],lC); \
|
|
(lP).w[3] = lP2.w[1] + lC; \
|
|
}
|
|
#define __mul_64x256_to_320(P, A, B) \
|
|
{ \
|
|
UINT128 lP0,lP1,lP2,lP3; \
|
|
UINT64 lC; \
|
|
__mul_64x64_to_128(lP0, A, (B).w[0]); \
|
|
__mul_64x64_to_128(lP1, A, (B).w[1]); \
|
|
__mul_64x64_to_128(lP2, A, (B).w[2]); \
|
|
__mul_64x64_to_128(lP3, A, (B).w[3]); \
|
|
(P).w[0] = lP0.w[0]; \
|
|
__add_carry_out((P).w[1],lC,lP1.w[0],lP0.w[1]); \
|
|
__add_carry_in_out((P).w[2],lC,lP2.w[0],lP1.w[1],lC); \
|
|
__add_carry_in_out((P).w[3],lC,lP3.w[0],lP2.w[1],lC); \
|
|
(P).w[4] = lP3.w[1] + lC; \
|
|
}
|
|
#define __mul_192x192_to_384(P, A, B) \
|
|
{ \
|
|
UINT256 P0,P1,P2; \
|
|
UINT64 CY; \
|
|
__mul_64x192_to_256(P0, (A).w[0], B); \
|
|
__mul_64x192_to_256(P1, (A).w[1], B); \
|
|
__mul_64x192_to_256(P2, (A).w[2], B); \
|
|
(P).w[0] = P0.w[0]; \
|
|
__add_carry_out((P).w[1],CY,P1.w[0],P0.w[1]); \
|
|
__add_carry_in_out((P).w[2],CY,P1.w[1],P0.w[2],CY); \
|
|
__add_carry_in_out((P).w[3],CY,P1.w[2],P0.w[3],CY); \
|
|
(P).w[4] = P1.w[3] + CY; \
|
|
__add_carry_out((P).w[2],CY,P2.w[0],(P).w[2]); \
|
|
__add_carry_in_out((P).w[3],CY,P2.w[1],(P).w[3],CY); \
|
|
__add_carry_in_out((P).w[4],CY,P2.w[2],(P).w[4],CY); \
|
|
(P).w[5] = P2.w[3] + CY; \
|
|
}
|
|
#define __mul_64x320_to_384(P, A, B) \
|
|
{ \
|
|
UINT128 lP0,lP1,lP2,lP3,lP4; \
|
|
UINT64 lC; \
|
|
__mul_64x64_to_128(lP0, A, (B).w[0]); \
|
|
__mul_64x64_to_128(lP1, A, (B).w[1]); \
|
|
__mul_64x64_to_128(lP2, A, (B).w[2]); \
|
|
__mul_64x64_to_128(lP3, A, (B).w[3]); \
|
|
__mul_64x64_to_128(lP4, A, (B).w[4]); \
|
|
(P).w[0] = lP0.w[0]; \
|
|
__add_carry_out((P).w[1],lC,lP1.w[0],lP0.w[1]); \
|
|
__add_carry_in_out((P).w[2],lC,lP2.w[0],lP1.w[1],lC); \
|
|
__add_carry_in_out((P).w[3],lC,lP3.w[0],lP2.w[1],lC); \
|
|
__add_carry_in_out((P).w[4],lC,lP4.w[0],lP3.w[1],lC); \
|
|
(P).w[5] = lP4.w[1] + lC; \
|
|
}
|
|
// A*A
|
|
// Full 128x128-bit product
|
|
#define __sqr128_to_256(P256, A) \
|
|
{ \
|
|
UINT128 Qll, Qlh, Qhh; \
|
|
UINT64 TMP_C1, TMP_C2; \
|
|
\
|
|
__mul_64x64_to_128(Qhh, A.w[1], A.w[1]); \
|
|
__mul_64x64_to_128(Qlh, A.w[0], A.w[1]); \
|
|
Qhh.w[1] += (Qlh.w[1]>>63); \
|
|
Qlh.w[1] = (Qlh.w[1]+Qlh.w[1])|(Qlh.w[0]>>63); \
|
|
Qlh.w[0] += Qlh.w[0]; \
|
|
__mul_64x64_to_128(Qll, A.w[0], A.w[0]); \
|
|
\
|
|
__add_carry_out((P256).w[1],TMP_C1, Qlh.w[0], Qll.w[1]); \
|
|
(P256).w[0] = Qll.w[0]; \
|
|
__add_carry_in_out((P256).w[2],TMP_C2, Qlh.w[1], Qhh.w[0], TMP_C1); \
|
|
(P256).w[3] = Qhh.w[1]+TMP_C2; \
|
|
}
|
|
#define __mul_128x128_to_256_low_high(PQh, PQl, A, B) \
|
|
{ \
|
|
UINT128 Qll, Qlh; \
|
|
UINT64 Phl, Phh, C1, C2; \
|
|
\
|
|
__mul_64x128_full(Phl, Qll, A.w[0], B); \
|
|
__mul_64x128_full(Phh, Qlh, A.w[1], B); \
|
|
(PQl).w[0] = Qll.w[0]; \
|
|
__add_carry_out((PQl).w[1],C1, Qlh.w[0], Qll.w[1]); \
|
|
__add_carry_in_out((PQh).w[0],C2, Qlh.w[1], Phl, C1); \
|
|
(PQh).w[1] = Phh + C2; \
|
|
}
|
|
#define __mul_256x256_to_512(P, A, B) \
|
|
{ \
|
|
UINT512 P0,P1,P2,P3; \
|
|
UINT64 CY; \
|
|
__mul_64x256_to_320(P0, (A).w[0], B); \
|
|
__mul_64x256_to_320(P1, (A).w[1], B); \
|
|
__mul_64x256_to_320(P2, (A).w[2], B); \
|
|
__mul_64x256_to_320(P3, (A).w[3], B); \
|
|
(P).w[0] = P0.w[0]; \
|
|
__add_carry_out((P).w[1],CY,P1.w[0],P0.w[1]); \
|
|
__add_carry_in_out((P).w[2],CY,P1.w[1],P0.w[2],CY); \
|
|
__add_carry_in_out((P).w[3],CY,P1.w[2],P0.w[3],CY); \
|
|
__add_carry_in_out((P).w[4],CY,P1.w[3],P0.w[4],CY); \
|
|
(P).w[5] = P1.w[4] + CY; \
|
|
__add_carry_out((P).w[2],CY,P2.w[0],(P).w[2]); \
|
|
__add_carry_in_out((P).w[3],CY,P2.w[1],(P).w[3],CY); \
|
|
__add_carry_in_out((P).w[4],CY,P2.w[2],(P).w[4],CY); \
|
|
__add_carry_in_out((P).w[5],CY,P2.w[3],(P).w[5],CY); \
|
|
(P).w[6] = P2.w[4] + CY; \
|
|
__add_carry_out((P).w[3],CY,P3.w[0],(P).w[3]); \
|
|
__add_carry_in_out((P).w[4],CY,P3.w[1],(P).w[4],CY); \
|
|
__add_carry_in_out((P).w[5],CY,P3.w[2],(P).w[5],CY); \
|
|
__add_carry_in_out((P).w[6],CY,P3.w[3],(P).w[6],CY); \
|
|
(P).w[7] = P3.w[4] + CY; \
|
|
}
|
|
#define __mul_192x256_to_448(P, A, B) \
|
|
{ \
|
|
UINT512 P0,P1,P2; \
|
|
UINT64 CY; \
|
|
__mul_64x256_to_320(P0, (A).w[0], B); \
|
|
__mul_64x256_to_320(P1, (A).w[1], B); \
|
|
__mul_64x256_to_320(P2, (A).w[2], B); \
|
|
(P).w[0] = P0.w[0]; \
|
|
__add_carry_out((P).w[1],CY,P1.w[0],P0.w[1]); \
|
|
__add_carry_in_out((P).w[2],CY,P1.w[1],P0.w[2],CY); \
|
|
__add_carry_in_out((P).w[3],CY,P1.w[2],P0.w[3],CY); \
|
|
__add_carry_in_out((P).w[4],CY,P1.w[3],P0.w[4],CY); \
|
|
(P).w[5] = P1.w[4] + CY; \
|
|
__add_carry_out((P).w[2],CY,P2.w[0],(P).w[2]); \
|
|
__add_carry_in_out((P).w[3],CY,P2.w[1],(P).w[3],CY); \
|
|
__add_carry_in_out((P).w[4],CY,P2.w[2],(P).w[4],CY); \
|
|
__add_carry_in_out((P).w[5],CY,P2.w[3],(P).w[5],CY); \
|
|
(P).w[6] = P2.w[4] + CY; \
|
|
}
|
|
#define __mul_320x320_to_640(P, A, B) \
|
|
{ \
|
|
UINT512 P0,P1,P2,P3; \
|
|
UINT64 CY; \
|
|
__mul_256x256_to_512((P), (A), B); \
|
|
__mul_64x256_to_320(P1, (A).w[4], B); \
|
|
__mul_64x256_to_320(P2, (B).w[4], A); \
|
|
__mul_64x64_to_128(P3, (A).w[4], (B).w[4]); \
|
|
__add_carry_out((P0).w[0],CY,P1.w[0],P2.w[0]); \
|
|
__add_carry_in_out((P0).w[1],CY,P1.w[1],P2.w[1],CY); \
|
|
__add_carry_in_out((P0).w[2],CY,P1.w[2],P2.w[2],CY); \
|
|
__add_carry_in_out((P0).w[3],CY,P1.w[3],P2.w[3],CY); \
|
|
__add_carry_in_out((P0).w[4],CY,P1.w[4],P2.w[4],CY); \
|
|
P3.w[1] += CY; \
|
|
__add_carry_out((P).w[4],CY,(P).w[4],P0.w[0]); \
|
|
__add_carry_in_out((P).w[5],CY,(P).w[5],P0.w[1],CY); \
|
|
__add_carry_in_out((P).w[6],CY,(P).w[6],P0.w[2],CY); \
|
|
__add_carry_in_out((P).w[7],CY,(P).w[7],P0.w[3],CY); \
|
|
__add_carry_in_out((P).w[8],CY,P3.w[0],P0.w[4],CY); \
|
|
(P).w[9] = P3.w[1] + CY; \
|
|
}
|
|
#define __mul_384x384_to_768(P, A, B) \
|
|
{ \
|
|
UINT512 P0,P1,P2,P3; \
|
|
UINT64 CY; \
|
|
__mul_320x320_to_640((P), (A), B); \
|
|
__mul_64x320_to_384(P1, (A).w[5], B); \
|
|
__mul_64x320_to_384(P2, (B).w[5], A); \
|
|
__mul_64x64_to_128(P3, (A).w[5], (B).w[5]); \
|
|
__add_carry_out((P0).w[0],CY,P1.w[0],P2.w[0]); \
|
|
__add_carry_in_out((P0).w[1],CY,P1.w[1],P2.w[1],CY); \
|
|
__add_carry_in_out((P0).w[2],CY,P1.w[2],P2.w[2],CY); \
|
|
__add_carry_in_out((P0).w[3],CY,P1.w[3],P2.w[3],CY); \
|
|
__add_carry_in_out((P0).w[4],CY,P1.w[4],P2.w[4],CY); \
|
|
__add_carry_in_out((P0).w[5],CY,P1.w[5],P2.w[5],CY); \
|
|
P3.w[1] += CY; \
|
|
__add_carry_out((P).w[5],CY,(P).w[5],P0.w[0]); \
|
|
__add_carry_in_out((P).w[6],CY,(P).w[6],P0.w[1],CY); \
|
|
__add_carry_in_out((P).w[7],CY,(P).w[7],P0.w[2],CY); \
|
|
__add_carry_in_out((P).w[8],CY,(P).w[8],P0.w[3],CY); \
|
|
__add_carry_in_out((P).w[9],CY,(P).w[9],P0.w[4],CY); \
|
|
__add_carry_in_out((P).w[10],CY,P3.w[0],P0.w[5],CY); \
|
|
(P).w[11] = P3.w[1] + CY; \
|
|
}
|
|
#define __mul_64x128_short(Ql, A, B) \
|
|
{ \
|
|
UINT64 ALBH_L; \
|
|
\
|
|
__mul_64x64_to_64(ALBH_L, (A),(B).w[1]); \
|
|
__mul_64x64_to_128((Ql), (A), (B).w[0]); \
|
|
\
|
|
(Ql).w[1] += ALBH_L; \
|
|
}
|
|
#define __scale128_10(D,_TMP) \
|
|
{ \
|
|
UINT128 _TMP2,_TMP8; \
|
|
_TMP2.w[1] = (_TMP.w[1]<<1)|(_TMP.w[0]>>63); \
|
|
_TMP2.w[0] = _TMP.w[0]<<1; \
|
|
_TMP8.w[1] = (_TMP.w[1]<<3)|(_TMP.w[0]>>61); \
|
|
_TMP8.w[0] = _TMP.w[0]<<3; \
|
|
__add_128_128(D, _TMP2, _TMP8); \
|
|
}
|
|
// 64x64-bit product
|
|
#define __mul_64x64_to_128MACH(P128, CX64, CY64) \
|
|
{ \
|
|
UINT64 CXH,CXL,CYH,CYL,PL,PH,PM,PM2; \
|
|
CXH = (CX64) >> 32; \
|
|
CXL = (UINT32)(CX64); \
|
|
CYH = (CY64) >> 32; \
|
|
CYL = (UINT32)(CY64); \
|
|
PM = CXH*CYL; \
|
|
PH = CXH*CYH; \
|
|
PL = CXL*CYL; \
|
|
PM2 = CXL*CYH; \
|
|
PH += (PM>>32); \
|
|
PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \
|
|
(P128).w[1] = PH + (PM>>32); \
|
|
(P128).w[0] = (PM<<32)+(UINT32)PL; \
|
|
}
|
|
// 64x64-bit product
|
|
#define __mul_64x64_to_128HIGH(P64, CX64, CY64) \
|
|
{ \
|
|
UINT64 CXH,CXL,CYH,CYL,PL,PH,PM,PM2; \
|
|
CXH = (CX64) >> 32; \
|
|
CXL = (UINT32)(CX64); \
|
|
CYH = (CY64) >> 32; \
|
|
CYL = (UINT32)(CY64); \
|
|
PM = CXH*CYL; \
|
|
PH = CXH*CYH; \
|
|
PL = CXL*CYL; \
|
|
PM2 = CXL*CYH; \
|
|
PH += (PM>>32); \
|
|
PM = (UINT64)((UINT32)PM)+PM2+(PL>>32); \
|
|
P64 = PH + (PM>>32); \
|
|
}
|
|
#define __mul_128x64_to_128(Q128, A64, B128) \
|
|
{ \
|
|
UINT64 ALBH_L; \
|
|
ALBH_L = (A64) * (B128).w[1]; \
|
|
__mul_64x64_to_128MACH((Q128), (A64), (B128).w[0]); \
|
|
(Q128).w[1] += ALBH_L; \
|
|
}
|
|
// might simplify by calculating just QM2.w[0]
|
|
#define __mul_64x128_to_128(Ql, A, B) \
|
|
{ \
|
|
UINT128 ALBL, ALBH, QM2; \
|
|
__mul_64x64_to_128(ALBH, (A), (B).w[1]); \
|
|
__mul_64x64_to_128(ALBL, (A), (B).w[0]); \
|
|
(Ql).w[0] = ALBL.w[0]; \
|
|
__add_128_64(QM2, ALBH, ALBL.w[1]); \
|
|
(Ql).w[1] = QM2.w[0]; \
|
|
}
|
|
/*********************************************************************
|
|
*
|
|
* BID Pack/Unpack Macros
|
|
*
|
|
*********************************************************************/
|
|
/////////////////////////////////////////
|
|
// BID64 definitions
|
|
////////////////////////////////////////
|
|
#define DECIMAL_MAX_EXPON_64 767
|
|
#define DECIMAL_EXPONENT_BIAS 398
|
|
#define MAX_FORMAT_DIGITS 16
|
|
/////////////////////////////////////////
|
|
// BID128 definitions
|
|
////////////////////////////////////////
|
|
#define DECIMAL_MAX_EXPON_128 12287
|
|
#define DECIMAL_EXPONENT_BIAS_128 6176
|
|
#define MAX_FORMAT_DIGITS_128 34
|
|
/////////////////////////////////////////
|
|
// BID32 definitions
|
|
////////////////////////////////////////
|
|
#define DECIMAL_MAX_EXPON_32 191
|
|
#define DECIMAL_EXPONENT_BIAS_32 101
|
|
#define MAX_FORMAT_DIGITS_32 7
|
|
////////////////////////////////////////
|
|
// Constant Definitions
|
|
///////////////////////////////////////
|
|
#define SPECIAL_ENCODING_MASK64 0x6000000000000000ull
|
|
#define INFINITY_MASK64 0x7800000000000000ull
|
|
#define SINFINITY_MASK64 0xf800000000000000ull
|
|
#define SSNAN_MASK64 0xfc00000000000000ull
|
|
#define NAN_MASK64 0x7c00000000000000ull
|
|
#define SNAN_MASK64 0x7e00000000000000ull
|
|
#define QUIET_MASK64 0xfdffffffffffffffull
|
|
#define LARGE_COEFF_MASK64 0x0007ffffffffffffull
|
|
#define LARGE_COEFF_HIGH_BIT64 0x0020000000000000ull
|
|
#define SMALL_COEFF_MASK64 0x001fffffffffffffull
|
|
#define EXPONENT_MASK64 0x3ff
|
|
#define EXPONENT_SHIFT_LARGE64 51
|
|
#define EXPONENT_SHIFT_SMALL64 53
|
|
#define LARGEST_BID64 0x77fb86f26fc0ffffull
|
|
#define SMALLEST_BID64 0xf7fb86f26fc0ffffull
|
|
#define SMALL_COEFF_MASK128 0x0001ffffffffffffull
|
|
#define LARGE_COEFF_MASK128 0x00007fffffffffffull
|
|
#define EXPONENT_MASK128 0x3fff
|
|
#define LARGEST_BID128_HIGH 0x5fffed09bead87c0ull
|
|
#define LARGEST_BID128_LOW 0x378d8e63ffffffffull
|
|
#define SPECIAL_ENCODING_MASK32 0x60000000ul
|
|
#define INFINITY_MASK32 0x78000000ul
|
|
#define LARGE_COEFF_MASK32 0x007ffffful
|
|
#define LARGE_COEFF_HIGH_BIT32 0x00800000ul
|
|
#define SMALL_COEFF_MASK32 0x001ffffful
|
|
#define EXPONENT_MASK32 0xff
|
|
#define LARGEST_BID32 0x77f8967f
|
|
#define NAN_MASK32 0x7c000000
|
|
#define SNAN_MASK32 0x7e000000
|
|
#define MASK_BINARY_EXPONENT 0x7ff0000000000000ull
|
|
#define BINARY_EXPONENT_BIAS 0x3ff
|
|
#define UPPER_EXPON_LIMIT 51
|
|
// data needed for BID pack/unpack macros
|
|
extern UINT64 round_const_table[][19];
|
|
extern UINT128 reciprocals10_128[];
|
|
extern int recip_scale[];
|
|
extern UINT128 power10_table_128[];
|
|
extern int estimate_decimal_digits[];
|
|
extern int estimate_bin_expon[];
|
|
extern UINT64 power10_index_binexp[];
|
|
extern int short_recip_scale[];
|
|
extern UINT64 reciprocals10_64[];
|
|
extern UINT128 power10_index_binexp_128[];
|
|
extern UINT128 round_const_table_128[][36];
|
|
|
|
|
|
//////////////////////////////////////////////
|
|
// Status Flag Handling
|
|
/////////////////////////////////////////////
|
|
#define __set_status_flags(fpsc, status) *(fpsc) |= status
|
|
#define is_inexact(fpsc) ((*(fpsc))&INEXACT_EXCEPTION)
|
|
|
|
__BID_INLINE__ UINT64
|
|
unpack_BID64 (UINT64 * psign_x, int *pexponent_x,
|
|
UINT64 * pcoefficient_x, UINT64 x) {
|
|
UINT64 tmp, coeff;
|
|
|
|
*psign_x = x & 0x8000000000000000ull;
|
|
|
|
if ((x & SPECIAL_ENCODING_MASK64) == SPECIAL_ENCODING_MASK64) {
|
|
// special encodings
|
|
// coefficient
|
|
coeff = (x & LARGE_COEFF_MASK64) | LARGE_COEFF_HIGH_BIT64;
|
|
|
|
if ((x & INFINITY_MASK64) == INFINITY_MASK64) {
|
|
*pexponent_x = 0;
|
|
*pcoefficient_x = x & 0xfe03ffffffffffffull;
|
|
if ((x & 0x0003ffffffffffffull) >= 1000000000000000ull)
|
|
*pcoefficient_x = x & 0xfe00000000000000ull;
|
|
if ((x & NAN_MASK64) == INFINITY_MASK64)
|
|
*pcoefficient_x = x & SINFINITY_MASK64;
|
|
return 0; // NaN or Infinity
|
|
}
|
|
// check for non-canonical values
|
|
if (coeff >= 10000000000000000ull)
|
|
coeff = 0;
|
|
*pcoefficient_x = coeff;
|
|
// get exponent
|
|
tmp = x >> EXPONENT_SHIFT_LARGE64;
|
|
*pexponent_x = (int) (tmp & EXPONENT_MASK64);
|
|
return coeff;
|
|
}
|
|
// exponent
|
|
tmp = x >> EXPONENT_SHIFT_SMALL64;
|
|
*pexponent_x = (int) (tmp & EXPONENT_MASK64);
|
|
// coefficient
|
|
*pcoefficient_x = (x & SMALL_COEFF_MASK64);
|
|
|
|
return *pcoefficient_x;
|
|
}
|
|
|
|
//
|
|
// BID64 pack macro (general form)
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
get_BID64 (UINT64 sgn, int expon, UINT64 coeff, int rmode,
|
|
unsigned *fpsc) {
|
|
UINT128 Stemp, Q_low;
|
|
UINT64 QH, r, mask, C64, remainder_h, CY, carry;
|
|
int extra_digits, amount, amount2;
|
|
unsigned status;
|
|
|
|
if (coeff > 9999999999999999ull) {
|
|
expon++;
|
|
coeff = 1000000000000000ull;
|
|
}
|
|
// check for possible underflow/overflow
|
|
if (((unsigned) expon) >= 3 * 256) {
|
|
if (expon < 0) {
|
|
// underflow
|
|
if (expon + MAX_FORMAT_DIGITS < 0) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc,
|
|
UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == ROUNDING_DOWN && sgn)
|
|
return 0x8000000000000001ull;
|
|
if (rmode == ROUNDING_UP && !sgn)
|
|
return 1ull;
|
|
#endif
|
|
#endif
|
|
// result is 0
|
|
return sgn;
|
|
}
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (sgn && (unsigned) (rmode - 1) < 2)
|
|
rmode = 3 - rmode;
|
|
#endif
|
|
#endif
|
|
// get digits to be shifted out
|
|
extra_digits = -expon;
|
|
coeff += round_const_table[rmode][extra_digits];
|
|
|
|
// get coeff*(2^M[extra_digits])/10^extra_digits
|
|
__mul_64x128_full (QH, Q_low, coeff,
|
|
reciprocals10_128[extra_digits]);
|
|
|
|
// now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128
|
|
amount = recip_scale[extra_digits];
|
|
|
|
C64 = QH >> amount;
|
|
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == 0) //ROUNDING_TO_NEAREST
|
|
#endif
|
|
if (C64 & 1) {
|
|
// check whether fractional part of initial_P/10^extra_digits is exactly .5
|
|
|
|
// get remainder
|
|
amount2 = 64 - amount;
|
|
remainder_h = 0;
|
|
remainder_h--;
|
|
remainder_h >>= amount2;
|
|
remainder_h = remainder_h & QH;
|
|
|
|
if (!remainder_h
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0]))) {
|
|
C64--;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef SET_STATUS_FLAGS
|
|
|
|
if (is_inexact (fpsc))
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION);
|
|
else {
|
|
status = INEXACT_EXCEPTION;
|
|
// get remainder
|
|
remainder_h = QH << (64 - amount);
|
|
|
|
switch (rmode) {
|
|
case ROUNDING_TO_NEAREST:
|
|
case ROUNDING_TIES_AWAY:
|
|
// test whether fractional part is 0
|
|
if (remainder_h == 0x8000000000000000ull
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
case ROUNDING_DOWN:
|
|
case ROUNDING_TO_ZERO:
|
|
if (!remainder_h
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
default:
|
|
// round up
|
|
__add_carry_out (Stemp.w[0], CY, Q_low.w[0],
|
|
reciprocals10_128[extra_digits].w[0]);
|
|
__add_carry_in_out (Stemp.w[1], carry, Q_low.w[1],
|
|
reciprocals10_128[extra_digits].w[1], CY);
|
|
if ((remainder_h >> (64 - amount)) + carry >=
|
|
(((UINT64) 1) << amount))
|
|
status = EXACT_STATUS;
|
|
}
|
|
|
|
if (status != EXACT_STATUS)
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status);
|
|
}
|
|
|
|
#endif
|
|
|
|
return sgn | C64;
|
|
}
|
|
while (coeff < 1000000000000000ull && expon >= 3 * 256) {
|
|
expon--;
|
|
coeff = (coeff << 3) + (coeff << 1);
|
|
}
|
|
if (expon > DECIMAL_MAX_EXPON_64) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
// overflow
|
|
r = sgn | INFINITY_MASK64;
|
|
switch (rmode) {
|
|
case ROUNDING_DOWN:
|
|
if (!sgn)
|
|
r = LARGEST_BID64;
|
|
break;
|
|
case ROUNDING_TO_ZERO:
|
|
r = sgn | LARGEST_BID64;
|
|
break;
|
|
case ROUNDING_UP:
|
|
// round up
|
|
if (sgn)
|
|
r = SMALLEST_BID64;
|
|
}
|
|
return r;
|
|
}
|
|
}
|
|
|
|
mask = 1;
|
|
mask <<= EXPONENT_SHIFT_SMALL64;
|
|
|
|
// check whether coefficient fits in 10*5+3 bits
|
|
if (coeff < mask) {
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (coeff | sgn);
|
|
return r;
|
|
}
|
|
// special format
|
|
|
|
// eliminate the case coeff==10^16 after rounding
|
|
if (coeff == 10000000000000000ull) {
|
|
r = expon + 1;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (1000000000000000ull | sgn);
|
|
return r;
|
|
}
|
|
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_LARGE64;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK64);
|
|
// add coeff, without leading bits
|
|
mask = (mask >> 2) - 1;
|
|
coeff &= mask;
|
|
r |= coeff;
|
|
|
|
return r;
|
|
}
|
|
|
|
|
|
|
|
|
|
//
|
|
// No overflow/underflow checking
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
fast_get_BID64 (UINT64 sgn, int expon, UINT64 coeff) {
|
|
UINT64 r, mask;
|
|
|
|
mask = 1;
|
|
mask <<= EXPONENT_SHIFT_SMALL64;
|
|
|
|
// check whether coefficient fits in 10*5+3 bits
|
|
if (coeff < mask) {
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (coeff | sgn);
|
|
return r;
|
|
}
|
|
// special format
|
|
|
|
// eliminate the case coeff==10^16 after rounding
|
|
if (coeff == 10000000000000000ull) {
|
|
r = expon + 1;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (1000000000000000ull | sgn);
|
|
return r;
|
|
}
|
|
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_LARGE64;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK64);
|
|
// add coeff, without leading bits
|
|
mask = (mask >> 2) - 1;
|
|
coeff &= mask;
|
|
r |= coeff;
|
|
|
|
return r;
|
|
}
|
|
|
|
|
|
//
|
|
// no underflow checking
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
fast_get_BID64_check_OF (UINT64 sgn, int expon, UINT64 coeff, int rmode,
|
|
unsigned *fpsc) {
|
|
UINT64 r, mask;
|
|
|
|
if (((unsigned) expon) >= 3 * 256 - 1) {
|
|
if ((expon == 3 * 256 - 1) && coeff == 10000000000000000ull) {
|
|
expon = 3 * 256;
|
|
coeff = 1000000000000000ull;
|
|
}
|
|
|
|
if (((unsigned) expon) >= 3 * 256) {
|
|
while (coeff < 1000000000000000ull && expon >= 3 * 256) {
|
|
expon--;
|
|
coeff = (coeff << 3) + (coeff << 1);
|
|
}
|
|
if (expon > DECIMAL_MAX_EXPON_64) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc,
|
|
OVERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
// overflow
|
|
r = sgn | INFINITY_MASK64;
|
|
switch (rmode) {
|
|
case ROUNDING_DOWN:
|
|
if (!sgn)
|
|
r = LARGEST_BID64;
|
|
break;
|
|
case ROUNDING_TO_ZERO:
|
|
r = sgn | LARGEST_BID64;
|
|
break;
|
|
case ROUNDING_UP:
|
|
// round up
|
|
if (sgn)
|
|
r = SMALLEST_BID64;
|
|
}
|
|
return r;
|
|
}
|
|
}
|
|
}
|
|
|
|
mask = 1;
|
|
mask <<= EXPONENT_SHIFT_SMALL64;
|
|
|
|
// check whether coefficient fits in 10*5+3 bits
|
|
if (coeff < mask) {
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (coeff | sgn);
|
|
return r;
|
|
}
|
|
// special format
|
|
|
|
// eliminate the case coeff==10^16 after rounding
|
|
if (coeff == 10000000000000000ull) {
|
|
r = expon + 1;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (1000000000000000ull | sgn);
|
|
return r;
|
|
}
|
|
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_LARGE64;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK64);
|
|
// add coeff, without leading bits
|
|
mask = (mask >> 2) - 1;
|
|
coeff &= mask;
|
|
r |= coeff;
|
|
|
|
return r;
|
|
}
|
|
|
|
|
|
//
|
|
// No overflow/underflow checking
|
|
// or checking for coefficients equal to 10^16 (after rounding)
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
very_fast_get_BID64 (UINT64 sgn, int expon, UINT64 coeff) {
|
|
UINT64 r, mask;
|
|
|
|
mask = 1;
|
|
mask <<= EXPONENT_SHIFT_SMALL64;
|
|
|
|
// check whether coefficient fits in 10*5+3 bits
|
|
if (coeff < mask) {
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (coeff | sgn);
|
|
return r;
|
|
}
|
|
// special format
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_LARGE64;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK64);
|
|
// add coeff, without leading bits
|
|
mask = (mask >> 2) - 1;
|
|
coeff &= mask;
|
|
r |= coeff;
|
|
|
|
return r;
|
|
}
|
|
|
|
//
|
|
// No overflow/underflow checking or checking for coefficients above 2^53
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
very_fast_get_BID64_small_mantissa (UINT64 sgn, int expon, UINT64 coeff) {
|
|
// no UF/OF
|
|
UINT64 r;
|
|
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (coeff | sgn);
|
|
return r;
|
|
}
|
|
|
|
|
|
//
|
|
// This pack macro is used when underflow is known to occur
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
get_BID64_UF (UINT64 sgn, int expon, UINT64 coeff, UINT64 R, int rmode,
|
|
unsigned *fpsc) {
|
|
UINT128 C128, Q_low, Stemp;
|
|
UINT64 C64, remainder_h, QH, carry, CY;
|
|
int extra_digits, amount, amount2;
|
|
unsigned status;
|
|
|
|
// underflow
|
|
if (expon + MAX_FORMAT_DIGITS < 0) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == ROUNDING_DOWN && sgn)
|
|
return 0x8000000000000001ull;
|
|
if (rmode == ROUNDING_UP && !sgn)
|
|
return 1ull;
|
|
#endif
|
|
#endif
|
|
// result is 0
|
|
return sgn;
|
|
}
|
|
// 10*coeff
|
|
coeff = (coeff << 3) + (coeff << 1);
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (sgn && (unsigned) (rmode - 1) < 2)
|
|
rmode = 3 - rmode;
|
|
#endif
|
|
#endif
|
|
if (R)
|
|
coeff |= 1;
|
|
// get digits to be shifted out
|
|
extra_digits = 1 - expon;
|
|
C128.w[0] = coeff + round_const_table[rmode][extra_digits];
|
|
|
|
// get coeff*(2^M[extra_digits])/10^extra_digits
|
|
__mul_64x128_full (QH, Q_low, C128.w[0],
|
|
reciprocals10_128[extra_digits]);
|
|
|
|
// now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128
|
|
amount = recip_scale[extra_digits];
|
|
|
|
C64 = QH >> amount;
|
|
//__shr_128(C128, Q_high, amount);
|
|
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == 0) //ROUNDING_TO_NEAREST
|
|
#endif
|
|
if (C64 & 1) {
|
|
// check whether fractional part of initial_P/10^extra_digits is exactly .5
|
|
|
|
// get remainder
|
|
amount2 = 64 - amount;
|
|
remainder_h = 0;
|
|
remainder_h--;
|
|
remainder_h >>= amount2;
|
|
remainder_h = remainder_h & QH;
|
|
|
|
if (!remainder_h
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0]))) {
|
|
C64--;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef SET_STATUS_FLAGS
|
|
|
|
if (is_inexact (fpsc))
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION);
|
|
else {
|
|
status = INEXACT_EXCEPTION;
|
|
// get remainder
|
|
remainder_h = QH << (64 - amount);
|
|
|
|
switch (rmode) {
|
|
case ROUNDING_TO_NEAREST:
|
|
case ROUNDING_TIES_AWAY:
|
|
// test whether fractional part is 0
|
|
if (remainder_h == 0x8000000000000000ull
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
case ROUNDING_DOWN:
|
|
case ROUNDING_TO_ZERO:
|
|
if (!remainder_h
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
default:
|
|
// round up
|
|
__add_carry_out (Stemp.w[0], CY, Q_low.w[0],
|
|
reciprocals10_128[extra_digits].w[0]);
|
|
__add_carry_in_out (Stemp.w[1], carry, Q_low.w[1],
|
|
reciprocals10_128[extra_digits].w[1], CY);
|
|
if ((remainder_h >> (64 - amount)) + carry >=
|
|
(((UINT64) 1) << amount))
|
|
status = EXACT_STATUS;
|
|
}
|
|
|
|
if (status != EXACT_STATUS)
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status);
|
|
}
|
|
|
|
#endif
|
|
|
|
return sgn | C64;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
// This pack macro doesnot check for coefficients above 2^53
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
get_BID64_small_mantissa (UINT64 sgn, int expon, UINT64 coeff,
|
|
int rmode, unsigned *fpsc) {
|
|
UINT128 C128, Q_low, Stemp;
|
|
UINT64 r, mask, C64, remainder_h, QH, carry, CY;
|
|
int extra_digits, amount, amount2;
|
|
unsigned status;
|
|
|
|
// check for possible underflow/overflow
|
|
if (((unsigned) expon) >= 3 * 256) {
|
|
if (expon < 0) {
|
|
// underflow
|
|
if (expon + MAX_FORMAT_DIGITS < 0) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc,
|
|
UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == ROUNDING_DOWN && sgn)
|
|
return 0x8000000000000001ull;
|
|
if (rmode == ROUNDING_UP && !sgn)
|
|
return 1ull;
|
|
#endif
|
|
#endif
|
|
// result is 0
|
|
return sgn;
|
|
}
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (sgn && (unsigned) (rmode - 1) < 2)
|
|
rmode = 3 - rmode;
|
|
#endif
|
|
#endif
|
|
// get digits to be shifted out
|
|
extra_digits = -expon;
|
|
C128.w[0] = coeff + round_const_table[rmode][extra_digits];
|
|
|
|
// get coeff*(2^M[extra_digits])/10^extra_digits
|
|
__mul_64x128_full (QH, Q_low, C128.w[0],
|
|
reciprocals10_128[extra_digits]);
|
|
|
|
// now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128
|
|
amount = recip_scale[extra_digits];
|
|
|
|
C64 = QH >> amount;
|
|
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == 0) //ROUNDING_TO_NEAREST
|
|
#endif
|
|
if (C64 & 1) {
|
|
// check whether fractional part of initial_P/10^extra_digits is exactly .5
|
|
|
|
// get remainder
|
|
amount2 = 64 - amount;
|
|
remainder_h = 0;
|
|
remainder_h--;
|
|
remainder_h >>= amount2;
|
|
remainder_h = remainder_h & QH;
|
|
|
|
if (!remainder_h
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0]))) {
|
|
C64--;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef SET_STATUS_FLAGS
|
|
|
|
if (is_inexact (fpsc))
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION);
|
|
else {
|
|
status = INEXACT_EXCEPTION;
|
|
// get remainder
|
|
remainder_h = QH << (64 - amount);
|
|
|
|
switch (rmode) {
|
|
case ROUNDING_TO_NEAREST:
|
|
case ROUNDING_TIES_AWAY:
|
|
// test whether fractional part is 0
|
|
if (remainder_h == 0x8000000000000000ull
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
case ROUNDING_DOWN:
|
|
case ROUNDING_TO_ZERO:
|
|
if (!remainder_h
|
|
&& (Q_low.w[1] < reciprocals10_128[extra_digits].w[1]
|
|
|| (Q_low.w[1] == reciprocals10_128[extra_digits].w[1]
|
|
&& Q_low.w[0] <
|
|
reciprocals10_128[extra_digits].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
default:
|
|
// round up
|
|
__add_carry_out (Stemp.w[0], CY, Q_low.w[0],
|
|
reciprocals10_128[extra_digits].w[0]);
|
|
__add_carry_in_out (Stemp.w[1], carry, Q_low.w[1],
|
|
reciprocals10_128[extra_digits].w[1], CY);
|
|
if ((remainder_h >> (64 - amount)) + carry >=
|
|
(((UINT64) 1) << amount))
|
|
status = EXACT_STATUS;
|
|
}
|
|
|
|
if (status != EXACT_STATUS)
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status);
|
|
}
|
|
|
|
#endif
|
|
|
|
return sgn | C64;
|
|
}
|
|
|
|
while (coeff < 1000000000000000ull && expon >= 3 * 256) {
|
|
expon--;
|
|
coeff = (coeff << 3) + (coeff << 1);
|
|
}
|
|
if (expon > DECIMAL_MAX_EXPON_64) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
// overflow
|
|
r = sgn | INFINITY_MASK64;
|
|
switch (rmode) {
|
|
case ROUNDING_DOWN:
|
|
if (!sgn)
|
|
r = LARGEST_BID64;
|
|
break;
|
|
case ROUNDING_TO_ZERO:
|
|
r = sgn | LARGEST_BID64;
|
|
break;
|
|
case ROUNDING_UP:
|
|
// round up
|
|
if (sgn)
|
|
r = SMALLEST_BID64;
|
|
}
|
|
return r;
|
|
} else {
|
|
mask = 1;
|
|
mask <<= EXPONENT_SHIFT_SMALL64;
|
|
if (coeff >= mask) {
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_LARGE64;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK64);
|
|
// add coeff, without leading bits
|
|
mask = (mask >> 2) - 1;
|
|
coeff &= mask;
|
|
r |= coeff;
|
|
return r;
|
|
}
|
|
}
|
|
}
|
|
|
|
r = expon;
|
|
r <<= EXPONENT_SHIFT_SMALL64;
|
|
r |= (coeff | sgn);
|
|
|
|
return r;
|
|
}
|
|
|
|
|
|
/*****************************************************************************
|
|
*
|
|
* BID128 pack/unpack macros
|
|
*
|
|
*****************************************************************************/
|
|
|
|
//
|
|
// Macro for handling BID128 underflow
|
|
// sticky bit given as additional argument
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
handle_UF_128_rem (UINT128 * pres, UINT64 sgn, int expon, UINT128 CQ,
|
|
UINT64 R, unsigned *prounding_mode, unsigned *fpsc) {
|
|
UINT128 T128, TP128, Qh, Ql, Qh1, Stemp, Tmp, Tmp1, CQ2, CQ8;
|
|
UINT64 carry, CY;
|
|
int ed2, amount;
|
|
unsigned rmode, status;
|
|
|
|
// UF occurs
|
|
if (expon + MAX_FORMAT_DIGITS_128 < 0) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
pres->w[1] = sgn;
|
|
pres->w[0] = 0;
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if ((sgn && *prounding_mode == ROUNDING_DOWN)
|
|
|| (!sgn && *prounding_mode == ROUNDING_UP))
|
|
pres->w[0] = 1ull;
|
|
#endif
|
|
#endif
|
|
return pres;
|
|
}
|
|
// CQ *= 10
|
|
CQ2.w[1] = (CQ.w[1] << 1) | (CQ.w[0] >> 63);
|
|
CQ2.w[0] = CQ.w[0] << 1;
|
|
CQ8.w[1] = (CQ.w[1] << 3) | (CQ.w[0] >> 61);
|
|
CQ8.w[0] = CQ.w[0] << 3;
|
|
__add_128_128 (CQ, CQ2, CQ8);
|
|
|
|
// add remainder
|
|
if (R)
|
|
CQ.w[0] |= 1;
|
|
|
|
ed2 = 1 - expon;
|
|
// add rounding constant to CQ
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
rmode = *prounding_mode;
|
|
if (sgn && (unsigned) (rmode - 1) < 2)
|
|
rmode = 3 - rmode;
|
|
#else
|
|
rmode = 0;
|
|
#endif
|
|
#else
|
|
rmode = 0;
|
|
#endif
|
|
T128 = round_const_table_128[rmode][ed2];
|
|
__add_carry_out (CQ.w[0], carry, T128.w[0], CQ.w[0]);
|
|
CQ.w[1] = CQ.w[1] + T128.w[1] + carry;
|
|
|
|
TP128 = reciprocals10_128[ed2];
|
|
__mul_128x128_full (Qh, Ql, CQ, TP128);
|
|
amount = recip_scale[ed2];
|
|
|
|
if (amount >= 64) {
|
|
CQ.w[0] = Qh.w[1] >> (amount - 64);
|
|
CQ.w[1] = 0;
|
|
} else {
|
|
__shr_128 (CQ, Qh, amount);
|
|
}
|
|
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (!(*prounding_mode))
|
|
#endif
|
|
if (CQ.w[0] & 1) {
|
|
// check whether fractional part of initial_P/10^ed1 is exactly .5
|
|
|
|
// get remainder
|
|
__shl_128_long (Qh1, Qh, (128 - amount));
|
|
|
|
if (!Qh1.w[1] && !Qh1.w[0]
|
|
&& (Ql.w[1] < reciprocals10_128[ed2].w[1]
|
|
|| (Ql.w[1] == reciprocals10_128[ed2].w[1]
|
|
&& Ql.w[0] < reciprocals10_128[ed2].w[0]))) {
|
|
CQ.w[0]--;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef SET_STATUS_FLAGS
|
|
|
|
if (is_inexact (fpsc))
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION);
|
|
else {
|
|
status = INEXACT_EXCEPTION;
|
|
// get remainder
|
|
__shl_128_long (Qh1, Qh, (128 - amount));
|
|
|
|
switch (rmode) {
|
|
case ROUNDING_TO_NEAREST:
|
|
case ROUNDING_TIES_AWAY:
|
|
// test whether fractional part is 0
|
|
if (Qh1.w[1] == 0x8000000000000000ull && (!Qh1.w[0])
|
|
&& (Ql.w[1] < reciprocals10_128[ed2].w[1]
|
|
|| (Ql.w[1] == reciprocals10_128[ed2].w[1]
|
|
&& Ql.w[0] < reciprocals10_128[ed2].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
case ROUNDING_DOWN:
|
|
case ROUNDING_TO_ZERO:
|
|
if ((!Qh1.w[1]) && (!Qh1.w[0])
|
|
&& (Ql.w[1] < reciprocals10_128[ed2].w[1]
|
|
|| (Ql.w[1] == reciprocals10_128[ed2].w[1]
|
|
&& Ql.w[0] < reciprocals10_128[ed2].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
default:
|
|
// round up
|
|
__add_carry_out (Stemp.w[0], CY, Ql.w[0],
|
|
reciprocals10_128[ed2].w[0]);
|
|
__add_carry_in_out (Stemp.w[1], carry, Ql.w[1],
|
|
reciprocals10_128[ed2].w[1], CY);
|
|
__shr_128_long (Qh, Qh1, (128 - amount));
|
|
Tmp.w[0] = 1;
|
|
Tmp.w[1] = 0;
|
|
__shl_128_long (Tmp1, Tmp, amount);
|
|
Qh.w[0] += carry;
|
|
if (Qh.w[0] < carry)
|
|
Qh.w[1]++;
|
|
if (__unsigned_compare_ge_128 (Qh, Tmp1))
|
|
status = EXACT_STATUS;
|
|
}
|
|
|
|
if (status != EXACT_STATUS)
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status);
|
|
}
|
|
|
|
#endif
|
|
|
|
pres->w[1] = sgn | CQ.w[1];
|
|
pres->w[0] = CQ.w[0];
|
|
|
|
return pres;
|
|
|
|
}
|
|
|
|
|
|
//
|
|
// Macro for handling BID128 underflow
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
handle_UF_128 (UINT128 * pres, UINT64 sgn, int expon, UINT128 CQ,
|
|
unsigned *prounding_mode, unsigned *fpsc) {
|
|
UINT128 T128, TP128, Qh, Ql, Qh1, Stemp, Tmp, Tmp1;
|
|
UINT64 carry, CY;
|
|
int ed2, amount;
|
|
unsigned rmode, status;
|
|
|
|
// UF occurs
|
|
if (expon + MAX_FORMAT_DIGITS_128 < 0) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
pres->w[1] = sgn;
|
|
pres->w[0] = 0;
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if ((sgn && *prounding_mode == ROUNDING_DOWN)
|
|
|| (!sgn && *prounding_mode == ROUNDING_UP))
|
|
pres->w[0] = 1ull;
|
|
#endif
|
|
#endif
|
|
return pres;
|
|
}
|
|
|
|
ed2 = 0 - expon;
|
|
// add rounding constant to CQ
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
rmode = *prounding_mode;
|
|
if (sgn && (unsigned) (rmode - 1) < 2)
|
|
rmode = 3 - rmode;
|
|
#else
|
|
rmode = 0;
|
|
#endif
|
|
#else
|
|
rmode = 0;
|
|
#endif
|
|
|
|
T128 = round_const_table_128[rmode][ed2];
|
|
__add_carry_out (CQ.w[0], carry, T128.w[0], CQ.w[0]);
|
|
CQ.w[1] = CQ.w[1] + T128.w[1] + carry;
|
|
|
|
TP128 = reciprocals10_128[ed2];
|
|
__mul_128x128_full (Qh, Ql, CQ, TP128);
|
|
amount = recip_scale[ed2];
|
|
|
|
if (amount >= 64) {
|
|
CQ.w[0] = Qh.w[1] >> (amount - 64);
|
|
CQ.w[1] = 0;
|
|
} else {
|
|
__shr_128 (CQ, Qh, amount);
|
|
}
|
|
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (!(*prounding_mode))
|
|
#endif
|
|
if (CQ.w[0] & 1) {
|
|
// check whether fractional part of initial_P/10^ed1 is exactly .5
|
|
|
|
// get remainder
|
|
__shl_128_long (Qh1, Qh, (128 - amount));
|
|
|
|
if (!Qh1.w[1] && !Qh1.w[0]
|
|
&& (Ql.w[1] < reciprocals10_128[ed2].w[1]
|
|
|| (Ql.w[1] == reciprocals10_128[ed2].w[1]
|
|
&& Ql.w[0] < reciprocals10_128[ed2].w[0]))) {
|
|
CQ.w[0]--;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef SET_STATUS_FLAGS
|
|
|
|
if (is_inexact (fpsc))
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION);
|
|
else {
|
|
status = INEXACT_EXCEPTION;
|
|
// get remainder
|
|
__shl_128_long (Qh1, Qh, (128 - amount));
|
|
|
|
switch (rmode) {
|
|
case ROUNDING_TO_NEAREST:
|
|
case ROUNDING_TIES_AWAY:
|
|
// test whether fractional part is 0
|
|
if (Qh1.w[1] == 0x8000000000000000ull && (!Qh1.w[0])
|
|
&& (Ql.w[1] < reciprocals10_128[ed2].w[1]
|
|
|| (Ql.w[1] == reciprocals10_128[ed2].w[1]
|
|
&& Ql.w[0] < reciprocals10_128[ed2].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
case ROUNDING_DOWN:
|
|
case ROUNDING_TO_ZERO:
|
|
if ((!Qh1.w[1]) && (!Qh1.w[0])
|
|
&& (Ql.w[1] < reciprocals10_128[ed2].w[1]
|
|
|| (Ql.w[1] == reciprocals10_128[ed2].w[1]
|
|
&& Ql.w[0] < reciprocals10_128[ed2].w[0])))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
default:
|
|
// round up
|
|
__add_carry_out (Stemp.w[0], CY, Ql.w[0],
|
|
reciprocals10_128[ed2].w[0]);
|
|
__add_carry_in_out (Stemp.w[1], carry, Ql.w[1],
|
|
reciprocals10_128[ed2].w[1], CY);
|
|
__shr_128_long (Qh, Qh1, (128 - amount));
|
|
Tmp.w[0] = 1;
|
|
Tmp.w[1] = 0;
|
|
__shl_128_long (Tmp1, Tmp, amount);
|
|
Qh.w[0] += carry;
|
|
if (Qh.w[0] < carry)
|
|
Qh.w[1]++;
|
|
if (__unsigned_compare_ge_128 (Qh, Tmp1))
|
|
status = EXACT_STATUS;
|
|
}
|
|
|
|
if (status != EXACT_STATUS)
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status);
|
|
}
|
|
|
|
#endif
|
|
|
|
pres->w[1] = sgn | CQ.w[1];
|
|
pres->w[0] = CQ.w[0];
|
|
|
|
return pres;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
// BID128 unpack, input passed by value
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
unpack_BID128_value (UINT64 * psign_x, int *pexponent_x,
|
|
UINT128 * pcoefficient_x, UINT128 x) {
|
|
UINT128 coeff, T33, T34;
|
|
UINT64 ex;
|
|
|
|
*psign_x = (x.w[1]) & 0x8000000000000000ull;
|
|
|
|
// special encodings
|
|
if ((x.w[1] & INFINITY_MASK64) >= SPECIAL_ENCODING_MASK64) {
|
|
if ((x.w[1] & INFINITY_MASK64) < INFINITY_MASK64) {
|
|
// non-canonical input
|
|
pcoefficient_x->w[0] = 0;
|
|
pcoefficient_x->w[1] = 0;
|
|
ex = (x.w[1]) >> 47;
|
|
*pexponent_x = ((int) ex) & EXPONENT_MASK128;
|
|
return 0;
|
|
}
|
|
// 10^33
|
|
T33 = power10_table_128[33];
|
|
/*coeff.w[0] = x.w[0];
|
|
coeff.w[1] = (x.w[1]) & LARGE_COEFF_MASK128;
|
|
pcoefficient_x->w[0] = x.w[0];
|
|
pcoefficient_x->w[1] = x.w[1];
|
|
if (__unsigned_compare_ge_128 (coeff, T33)) // non-canonical
|
|
pcoefficient_x->w[1] &= (~LARGE_COEFF_MASK128); */
|
|
|
|
pcoefficient_x->w[0] = x.w[0];
|
|
pcoefficient_x->w[1] = (x.w[1]) & 0x00003fffffffffffull;
|
|
if (__unsigned_compare_ge_128 ((*pcoefficient_x), T33)) // non-canonical
|
|
{
|
|
pcoefficient_x->w[1] = (x.w[1]) & 0xfe00000000000000ull;
|
|
pcoefficient_x->w[0] = 0;
|
|
} else
|
|
pcoefficient_x->w[1] = (x.w[1]) & 0xfe003fffffffffffull;
|
|
if ((x.w[1] & NAN_MASK64) == INFINITY_MASK64) {
|
|
pcoefficient_x->w[0] = 0;
|
|
pcoefficient_x->w[1] = x.w[1] & SINFINITY_MASK64;
|
|
}
|
|
*pexponent_x = 0;
|
|
return 0; // NaN or Infinity
|
|
}
|
|
|
|
coeff.w[0] = x.w[0];
|
|
coeff.w[1] = (x.w[1]) & SMALL_COEFF_MASK128;
|
|
|
|
// 10^34
|
|
T34 = power10_table_128[34];
|
|
// check for non-canonical values
|
|
if (__unsigned_compare_ge_128 (coeff, T34))
|
|
coeff.w[0] = coeff.w[1] = 0;
|
|
|
|
pcoefficient_x->w[0] = coeff.w[0];
|
|
pcoefficient_x->w[1] = coeff.w[1];
|
|
|
|
ex = (x.w[1]) >> 49;
|
|
*pexponent_x = ((int) ex) & EXPONENT_MASK128;
|
|
|
|
return coeff.w[0] | coeff.w[1];
|
|
}
|
|
|
|
|
|
//
|
|
// BID128 unpack, input pased by reference
|
|
//
|
|
__BID_INLINE__ UINT64
|
|
unpack_BID128 (UINT64 * psign_x, int *pexponent_x,
|
|
UINT128 * pcoefficient_x, UINT128 * px) {
|
|
UINT128 coeff, T33, T34;
|
|
UINT64 ex;
|
|
|
|
*psign_x = (px->w[1]) & 0x8000000000000000ull;
|
|
|
|
// special encodings
|
|
if ((px->w[1] & INFINITY_MASK64) >= SPECIAL_ENCODING_MASK64) {
|
|
if ((px->w[1] & INFINITY_MASK64) < INFINITY_MASK64) {
|
|
// non-canonical input
|
|
pcoefficient_x->w[0] = 0;
|
|
pcoefficient_x->w[1] = 0;
|
|
ex = (px->w[1]) >> 47;
|
|
*pexponent_x = ((int) ex) & EXPONENT_MASK128;
|
|
return 0;
|
|
}
|
|
// 10^33
|
|
T33 = power10_table_128[33];
|
|
coeff.w[0] = px->w[0];
|
|
coeff.w[1] = (px->w[1]) & LARGE_COEFF_MASK128;
|
|
pcoefficient_x->w[0] = px->w[0];
|
|
pcoefficient_x->w[1] = px->w[1];
|
|
if (__unsigned_compare_ge_128 (coeff, T33)) { // non-canonical
|
|
pcoefficient_x->w[1] &= (~LARGE_COEFF_MASK128);
|
|
pcoefficient_x->w[0] = 0;
|
|
}
|
|
*pexponent_x = 0;
|
|
return 0; // NaN or Infinity
|
|
}
|
|
|
|
coeff.w[0] = px->w[0];
|
|
coeff.w[1] = (px->w[1]) & SMALL_COEFF_MASK128;
|
|
|
|
// 10^34
|
|
T34 = power10_table_128[34];
|
|
// check for non-canonical values
|
|
if (__unsigned_compare_ge_128 (coeff, T34))
|
|
coeff.w[0] = coeff.w[1] = 0;
|
|
|
|
pcoefficient_x->w[0] = coeff.w[0];
|
|
pcoefficient_x->w[1] = coeff.w[1];
|
|
|
|
ex = (px->w[1]) >> 49;
|
|
*pexponent_x = ((int) ex) & EXPONENT_MASK128;
|
|
|
|
return coeff.w[0] | coeff.w[1];
|
|
}
|
|
|
|
//
|
|
// Pack macro checks for overflow, but not underflow
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
get_BID128_very_fast_OF (UINT128 * pres, UINT64 sgn, int expon,
|
|
UINT128 coeff, unsigned *prounding_mode,
|
|
unsigned *fpsc) {
|
|
UINT128 T;
|
|
UINT64 tmp, tmp2;
|
|
|
|
if ((unsigned) expon > DECIMAL_MAX_EXPON_128) {
|
|
|
|
if (expon - MAX_FORMAT_DIGITS_128 <= DECIMAL_MAX_EXPON_128) {
|
|
T = power10_table_128[MAX_FORMAT_DIGITS_128 - 1];
|
|
while (__unsigned_compare_gt_128 (T, coeff)
|
|
&& expon > DECIMAL_MAX_EXPON_128) {
|
|
coeff.w[1] =
|
|
(coeff.w[1] << 3) + (coeff.w[1] << 1) + (coeff.w[0] >> 61) +
|
|
(coeff.w[0] >> 63);
|
|
tmp2 = coeff.w[0] << 3;
|
|
coeff.w[0] = (coeff.w[0] << 1) + tmp2;
|
|
if (coeff.w[0] < tmp2)
|
|
coeff.w[1]++;
|
|
|
|
expon--;
|
|
}
|
|
}
|
|
if ((unsigned) expon > DECIMAL_MAX_EXPON_128) {
|
|
// OF
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (*prounding_mode == ROUNDING_TO_ZERO
|
|
|| (sgn && *prounding_mode == ROUNDING_UP) || (!sgn
|
|
&&
|
|
*prounding_mode
|
|
==
|
|
ROUNDING_DOWN))
|
|
{
|
|
pres->w[1] = sgn | LARGEST_BID128_HIGH;
|
|
pres->w[0] = LARGEST_BID128_LOW;
|
|
} else
|
|
#endif
|
|
#endif
|
|
{
|
|
pres->w[1] = sgn | INFINITY_MASK64;
|
|
pres->w[0] = 0;
|
|
}
|
|
return pres;
|
|
}
|
|
}
|
|
|
|
pres->w[0] = coeff.w[0];
|
|
tmp = expon;
|
|
tmp <<= 49;
|
|
pres->w[1] = sgn | tmp | coeff.w[1];
|
|
|
|
return pres;
|
|
}
|
|
|
|
|
|
//
|
|
// No overflow/underflow checks
|
|
// No checking for coefficient == 10^34 (rounding artifact)
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
get_BID128_very_fast (UINT128 * pres, UINT64 sgn, int expon,
|
|
UINT128 coeff) {
|
|
UINT64 tmp;
|
|
|
|
pres->w[0] = coeff.w[0];
|
|
tmp = expon;
|
|
tmp <<= 49;
|
|
pres->w[1] = sgn | tmp | coeff.w[1];
|
|
|
|
return pres;
|
|
}
|
|
|
|
//
|
|
// No overflow/underflow checks
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
get_BID128_fast (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff) {
|
|
UINT64 tmp;
|
|
|
|
// coeff==10^34?
|
|
if (coeff.w[1] == 0x0001ed09bead87c0ull
|
|
&& coeff.w[0] == 0x378d8e6400000000ull) {
|
|
expon++;
|
|
// set coefficient to 10^33
|
|
coeff.w[1] = 0x0000314dc6448d93ull;
|
|
coeff.w[0] = 0x38c15b0a00000000ull;
|
|
}
|
|
|
|
pres->w[0] = coeff.w[0];
|
|
tmp = expon;
|
|
tmp <<= 49;
|
|
pres->w[1] = sgn | tmp | coeff.w[1];
|
|
|
|
return pres;
|
|
}
|
|
|
|
//
|
|
// General BID128 pack macro
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
get_BID128 (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff,
|
|
unsigned *prounding_mode, unsigned *fpsc) {
|
|
UINT128 T;
|
|
UINT64 tmp, tmp2;
|
|
|
|
// coeff==10^34?
|
|
if (coeff.w[1] == 0x0001ed09bead87c0ull
|
|
&& coeff.w[0] == 0x378d8e6400000000ull) {
|
|
expon++;
|
|
// set coefficient to 10^33
|
|
coeff.w[1] = 0x0000314dc6448d93ull;
|
|
coeff.w[0] = 0x38c15b0a00000000ull;
|
|
}
|
|
// check OF, UF
|
|
if (expon < 0 || expon > DECIMAL_MAX_EXPON_128) {
|
|
// check UF
|
|
if (expon < 0) {
|
|
return handle_UF_128 (pres, sgn, expon, coeff, prounding_mode,
|
|
fpsc);
|
|
}
|
|
|
|
if (expon - MAX_FORMAT_DIGITS_128 <= DECIMAL_MAX_EXPON_128) {
|
|
T = power10_table_128[MAX_FORMAT_DIGITS_128 - 1];
|
|
while (__unsigned_compare_gt_128 (T, coeff)
|
|
&& expon > DECIMAL_MAX_EXPON_128) {
|
|
coeff.w[1] =
|
|
(coeff.w[1] << 3) + (coeff.w[1] << 1) + (coeff.w[0] >> 61) +
|
|
(coeff.w[0] >> 63);
|
|
tmp2 = coeff.w[0] << 3;
|
|
coeff.w[0] = (coeff.w[0] << 1) + tmp2;
|
|
if (coeff.w[0] < tmp2)
|
|
coeff.w[1]++;
|
|
|
|
expon--;
|
|
}
|
|
}
|
|
if (expon > DECIMAL_MAX_EXPON_128) {
|
|
if (!(coeff.w[1] | coeff.w[0])) {
|
|
pres->w[1] = sgn | (((UINT64) DECIMAL_MAX_EXPON_128) << 49);
|
|
pres->w[0] = 0;
|
|
return pres;
|
|
}
|
|
// OF
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (*prounding_mode == ROUNDING_TO_ZERO
|
|
|| (sgn && *prounding_mode == ROUNDING_UP) || (!sgn
|
|
&&
|
|
*prounding_mode
|
|
==
|
|
ROUNDING_DOWN))
|
|
{
|
|
pres->w[1] = sgn | LARGEST_BID128_HIGH;
|
|
pres->w[0] = LARGEST_BID128_LOW;
|
|
} else
|
|
#endif
|
|
#endif
|
|
{
|
|
pres->w[1] = sgn | INFINITY_MASK64;
|
|
pres->w[0] = 0;
|
|
}
|
|
return pres;
|
|
}
|
|
}
|
|
|
|
pres->w[0] = coeff.w[0];
|
|
tmp = expon;
|
|
tmp <<= 49;
|
|
pres->w[1] = sgn | tmp | coeff.w[1];
|
|
|
|
return pres;
|
|
}
|
|
|
|
|
|
//
|
|
// Macro used for conversions from string
|
|
// (no additional arguments given for rounding mode, status flags)
|
|
//
|
|
__BID_INLINE__ UINT128 *
|
|
get_BID128_string (UINT128 * pres, UINT64 sgn, int expon, UINT128 coeff) {
|
|
UINT128 D2, D8;
|
|
UINT64 tmp;
|
|
unsigned rmode = 0, status;
|
|
|
|
// coeff==10^34?
|
|
if (coeff.w[1] == 0x0001ed09bead87c0ull
|
|
&& coeff.w[0] == 0x378d8e6400000000ull) {
|
|
expon++;
|
|
// set coefficient to 10^33
|
|
coeff.w[1] = 0x0000314dc6448d93ull;
|
|
coeff.w[0] = 0x38c15b0a00000000ull;
|
|
}
|
|
// check OF, UF
|
|
if ((unsigned) expon > DECIMAL_MAX_EXPON_128) {
|
|
// check UF
|
|
if (expon < 0)
|
|
return handle_UF_128 (pres, sgn, expon, coeff, &rmode, &status);
|
|
|
|
// OF
|
|
|
|
if (expon < DECIMAL_MAX_EXPON_128 + 34) {
|
|
while (expon > DECIMAL_MAX_EXPON_128 &&
|
|
(coeff.w[1] < power10_table_128[33].w[1] ||
|
|
(coeff.w[1] == power10_table_128[33].w[1]
|
|
&& coeff.w[0] < power10_table_128[33].w[0]))) {
|
|
D2.w[1] = (coeff.w[1] << 1) | (coeff.w[0] >> 63);
|
|
D2.w[0] = coeff.w[0] << 1;
|
|
D8.w[1] = (coeff.w[1] << 3) | (coeff.w[0] >> 61);
|
|
D8.w[0] = coeff.w[0] << 3;
|
|
|
|
__add_128_128 (coeff, D2, D8);
|
|
expon--;
|
|
}
|
|
} else if (!(coeff.w[0] | coeff.w[1]))
|
|
expon = DECIMAL_MAX_EXPON_128;
|
|
|
|
if (expon > DECIMAL_MAX_EXPON_128) {
|
|
pres->w[1] = sgn | INFINITY_MASK64;
|
|
pres->w[0] = 0;
|
|
switch (rmode) {
|
|
case ROUNDING_DOWN:
|
|
if (!sgn) {
|
|
pres->w[1] = LARGEST_BID128_HIGH;
|
|
pres->w[0] = LARGEST_BID128_LOW;
|
|
}
|
|
break;
|
|
case ROUNDING_TO_ZERO:
|
|
pres->w[1] = sgn | LARGEST_BID128_HIGH;
|
|
pres->w[0] = LARGEST_BID128_LOW;
|
|
break;
|
|
case ROUNDING_UP:
|
|
// round up
|
|
if (sgn) {
|
|
pres->w[1] = sgn | LARGEST_BID128_HIGH;
|
|
pres->w[0] = LARGEST_BID128_LOW;
|
|
}
|
|
break;
|
|
}
|
|
|
|
return pres;
|
|
}
|
|
}
|
|
|
|
pres->w[0] = coeff.w[0];
|
|
tmp = expon;
|
|
tmp <<= 49;
|
|
pres->w[1] = sgn | tmp | coeff.w[1];
|
|
|
|
return pres;
|
|
}
|
|
|
|
|
|
|
|
/*****************************************************************************
|
|
*
|
|
* BID32 pack/unpack macros
|
|
*
|
|
*****************************************************************************/
|
|
|
|
|
|
__BID_INLINE__ UINT32
|
|
unpack_BID32 (UINT32 * psign_x, int *pexponent_x,
|
|
UINT32 * pcoefficient_x, UINT32 x) {
|
|
UINT32 tmp;
|
|
|
|
*psign_x = x & 0x80000000;
|
|
|
|
if ((x & SPECIAL_ENCODING_MASK32) == SPECIAL_ENCODING_MASK32) {
|
|
// special encodings
|
|
if ((x & INFINITY_MASK32) == INFINITY_MASK32) {
|
|
*pcoefficient_x = x & 0xfe0fffff;
|
|
if ((x & 0x000fffff) >= 1000000)
|
|
*pcoefficient_x = x & 0xfe000000;
|
|
if ((x & NAN_MASK32) == INFINITY_MASK32)
|
|
*pcoefficient_x = x & 0xf8000000;
|
|
*pexponent_x = 0;
|
|
return 0; // NaN or Infinity
|
|
}
|
|
// coefficient
|
|
*pcoefficient_x = (x & SMALL_COEFF_MASK32) | LARGE_COEFF_HIGH_BIT32;
|
|
// check for non-canonical value
|
|
if (*pcoefficient_x >= 10000000)
|
|
*pcoefficient_x = 0;
|
|
// get exponent
|
|
tmp = x >> 21;
|
|
*pexponent_x = tmp & EXPONENT_MASK32;
|
|
return 1;
|
|
}
|
|
// exponent
|
|
tmp = x >> 23;
|
|
*pexponent_x = tmp & EXPONENT_MASK32;
|
|
// coefficient
|
|
*pcoefficient_x = (x & LARGE_COEFF_MASK32);
|
|
|
|
return *pcoefficient_x;
|
|
}
|
|
|
|
//
|
|
// General pack macro for BID32
|
|
//
|
|
__BID_INLINE__ UINT32
|
|
get_BID32 (UINT32 sgn, int expon, UINT64 coeff, int rmode,
|
|
unsigned *fpsc) {
|
|
UINT128 Q;
|
|
UINT64 C64, remainder_h, carry, Stemp;
|
|
UINT32 r, mask;
|
|
int extra_digits, amount, amount2;
|
|
unsigned status;
|
|
|
|
if (coeff > 9999999ull) {
|
|
expon++;
|
|
coeff = 1000000ull;
|
|
}
|
|
// check for possible underflow/overflow
|
|
if (((unsigned) expon) > DECIMAL_MAX_EXPON_32) {
|
|
if (expon < 0) {
|
|
// underflow
|
|
if (expon + MAX_FORMAT_DIGITS_32 < 0) {
|
|
#ifdef SET_STATUS_FLAGS
|
|
__set_status_flags (fpsc,
|
|
UNDERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == ROUNDING_DOWN && sgn)
|
|
return 0x80000001;
|
|
if (rmode == ROUNDING_UP && !sgn)
|
|
return 1;
|
|
#endif
|
|
#endif
|
|
// result is 0
|
|
return sgn;
|
|
}
|
|
// get digits to be shifted out
|
|
#ifdef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
rmode = 0;
|
|
#endif
|
|
#ifdef IEEE_ROUND_NEAREST
|
|
rmode = 0;
|
|
#endif
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (sgn && (unsigned) (rmode - 1) < 2)
|
|
rmode = 3 - rmode;
|
|
#endif
|
|
#endif
|
|
|
|
extra_digits = -expon;
|
|
coeff += round_const_table[rmode][extra_digits];
|
|
|
|
// get coeff*(2^M[extra_digits])/10^extra_digits
|
|
__mul_64x64_to_128 (Q, coeff, reciprocals10_64[extra_digits]);
|
|
|
|
// now get P/10^extra_digits: shift Q_high right by M[extra_digits]-128
|
|
amount = short_recip_scale[extra_digits];
|
|
|
|
C64 = Q.w[1] >> amount;
|
|
|
|
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
|
|
#ifndef IEEE_ROUND_NEAREST
|
|
if (rmode == 0) //ROUNDING_TO_NEAREST
|
|
#endif
|
|
if (C64 & 1) {
|
|
// check whether fractional part of initial_P/10^extra_digits is exactly .5
|
|
|
|
// get remainder
|
|
amount2 = 64 - amount;
|
|
remainder_h = 0;
|
|
remainder_h--;
|
|
remainder_h >>= amount2;
|
|
remainder_h = remainder_h & Q.w[1];
|
|
|
|
if (!remainder_h && (Q.w[0] < reciprocals10_64[extra_digits])) {
|
|
C64--;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#ifdef SET_STATUS_FLAGS
|
|
|
|
if (is_inexact (fpsc))
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION);
|
|
else {
|
|
status = INEXACT_EXCEPTION;
|
|
// get remainder
|
|
remainder_h = Q.w[1] << (64 - amount);
|
|
|
|
switch (rmode) {
|
|
case ROUNDING_TO_NEAREST:
|
|
case ROUNDING_TIES_AWAY:
|
|
// test whether fractional part is 0
|
|
if (remainder_h == 0x8000000000000000ull
|
|
&& (Q.w[0] < reciprocals10_64[extra_digits]))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
case ROUNDING_DOWN:
|
|
case ROUNDING_TO_ZERO:
|
|
if (!remainder_h && (Q.w[0] < reciprocals10_64[extra_digits]))
|
|
status = EXACT_STATUS;
|
|
break;
|
|
default:
|
|
// round up
|
|
__add_carry_out (Stemp, carry, Q.w[0],
|
|
reciprocals10_64[extra_digits]);
|
|
if ((remainder_h >> (64 - amount)) + carry >=
|
|
(((UINT64) 1) << amount))
|
|
status = EXACT_STATUS;
|
|
}
|
|
|
|
if (status != EXACT_STATUS)
|
|
__set_status_flags (fpsc, UNDERFLOW_EXCEPTION | status);
|
|
}
|
|
|
|
#endif
|
|
|
|
return sgn | (UINT32) C64;
|
|
}
|
|
|
|
while (coeff < 1000000 && expon > DECIMAL_MAX_EXPON_32) {
|
|
coeff = (coeff << 3) + (coeff << 1);
|
|
expon--;
|
|
}
|
|
if (((unsigned) expon) > DECIMAL_MAX_EXPON_32) {
|
|
__set_status_flags (fpsc, OVERFLOW_EXCEPTION | INEXACT_EXCEPTION);
|
|
// overflow
|
|
r = sgn | INFINITY_MASK32;
|
|
switch (rmode) {
|
|
case ROUNDING_DOWN:
|
|
if (!sgn)
|
|
r = LARGEST_BID32;
|
|
break;
|
|
case ROUNDING_TO_ZERO:
|
|
r = sgn | LARGEST_BID32;
|
|
break;
|
|
case ROUNDING_UP:
|
|
// round up
|
|
if (sgn)
|
|
r = sgn | LARGEST_BID32;
|
|
}
|
|
return r;
|
|
}
|
|
}
|
|
|
|
mask = 1 << 23;
|
|
|
|
// check whether coefficient fits in DECIMAL_COEFF_FIT bits
|
|
if (coeff < mask) {
|
|
r = expon;
|
|
r <<= 23;
|
|
r |= ((UINT32) coeff | sgn);
|
|
return r;
|
|
}
|
|
// special format
|
|
|
|
r = expon;
|
|
r <<= 21;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK32);
|
|
// add coeff, without leading bits
|
|
mask = (1 << 21) - 1;
|
|
r |= (((UINT32) coeff) & mask);
|
|
|
|
return r;
|
|
}
|
|
|
|
|
|
|
|
//
|
|
// no overflow/underflow checks
|
|
//
|
|
__BID_INLINE__ UINT32
|
|
very_fast_get_BID32 (UINT32 sgn, int expon, UINT32 coeff) {
|
|
UINT32 r, mask;
|
|
|
|
mask = 1 << 23;
|
|
|
|
// check whether coefficient fits in 10*2+3 bits
|
|
if (coeff < mask) {
|
|
r = expon;
|
|
r <<= 23;
|
|
r |= (coeff | sgn);
|
|
return r;
|
|
}
|
|
// special format
|
|
r = expon;
|
|
r <<= 21;
|
|
r |= (sgn | SPECIAL_ENCODING_MASK32);
|
|
// add coeff, without leading bits
|
|
mask = (1 << 21) - 1;
|
|
coeff &= mask;
|
|
r |= coeff;
|
|
|
|
return r;
|
|
}
|
|
|
|
|
|
|
|
/*************************************************************
|
|
*
|
|
*************************************************************/
|
|
typedef
|
|
ALIGN (16)
|
|
struct {
|
|
UINT64 w[6];
|
|
} UINT384;
|
|
typedef ALIGN (16)
|
|
struct {
|
|
UINT64 w[8];
|
|
} UINT512;
|
|
|
|
// #define P 34
|
|
#define MASK_STEERING_BITS 0x6000000000000000ull
|
|
#define MASK_BINARY_EXPONENT1 0x7fe0000000000000ull
|
|
#define MASK_BINARY_SIG1 0x001fffffffffffffull
|
|
#define MASK_BINARY_EXPONENT2 0x1ff8000000000000ull
|
|
//used to take G[2:w+3] (sec 3.3)
|
|
#define MASK_BINARY_SIG2 0x0007ffffffffffffull
|
|
//used to mask out G4:T0 (sec 3.3)
|
|
#define MASK_BINARY_OR2 0x0020000000000000ull
|
|
//used to prefix 8+G4 to T (sec 3.3)
|
|
#define UPPER_EXPON_LIMIT 51
|
|
#define MASK_EXP 0x7ffe000000000000ull
|
|
#define MASK_SPECIAL 0x7800000000000000ull
|
|
#define MASK_NAN 0x7c00000000000000ull
|
|
#define MASK_SNAN 0x7e00000000000000ull
|
|
#define MASK_ANY_INF 0x7c00000000000000ull
|
|
#define MASK_INF 0x7800000000000000ull
|
|
#define MASK_SIGN 0x8000000000000000ull
|
|
#define MASK_COEFF 0x0001ffffffffffffull
|
|
#define BIN_EXP_BIAS (0x1820ull << 49)
|
|
|
|
#define EXP_MIN 0x0000000000000000ull
|
|
// EXP_MIN = (-6176 + 6176) << 49
|
|
#define EXP_MAX 0x5ffe000000000000ull
|
|
// EXP_MAX = (6111 + 6176) << 49
|
|
#define EXP_MAX_P1 0x6000000000000000ull
|
|
// EXP_MAX + 1 = (6111 + 6176 + 1) << 49
|
|
#define EXP_P1 0x0002000000000000ull
|
|
// EXP_ P1= 1 << 49
|
|
#define expmin -6176
|
|
// min unbiased exponent
|
|
#define expmax 6111
|
|
// max unbiased exponent
|
|
#define expmin16 -398
|
|
// min unbiased exponent
|
|
#define expmax16 369
|
|
// max unbiased exponent
|
|
|
|
#define SIGNMASK32 0x80000000
|
|
#define BID64_SIG_MAX 0x002386F26FC0ffffull
|
|
#define SIGNMASK64 0x8000000000000000ull
|
|
|
|
// typedef unsigned int FPSC; // floating-point status and control
|
|
// bit31:
|
|
// bit30:
|
|
// bit29:
|
|
// bit28:
|
|
// bit27:
|
|
// bit26:
|
|
// bit25:
|
|
// bit24:
|
|
// bit23:
|
|
// bit22:
|
|
// bit21:
|
|
// bit20:
|
|
// bit19:
|
|
// bit18:
|
|
// bit17:
|
|
// bit16:
|
|
// bit15:
|
|
// bit14: RC:2
|
|
// bit13: RC:1
|
|
// bit12: RC:0
|
|
// bit11: PM
|
|
// bit10: UM
|
|
// bit9: OM
|
|
// bit8: ZM
|
|
// bit7: DM
|
|
// bit6: IM
|
|
// bit5: PE
|
|
// bit4: UE
|
|
// bit3: OE
|
|
// bit2: ZE
|
|
// bit1: DE
|
|
// bit0: IE
|
|
|
|
#define ROUNDING_MODE_MASK 0x00007000
|
|
|
|
typedef struct _DEC_DIGITS {
|
|
unsigned int digits;
|
|
UINT64 threshold_hi;
|
|
UINT64 threshold_lo;
|
|
unsigned int digits1;
|
|
} DEC_DIGITS;
|
|
|
|
extern DEC_DIGITS nr_digits[];
|
|
extern UINT64 midpoint64[];
|
|
extern UINT128 midpoint128[];
|
|
extern UINT192 midpoint192[];
|
|
extern UINT256 midpoint256[];
|
|
extern UINT64 ten2k64[];
|
|
extern UINT128 ten2k128[];
|
|
extern UINT256 ten2k256[];
|
|
extern UINT128 ten2mk128[];
|
|
extern UINT64 ten2mk64[];
|
|
extern UINT128 ten2mk128trunc[];
|
|
extern int shiftright128[];
|
|
extern UINT64 maskhigh128[];
|
|
extern UINT64 maskhigh128M[];
|
|
extern UINT64 maskhigh192M[];
|
|
extern UINT64 maskhigh256M[];
|
|
extern UINT64 onehalf128[];
|
|
extern UINT64 onehalf128M[];
|
|
extern UINT64 onehalf192M[];
|
|
extern UINT64 onehalf256M[];
|
|
extern UINT128 ten2mk128M[];
|
|
extern UINT128 ten2mk128truncM[];
|
|
extern UINT192 ten2mk192truncM[];
|
|
extern UINT256 ten2mk256truncM[];
|
|
extern int shiftright128M[];
|
|
extern int shiftright192M[];
|
|
extern int shiftright256M[];
|
|
extern UINT192 ten2mk192M[];
|
|
extern UINT256 ten2mk256M[];
|
|
extern unsigned char char_table2[];
|
|
extern unsigned char char_table3[];
|
|
|
|
extern UINT64 ten2m3k64[];
|
|
extern unsigned int shift_ten2m3k64[];
|
|
extern UINT128 ten2m3k128[];
|
|
extern unsigned int shift_ten2m3k128[];
|
|
|
|
|
|
|
|
/***************************************************************************
|
|
*************** TABLES FOR GENERAL ROUNDING FUNCTIONS *********************
|
|
***************************************************************************/
|
|
|
|
extern UINT64 Kx64[];
|
|
extern unsigned int Ex64m64[];
|
|
extern UINT64 half64[];
|
|
extern UINT64 mask64[];
|
|
extern UINT64 ten2mxtrunc64[];
|
|
|
|
extern UINT128 Kx128[];
|
|
extern unsigned int Ex128m128[];
|
|
extern UINT64 half128[];
|
|
extern UINT64 mask128[];
|
|
extern UINT128 ten2mxtrunc128[];
|
|
|
|
extern UINT192 Kx192[];
|
|
extern unsigned int Ex192m192[];
|
|
extern UINT64 half192[];
|
|
extern UINT64 mask192[];
|
|
extern UINT192 ten2mxtrunc192[];
|
|
|
|
extern UINT256 Kx256[];
|
|
extern unsigned int Ex256m256[];
|
|
extern UINT64 half256[];
|
|
extern UINT64 mask256[];
|
|
extern UINT256 ten2mxtrunc256[];
|
|
|
|
typedef union __bid64_128 {
|
|
UINT64 b64;
|
|
UINT128 b128;
|
|
} BID64_128;
|
|
|
|
BID64_128 bid_fma (unsigned int P0,
|
|
BID64_128 x1, unsigned int P1,
|
|
BID64_128 y1, unsigned int P2,
|
|
BID64_128 z1, unsigned int P3,
|
|
unsigned int rnd_mode, FPSC * fpsc);
|
|
|
|
#define P16 16
|
|
#define P34 34
|
|
|
|
union __int_double {
|
|
UINT64 i;
|
|
double d;
|
|
};
|
|
typedef union __int_double int_double;
|
|
|
|
|
|
union __int_float {
|
|
UINT32 i;
|
|
float d;
|
|
};
|
|
typedef union __int_float int_float;
|
|
|
|
#define SWAP(A,B,T) {\
|
|
T = A; \
|
|
A = B; \
|
|
B = T; \
|
|
}
|
|
|
|
// this macro will find coefficient_x to be in [2^A, 2^(A+1) )
|
|
// ie it knows that it is A bits long
|
|
#define NUMBITS(A, coefficient_x, tempx){\
|
|
temp_x.d=(float)coefficient_x;\
|
|
A=((tempx.i >>23) & EXPONENT_MASK32) - 0x7f;\
|
|
}
|
|
|
|
enum class_types {
|
|
signalingNaN,
|
|
quietNaN,
|
|
negativeInfinity,
|
|
negativeNormal,
|
|
negativeSubnormal,
|
|
negativeZero,
|
|
positiveZero,
|
|
positiveSubnormal,
|
|
positiveNormal,
|
|
positiveInfinity
|
|
};
|
|
|
|
typedef union {
|
|
UINT64 ui64;
|
|
double d;
|
|
} BID_UI64DOUBLE;
|
|
|
|
#endif
|