gcc/libgcc/config/libbid/bid64_fma.c

512 lines
15 KiB
C

/* Copyright (C) 2007 Free Software Foundation, Inc.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify it under
the terms of the GNU General Public License as published by the Free
Software Foundation; either version 2, or (at your option) any later
version.
In addition to the permissions in the GNU General Public License, the
Free Software Foundation gives you unlimited permission to link the
compiled version of this file into combinations with other programs,
and to distribute those combinations without any restriction coming
from the use of this file. (The General Public License restrictions
do apply in other respects; for example, they cover modification of
the file, and distribution when not linked into a combine
executable.)
GCC is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
for more details.
You should have received a copy of the GNU General Public License
along with GCC; see the file COPYING. If not, write to the Free
Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
02110-1301, USA. */
/*****************************************************************************
* BID64 fma
*****************************************************************************
*
* Algorithm description:
*
* if multiplication is guranteed exact (short coefficients)
* call the unpacked arg. equivalent of bid64_add(x*y, z)
* else
* get full coefficient_x*coefficient_y product
* call subroutine to perform addition of 64-bit argument
* to 128-bit product
*
****************************************************************************/
#include "bid_inline_add.h"
#if DECIMAL_CALL_BY_REFERENCE
extern void bid64_mul (UINT64 * pres, UINT64 * px,
UINT64 *
py _RND_MODE_PARAM _EXC_FLAGS_PARAM
_EXC_MASKS_PARAM _EXC_INFO_PARAM);
#else
extern UINT64 bid64_mul (UINT64 x,
UINT64 y _RND_MODE_PARAM
_EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM);
#endif
#if DECIMAL_CALL_BY_REFERENCE
void
bid64_fma (UINT64 * pres, UINT64 * px, UINT64 * py,
UINT64 *
pz _RND_MODE_PARAM _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x, y, z;
#else
UINT64
bid64_fma (UINT64 x, UINT64 y,
UINT64 z _RND_MODE_PARAM _EXC_FLAGS_PARAM
_EXC_MASKS_PARAM _EXC_INFO_PARAM) {
#endif
UINT128 P, PU, CT, CZ;
UINT64 sign_x, sign_y, coefficient_x, coefficient_y, sign_z,
coefficient_z;
UINT64 C64, remainder_y, res;
UINT64 CYh, CY0L, T, valid_x, valid_y, valid_z;
int_double tempx, tempy;
int extra_digits, exponent_x, exponent_y, bin_expon_cx, bin_expon_cy,
bin_expon_product, rmode;
int digits_p, bp, final_exponent, exponent_z, digits_z, ez, ey,
scale_z, uf_status;
#if DECIMAL_CALL_BY_REFERENCE
#if !DECIMAL_GLOBAL_ROUNDING
_IDEC_round rnd_mode = *prnd_mode;
#endif
x = *px;
y = *py;
z = *pz;
#endif
valid_x = unpack_BID64 (&sign_x, &exponent_x, &coefficient_x, x);
valid_y = unpack_BID64 (&sign_y, &exponent_y, &coefficient_y, y);
valid_z = unpack_BID64 (&sign_z, &exponent_z, &coefficient_z, z);
// unpack arguments, check for NaN, Infinity, or 0
if (!valid_x || !valid_y || !valid_z) {
if ((y & MASK_NAN) == MASK_NAN) { // y is NAN
// if x = {0, f, inf, NaN}, y = NaN, z = {0, f, inf, NaN} then res = Q (y)
// check first for non-canonical NaN payload
y = y & 0xfe03ffffffffffffull; // clear G6-G12
if ((y & 0x0003ffffffffffffull) > 999999999999999ull) {
y = y & 0xfe00000000000000ull; // clear G6-G12 and the payload bits
}
if ((y & MASK_SNAN) == MASK_SNAN) { // y is SNAN
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return quiet (y)
res = y & 0xfdffffffffffffffull;
} else { // y is QNaN
// return y
res = y;
// if z = SNaN or x = SNaN signal invalid exception
if ((z & MASK_SNAN) == MASK_SNAN
|| (x & MASK_SNAN) == MASK_SNAN) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
}
}
BID_RETURN (res)
} else if ((z & MASK_NAN) == MASK_NAN) { // z is NAN
// if x = {0, f, inf, NaN}, y = {0, f, inf}, z = NaN then res = Q (z)
// check first for non-canonical NaN payload
z = z & 0xfe03ffffffffffffull; // clear G6-G12
if ((z & 0x0003ffffffffffffull) > 999999999999999ull) {
z = z & 0xfe00000000000000ull; // clear G6-G12 and the payload bits
}
if ((z & MASK_SNAN) == MASK_SNAN) { // z is SNAN
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return quiet (z)
res = z & 0xfdffffffffffffffull;
} else { // z is QNaN
// return z
res = z;
// if x = SNaN signal invalid exception
if ((x & MASK_SNAN) == MASK_SNAN) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
}
}
BID_RETURN (res)
} else if ((x & MASK_NAN) == MASK_NAN) { // x is NAN
// if x = NaN, y = {0, f, inf}, z = {0, f, inf} then res = Q (x)
// check first for non-canonical NaN payload
x = x & 0xfe03ffffffffffffull; // clear G6-G12
if ((x & 0x0003ffffffffffffull) > 999999999999999ull) {
x = x & 0xfe00000000000000ull; // clear G6-G12 and the payload bits
}
if ((x & MASK_SNAN) == MASK_SNAN) { // x is SNAN
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return quiet (x)
res = x & 0xfdffffffffffffffull;
} else { // x is QNaN
// return x
res = x; // clear out G[6]-G[16]
}
BID_RETURN (res)
}
if (!valid_x) {
// x is Inf. or 0
// x is Infinity?
if ((x & 0x7800000000000000ull) == 0x7800000000000000ull) {
// check if y is 0
if (!coefficient_y) {
// y==0, return NaN
#ifdef SET_STATUS_FLAGS
if ((z & 0x7e00000000000000ull) != 0x7c00000000000000ull)
__set_status_flags (pfpsf, INVALID_EXCEPTION);
#endif
BID_RETURN (0x7c00000000000000ull);
}
// test if z is Inf of oposite sign
if (((z & 0x7c00000000000000ull) == 0x7800000000000000ull)
&& (((x ^ y) ^ z) & 0x8000000000000000ull)) {
// return NaN
#ifdef SET_STATUS_FLAGS
__set_status_flags (pfpsf, INVALID_EXCEPTION);
#endif
BID_RETURN (0x7c00000000000000ull);
}
// otherwise return +/-Inf
BID_RETURN (((x ^ y) & 0x8000000000000000ull) |
0x7800000000000000ull);
}
// x is 0
if (((y & 0x7800000000000000ull) != 0x7800000000000000ull)
&& ((z & 0x7800000000000000ull) != 0x7800000000000000ull)) {
if (coefficient_z) {
exponent_y = exponent_x - DECIMAL_EXPONENT_BIAS + exponent_y;
sign_z = z & 0x8000000000000000ull;
if (exponent_y >= exponent_z)
BID_RETURN (z);
res =
add_zero64 (exponent_y, sign_z, exponent_z, coefficient_z,
&rnd_mode, pfpsf);
BID_RETURN (res);
}
}
}
if (!valid_y) {
// y is Inf. or 0
// y is Infinity?
if ((y & 0x7800000000000000ull) == 0x7800000000000000ull) {
// check if x is 0
if (!coefficient_x) {
// y==0, return NaN
#ifdef SET_STATUS_FLAGS
__set_status_flags (pfpsf, INVALID_EXCEPTION);
#endif
BID_RETURN (0x7c00000000000000ull);
}
// test if z is Inf of oposite sign
if (((z & 0x7c00000000000000ull) == 0x7800000000000000ull)
&& (((x ^ y) ^ z) & 0x8000000000000000ull)) {
#ifdef SET_STATUS_FLAGS
__set_status_flags (pfpsf, INVALID_EXCEPTION);
#endif
// return NaN
BID_RETURN (0x7c00000000000000ull);
}
// otherwise return +/-Inf
BID_RETURN (((x ^ y) & 0x8000000000000000ull) |
0x7800000000000000ull);
}
// y is 0
if (((z & 0x7800000000000000ull) != 0x7800000000000000ull)) {
if (coefficient_z) {
exponent_y += exponent_x - DECIMAL_EXPONENT_BIAS;
sign_z = z & 0x8000000000000000ull;
if (exponent_y >= exponent_z)
BID_RETURN (z);
res =
add_zero64 (exponent_y, sign_z, exponent_z, coefficient_z,
&rnd_mode, pfpsf);
BID_RETURN (res);
}
}
}
if (!valid_z) {
// y is Inf. or 0
// test if y is NaN/Inf
if ((z & 0x7800000000000000ull) == 0x7800000000000000ull) {
BID_RETURN (coefficient_z & QUIET_MASK64);
}
// z is 0, return x*y
if ((!coefficient_x) || (!coefficient_y)) {
//0+/-0
exponent_x += exponent_y - DECIMAL_EXPONENT_BIAS;
if (exponent_x > DECIMAL_MAX_EXPON_64)
exponent_x = DECIMAL_MAX_EXPON_64;
else if (exponent_x < 0)
exponent_x = 0;
if (exponent_x <= exponent_z)
res = ((UINT64) exponent_x) << 53;
else
res = ((UINT64) exponent_z) << 53;
if ((sign_x ^ sign_y) == sign_z)
res |= sign_z;
#ifndef IEEE_ROUND_NEAREST_TIES_AWAY
#ifndef IEEE_ROUND_NEAREST
else if (rnd_mode == ROUNDING_DOWN)
res |= 0x8000000000000000ull;
#endif
#endif
BID_RETURN (res);
}
}
}
/* get binary coefficients of x and y */
//--- get number of bits in the coefficients of x and y ---
// version 2 (original)
tempx.d = (double) coefficient_x;
bin_expon_cx = ((tempx.i & MASK_BINARY_EXPONENT) >> 52);
tempy.d = (double) coefficient_y;
bin_expon_cy = ((tempy.i & MASK_BINARY_EXPONENT) >> 52);
// magnitude estimate for coefficient_x*coefficient_y is
// 2^(unbiased_bin_expon_cx + unbiased_bin_expon_cx)
bin_expon_product = bin_expon_cx + bin_expon_cy;
// check if coefficient_x*coefficient_y<2^(10*k+3)
// equivalent to unbiased_bin_expon_cx + unbiased_bin_expon_cx < 10*k+1
if (bin_expon_product < UPPER_EXPON_LIMIT + 2 * BINARY_EXPONENT_BIAS) {
// easy multiply
C64 = coefficient_x * coefficient_y;
final_exponent = exponent_x + exponent_y - DECIMAL_EXPONENT_BIAS;
if ((final_exponent > 0) || (!coefficient_z)) {
res =
get_add64 (sign_x ^ sign_y,
final_exponent, C64, sign_z, exponent_z, coefficient_z, rnd_mode, pfpsf);
BID_RETURN (res);
} else {
P.w[0] = C64;
P.w[1] = 0;
extra_digits = 0;
}
} else {
if (!coefficient_z) {
#if DECIMAL_CALL_BY_REFERENCE
bid64_mul (&res, px,
py _RND_MODE_ARG _EXC_FLAGS_ARG _EXC_MASKS_ARG
_EXC_INFO_ARG);
#else
res =
bid64_mul (x,
y _RND_MODE_ARG _EXC_FLAGS_ARG _EXC_MASKS_ARG
_EXC_INFO_ARG);
#endif
BID_RETURN (res);
}
// get 128-bit product: coefficient_x*coefficient_y
__mul_64x64_to_128 (P, coefficient_x, coefficient_y);
// tighten binary range of P: leading bit is 2^bp
// unbiased_bin_expon_product <= bp <= unbiased_bin_expon_product+1
bin_expon_product -= 2 * BINARY_EXPONENT_BIAS;
__tight_bin_range_128 (bp, P, bin_expon_product);
// get number of decimal digits in the product
digits_p = estimate_decimal_digits[bp];
if (!(__unsigned_compare_gt_128 (power10_table_128[digits_p], P)))
digits_p++; // if power10_table_128[digits_p] <= P
// determine number of decimal digits to be rounded out
extra_digits = digits_p - MAX_FORMAT_DIGITS;
final_exponent =
exponent_x + exponent_y + extra_digits - DECIMAL_EXPONENT_BIAS;
}
if (((unsigned) final_exponent) >= 3 * 256) {
if (final_exponent < 0) {
//--- get number of bits in the coefficients of z ---
tempx.d = (double) coefficient_z;
bin_expon_cx = ((tempx.i & MASK_BINARY_EXPONENT) >> 52) - 0x3ff;
// get number of decimal digits in the coeff_x
digits_z = estimate_decimal_digits[bin_expon_cx];
if (coefficient_z >= power10_table_128[digits_z].w[0])
digits_z++;
// underflow
if ((final_exponent + 16 < 0)
|| (exponent_z + digits_z > 33 + final_exponent)) {
res =
BID_normalize (sign_z, exponent_z, coefficient_z,
sign_x ^ sign_y, 1, rnd_mode, pfpsf);
BID_RETURN (res);
}
ez = exponent_z + digits_z - 16;
if (ez < 0)
ez = 0;
scale_z = exponent_z - ez;
coefficient_z *= power10_table_128[scale_z].w[0];
ey = final_exponent - extra_digits;
extra_digits = ez - ey;
if (extra_digits > 33) {
res =
BID_normalize (sign_z, exponent_z, coefficient_z,
sign_x ^ sign_y, 1, rnd_mode, pfpsf);
BID_RETURN (res);
}
//else // extra_digits<=32
if (extra_digits > 17) {
CYh = __truncate (P, 16);
// get remainder
T = power10_table_128[16].w[0];
__mul_64x64_to_64 (CY0L, CYh, T);
remainder_y = P.w[0] - CY0L;
extra_digits -= 16;
P.w[0] = CYh;
P.w[1] = 0;
} else
remainder_y = 0;
// align coeff_x, CYh
__mul_64x64_to_128 (CZ, coefficient_z,
power10_table_128[extra_digits].w[0]);
if (sign_z == (sign_y ^ sign_x)) {
__add_128_128 (CT, CZ, P);
if (__unsigned_compare_ge_128
(CT, power10_table_128[16 + extra_digits])) {
extra_digits++;
ez++;
}
} else {
if (remainder_y && (__unsigned_compare_ge_128 (CZ, P))) {
P.w[0]++;
if (!P.w[0])
P.w[1]++;
}
__sub_128_128 (CT, CZ, P);
if (((SINT64) CT.w[1]) < 0) {
sign_z = sign_y ^ sign_x;
CT.w[0] = 0 - CT.w[0];
CT.w[1] = 0 - CT.w[1];
if (CT.w[0])
CT.w[1]--;
} else if(!(CT.w[1]|CT.w[0]))
sign_z = (rnd_mode!=ROUNDING_DOWN)? 0: 0x8000000000000000ull;
if (ez
&&
(__unsigned_compare_gt_128
(power10_table_128[15 + extra_digits], CT))) {
extra_digits--;
ez--;
}
}
#ifdef SET_STATUS_FLAGS
uf_status = 0;
if ((!ez)
&&
__unsigned_compare_gt_128 (power10_table_128
[extra_digits + 15], CT)) {
rmode = rnd_mode;
if (sign_z && (unsigned) (rmode - 1) < 2)
rmode = 3 - rmode;
//__add_128_64(PU, CT, round_const_table[rmode][extra_digits]);
PU = power10_table_128[extra_digits + 15];
PU.w[0]--;
if (__unsigned_compare_gt_128 (PU, CT)
|| (rmode == ROUNDING_DOWN)
|| (rmode == ROUNDING_TO_ZERO))
uf_status = UNDERFLOW_EXCEPTION;
else if (extra_digits < 2) {
if ((rmode == ROUNDING_UP)) {
if (!extra_digits)
uf_status = UNDERFLOW_EXCEPTION;
else {
if (remainder_y && (sign_z != (sign_y ^ sign_x)))
remainder_y = power10_table_128[16].w[0] - remainder_y;
if (power10_table_128[15].w[0] > remainder_y)
uf_status = UNDERFLOW_EXCEPTION;
}
} else // RN or RN_away
{
if (remainder_y && (sign_z != (sign_y ^ sign_x)))
remainder_y = power10_table_128[16].w[0] - remainder_y;
if (!extra_digits) {
remainder_y += round_const_table[rmode][15];
if (remainder_y < power10_table_128[16].w[0])
uf_status = UNDERFLOW_EXCEPTION;
} else {
if (remainder_y < round_const_table[rmode][16])
uf_status = UNDERFLOW_EXCEPTION;
}
}
//__set_status_flags (pfpsf, uf_status);
}
}
#endif
res =
__bid_full_round64_remainder (sign_z, ez - extra_digits, CT,
extra_digits, remainder_y,
rnd_mode, pfpsf, uf_status);
BID_RETURN (res);
} else {
if ((sign_z == (sign_x ^ sign_y))
|| (final_exponent > 3 * 256 + 15)) {
res =
fast_get_BID64_check_OF (sign_x ^ sign_y, final_exponent,
1000000000000000ull, rnd_mode,
pfpsf);
BID_RETURN (res);
}
}
}
if (extra_digits > 0) {
res =
get_add128 (sign_z, exponent_z, coefficient_z, sign_x ^ sign_y,
final_exponent, P, extra_digits, rnd_mode, pfpsf);
BID_RETURN (res);
}
// go to convert_format and exit
else {
C64 = __low_64 (P);
res =
get_add64 (sign_x ^ sign_y,
exponent_x + exponent_y - DECIMAL_EXPONENT_BIAS, C64,
sign_z, exponent_z, coefficient_z,
rnd_mode, pfpsf);
BID_RETURN (res);
}
}