aboutsummaryrefslogtreecommitdiff
path: root/libgcc/config/libbid/bid64_to_int32.c
diff options
context:
space:
mode:
authorAndrea Ornstein <andrea.ornstein@st.com>2009-04-20 16:01:58 +0000
committerAndrea Ornstein <andrea.ornstein@st.com>2009-04-20 16:01:58 +0000
commite916c2f6d80dbd955333bed4b12639fcc1ebc52e (patch)
tree46ac7b490c11d8396b9b7ff4197411626eb42c47 /libgcc/config/libbid/bid64_to_int32.c
parent2eaa0a1672c2d87d390584a93bacefa2c0c14b1d (diff)
merge from trunk revisions 127002-132392, 7/N
git-svn-id: https://gcc.gnu.org/svn/gcc/branches/st/cli-be@146438 138bc75d-0d04-0410-961f-82ee72b054a4
Diffstat (limited to 'libgcc/config/libbid/bid64_to_int32.c')
-rw-r--r--libgcc/config/libbid/bid64_to_int32.c1132
1 files changed, 565 insertions, 567 deletions
diff --git a/libgcc/config/libbid/bid64_to_int32.c b/libgcc/config/libbid/bid64_to_int32.c
index d0213200933..b4e62371b6a 100644
--- a/libgcc/config/libbid/bid64_to_int32.c
+++ b/libgcc/config/libbid/bid64_to_int32.c
@@ -34,27 +34,27 @@ Software Foundation, 51 Franklin Street, Fifth Floor, Boston, MA
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_rnint (int *pres,
+bid64_to_int32_rnint (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -67,17 +67,17 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -91,50 +91,50 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n < -2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) > 2^31+1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x500000005, 1<=q<=16
// <=> C * 10^(11-q) > 0x500000005, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 > 0x500000005ull) {
// set invalid flag
@@ -145,12 +145,12 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) > 0x500000005 <=>
// C > 0x500000005 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1/2 up)
// Note: 0x500000005*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000005ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000005ull * ten2k64[q - 11];
if (C1 > tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -161,14 +161,14 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31-1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x4fffffffb, 1<=q<=16
// <=> C * 10^(11-q) >= 0x4fffffffb, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x4fffffffbull) {
// set invalid flag
@@ -179,12 +179,12 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x4fffffffb <=>
// C >= 0x4fffffffb * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1/2 up)
// Note: 0x4fffffffb*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x4fffffffbull * __bid_ten2k64[q - 11];
+ tmp64 = 0x4fffffffbull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -199,45 +199,45 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1/2 <= n < 2^31 - 1/2
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
+ if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
// return 0
res = 0x00000000;
BID_RETURN (res);
- } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
+ } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
// if 0.c(0)c(1)...c(q-1) <= 0.5 <=> c(0)c(1)...c(q-1) <= 5 * 10^(q-1)
// res = 0
// else
// res = +/-1
ind = q - 1;
- if (C1 <= __bid_midpoint64[ind]) {
- res = 0x00000000; // return 0
- } else if (x_sign) { // n < 0
- res = 0xffffffff; // return -1
- } else { // n > 0
- res = 0x00000001; // return +1
+ if (C1 <= midpoint64[ind]) {
+ res = 0x00000000; // return 0
+ } else if (x_sign) { // n < 0
+ res = 0xffffffff; // return -1
+ } else { // n > 0
+ res = 0x00000001; // return +1
}
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1/2 <= x <= -1 or 1 <= x < 2^31-1/2 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 = C1 + 1/2 * 10^ind where the result C1 fits in 64 bits
- C1 = C1 + __bid_midpoint64[ind - 1];
+ C1 = C1 + midpoint64[ind - 1];
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = (C1 + 1/2 * 10^x) * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// if (0 < f* < 10^(-x)) then the result is a midpoint
// if floor(C*) is even then C* = floor(C*) - logical right
// shift; C* has p decimal digits, correct by Prop. 1)
@@ -248,22 +248,22 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// if the result was a midpoint it was rounded away from zero, so
// it will need a correction
// check for midpoints
if ((fstar.w[1] == 0) && fstar.w[0]
- && (fstar.w[0] <= __bid_ten2mk128trunc[ind - 1].w[1])) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ && (fstar.w[0] <= ten2mk128trunc[ind - 1].w[1])) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// the result is a midpoint; round to nearest
- if (Cstar & 0x01) { // Cstar is odd; MP in [EVEN, ODD]
+ if (Cstar & 0x01) { // Cstar is odd; MP in [EVEN, ODD]
// if floor(C*) is odd C = floor(C*) - 1; the result >= 1
- Cstar--; // Cstar is now even
- } // else MP in [ODD, EVEN]
+ Cstar--; // Cstar is now even
+ } // else MP in [ODD, EVEN]
}
if (x_sign)
res = -Cstar;
@@ -276,12 +276,12 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -293,27 +293,27 @@ __bid64_to_int32_rnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_xrnint (int *pres,
+bid64_to_int32_xrnint (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -326,17 +326,17 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -350,50 +350,50 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n < -2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) > 2^31+1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x500000005, 1<=q<=16
// <=> C * 10^(11-q) > 0x500000005, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 > 0x500000005ull) {
// set invalid flag
@@ -404,12 +404,12 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) > 0x500000005 <=>
// C > 0x500000005 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1/2 up)
// Note: 0x500000005*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000005ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000005ull * ten2k64[q - 11];
if (C1 > tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -420,14 +420,14 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31-1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x4fffffffb, 1<=q<=16
// <=> C * 10^(11-q) >= 0x4fffffffb, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x4fffffffbull) {
// set invalid flag
@@ -438,12 +438,12 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x4fffffffb <=>
// C >= 0x4fffffffb * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1/2 up)
// Note: 0x4fffffffb*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x4fffffffbull * __bid_ten2k64[q - 11];
+ tmp64 = 0x4fffffffbull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -458,49 +458,49 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1/2 < n < 2^31 - 1/2
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
+ if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
// return 0
res = 0x00000000;
BID_RETURN (res);
- } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
+ } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
// if 0.c(0)c(1)...c(q-1) <= 0.5 <=> c(0)c(1)...c(q-1) <= 5 * 10^(q-1)
// res = 0
// else
// res = +/-1
ind = q - 1;
- if (C1 <= __bid_midpoint64[ind]) {
- res = 0x00000000; // return 0
- } else if (x_sign) { // n < 0
- res = 0xffffffff; // return -1
- } else { // n > 0
- res = 0x00000001; // return +1
+ if (C1 <= midpoint64[ind]) {
+ res = 0x00000000; // return 0
+ } else if (x_sign) { // n < 0
+ res = 0xffffffff; // return -1
+ } else { // n > 0
+ res = 0x00000001; // return +1
}
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1/2 <= x <= -1 or 1 <= x < 2^31-1/2 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 = C1 + 1/2 * 10^ind where the result C1 fits in 64 bits
- C1 = C1 + __bid_midpoint64[ind - 1];
+ C1 = C1 + midpoint64[ind - 1];
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = (C1 + 1/2 * 10^x) * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// if (0 < f* < 10^(-x)) then the result is a midpoint
// if floor(C*) is even then C* = floor(C*) - logical right
// shift; C* has p decimal digits, correct by Prop. 1)
@@ -511,8 +511,8 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* - 1/2 < 10^(-x)) then
@@ -522,31 +522,30 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
if (ind - 1 <= 2) {
if (fstar.w[0] > 0x8000000000000000ull) {
// f* > 1/2 and the result may be exact
- tmp64 = fstar.w[0] - 0x8000000000000000ull; // f* - 1/2
- if ((tmp64 > __bid_ten2mk128trunc[ind - 1].w[1])) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ tmp64 = fstar.w[0] - 0x8000000000000000ull; // f* - 1/2
+ if ((tmp64 > ten2mk128trunc[ind - 1].w[1])) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // the result is inexact; f2* <= 1/2
+ } // else the result is exact
+ } else { // the result is inexact; f2* <= 1/2
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
}
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] > __bid_one_half128[ind - 1]
- || (fstar.w[1] == __bid_one_half128[ind - 1]
- && fstar.w[0])) {
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] > onehalf128[ind - 1] ||
+ (fstar.w[1] == onehalf128[ind - 1] && fstar.w[0])) {
// f2* > 1/2 and the result may be exact
// Calculate f2* - 1/2
- tmp64 = fstar.w[1] - __bid_one_half128[ind - 1];
- if (tmp64 || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ tmp64 = fstar.w[1] - onehalf128[ind - 1];
+ if (tmp64 || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // the result is inexact; f2* <= 1/2
+ } // else the result is exact
+ } else { // the result is inexact; f2* <= 1/2
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
}
@@ -556,14 +555,14 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// it will need a correction
// check for midpoints
if ((fstar.w[1] == 0) && fstar.w[0]
- && (fstar.w[0] <= __bid_ten2mk128trunc[ind - 1].w[1])) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ && (fstar.w[0] <= ten2mk128trunc[ind - 1].w[1])) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// the result is a midpoint; round to nearest
- if (Cstar & 0x01) { // Cstar is odd; MP in [EVEN, ODD]
+ if (Cstar & 0x01) { // Cstar is odd; MP in [EVEN, ODD]
// if floor(C*) is odd C = floor(C*) - 1; the result >= 1
- Cstar--; // Cstar is now even
- } // else MP in [ODD, EVEN]
+ Cstar--; // Cstar is now even
+ } // else MP in [ODD, EVEN]
}
if (x_sign)
res = -Cstar;
@@ -576,12 +575,12 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -593,27 +592,27 @@ __bid64_to_int32_xrnint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_floor (int *pres,
+bid64_to_int32_floor (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -626,17 +625,17 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -650,50 +649,50 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n < -2^31 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) > 2^31
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x500000000, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000000, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000000 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 > 0x500000000ull) {
// set invalid flag
@@ -704,12 +703,12 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) > 0x500000000 <=>
// C > 0x500000000 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1 up)
// Note: 0x500000000*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000000ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000000ull * ten2k64[q - 11];
if (C1 > tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -720,14 +719,14 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x500000000, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000000, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000000 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x500000000ull) {
// set invalid flag
@@ -738,12 +737,12 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x500000000 <=>
// C >= 0x500000000 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1 up)
// Note: 0x500000000*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000000ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000000ull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -758,40 +757,40 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 <= n < 2^31
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
+ if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
// return -1 or 0
if (x_sign)
res = 0xffffffff;
else
res = 0x00000000;
BID_RETURN (res);
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1 < x <= -1 or 1 <= x < 2^31 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 fits in 64 bits
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = C1 * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*) (logical right shift; C has p decimal digits,
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* < 10^(-x)) then
@@ -799,21 +798,21 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else // if (f* > T*) then
// the result is inexact
if (ind - 1 <= 2) {
- if (fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (x_sign) { // negative and inexact
+ if (fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (x_sign) { // negative and inexact
Cstar++;
}
- } // else the result is exact
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (x_sign) { // negative and inexact
+ } // else the result is exact
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (x_sign) { // negative and inexact
Cstar++;
}
- } // else the result is exact
+ } // else the result is exact
}
if (x_sign)
@@ -827,12 +826,12 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -844,27 +843,27 @@ __bid64_to_int32_floor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_xfloor (int *pres,
+bid64_to_int32_xfloor (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -877,17 +876,17 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -901,50 +900,50 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n < -2^31 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) > 2^31
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x500000000, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000000, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000000 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 > 0x500000000ull) {
// set invalid flag
@@ -955,12 +954,12 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) > 0x500000000 <=>
// C > 0x500000000 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1 up)
// Note: 0x500000000*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000000ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000000ull * ten2k64[q - 11];
if (C1 > tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -971,14 +970,14 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x500000000, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000000, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000000 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x500000000ull) {
// set invalid flag
@@ -989,12 +988,12 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x500000000 <=>
// C >= 0x500000000 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1 up)
// Note: 0x500000000*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000000ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000000ull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1009,7 +1008,7 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 <= n < 2^31
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
+ if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
// return -1 or 0
@@ -1018,33 +1017,33 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
else
res = 0x00000000;
BID_RETURN (res);
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1 < x <= -1 or 1 <= x < 2^31 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 fits in 64 bits
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = C1 * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*) (logical right shift; C has p decimal digits,
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* < 10^(-x)) then
@@ -1052,25 +1051,25 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else // if (f* > T*) then
// the result is inexact
if (ind - 1 <= 2) {
- if (fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (x_sign) { // negative and inexact
+ if (fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (x_sign) { // negative and inexact
Cstar++;
}
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (x_sign) { // negative and inexact
+ } // else the result is exact
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (x_sign) { // negative and inexact
Cstar++;
}
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
+ } // else the result is exact
}
if (x_sign)
@@ -1084,12 +1083,12 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -1101,27 +1100,27 @@ __bid64_to_int32_xfloor (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_ceil (int *pres,
+bid64_to_int32_ceil (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -1134,17 +1133,17 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -1158,50 +1157,50 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n <= -2^31 - 1 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31+1
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x50000000a, 1<=q<=16
// <=> C * 10^(11-q) >= 0x50000000a, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x50000000a has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x50000000aull) {
// set invalid flag
@@ -1212,12 +1211,12 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x50000000a <=>
// C >= 0x50000000a * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1 up)
// Note: 0x50000000a*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x50000000aull * __bid_ten2k64[q - 11];
+ tmp64 = 0x50000000aull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1228,14 +1227,14 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n > 2^31 - 1 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) > 2^31 - 1
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x4fffffff6, 1<=q<=16
// <=> C * 10^(11-q) > 0x4fffffff6, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x4fffffff6 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 > 0x4fffffff6ull) {
// set invalid flag
@@ -1246,12 +1245,12 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) > 0x4fffffff6 <=>
// C > 0x4fffffff6 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1 up)
// Note: 0x4fffffff6*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x4fffffff6ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x4fffffff6ull * ten2k64[q - 11];
if (C1 > tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1266,40 +1265,40 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1 < n <= 2^31 - 1
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
+ if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
// return 0 or 1
if (x_sign)
res = 0x00000000;
else
res = 0x00000001;
BID_RETURN (res);
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1 < x <= -1 or 1 <= x <= 2^31-1 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 fits in 64 bits
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = C1 * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*) (logical right shift; C has p decimal digits,
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* < 10^(-x)) then
@@ -1307,21 +1306,21 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else // if (f* > T*) then
// the result is inexact
if (ind - 1 <= 2) {
- if (fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (!x_sign) { // positive and inexact
+ if (fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (!x_sign) { // positive and inexact
Cstar++;
}
- } // else the result is exact
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (!x_sign) { // positive and inexact
+ } // else the result is exact
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (!x_sign) { // positive and inexact
Cstar++;
}
- } // else the result is exact
+ } // else the result is exact
}
if (x_sign)
@@ -1335,12 +1334,12 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -1352,27 +1351,27 @@ __bid64_to_int32_ceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_xceil (int *pres,
+bid64_to_int32_xceil (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -1385,17 +1384,17 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -1409,50 +1408,50 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n <= -2^31 - 1 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31+1
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x50000000a, 1<=q<=16
// <=> C * 10^(11-q) >= 0x50000000a, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x50000000a has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x50000000aull) {
// set invalid flag
@@ -1463,12 +1462,12 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x50000000a <=>
// C >= 0x50000000a * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1 up)
// Note: 0x50000000a*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x50000000aull * __bid_ten2k64[q - 11];
+ tmp64 = 0x50000000aull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1479,14 +1478,14 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n > 2^31 - 1 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) > 2^31 - 1
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x4fffffff6, 1<=q<=16
// <=> C * 10^(11-q) > 0x4fffffff6, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x4fffffff6 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 > 0x4fffffff6ull) {
// set invalid flag
@@ -1497,12 +1496,12 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) > 0x4fffffff6 <=>
// C > 0x4fffffff6 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1 up)
// Note: 0x4fffffff6*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x4fffffff6ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x4fffffff6ull * ten2k64[q - 11];
if (C1 > tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1517,7 +1516,7 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1 < n <= 2^31 - 1
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
+ if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
// return 0 or 1
@@ -1526,33 +1525,33 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
else
res = 0x00000001;
BID_RETURN (res);
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1 < x <= -1 or 1 <= x <= 2^31-1 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 fits in 64 bits
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = C1 * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*) (logical right shift; C has p decimal digits,
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* < 10^(-x)) then
@@ -1560,25 +1559,25 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else // if (f* > T*) then
// the result is inexact
if (ind - 1 <= 2) {
- if (fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (!x_sign) { // positive and inexact
+ if (fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (!x_sign) { // positive and inexact
Cstar++;
}
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
- if (!x_sign) { // positive and inexact
+ } // else the result is exact
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
+ if (!x_sign) { // positive and inexact
Cstar++;
}
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
+ } // else the result is exact
}
if (x_sign)
@@ -1592,12 +1591,12 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -1609,27 +1608,27 @@ __bid64_to_int32_xceil (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_int (int *pres,
+bid64_to_int32_int (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 P128;
// check for NaN or Infinity
@@ -1641,17 +1640,17 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -1665,50 +1664,50 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n <= -2^31 - 1 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31+1
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x50000000a, 1<=q<=16
// <=> C * 10^(11-q) >= 0x50000000a, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x50000000a has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x50000000aull) {
// set invalid flag
@@ -1719,12 +1718,12 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x50000000a <=>
// C >= 0x50000000a * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1 up)
// Note: 0x50000000a*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x50000000aull * __bid_ten2k64[q - 11];
+ tmp64 = 0x50000000aull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1735,14 +1734,14 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x500000000, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000000, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000000 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x500000000ull) {
// set invalid flag
@@ -1753,12 +1752,12 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x500000000 <=>
// C >= 0x500000000 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1 up)
// Note: 0x500000000*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000000ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000000ull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1773,35 +1772,35 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1 < n < 2^31
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
+ if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
// return 0
res = 0x00000000;
BID_RETURN (res);
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1 < x <= -1 or 1 <= x < 2^31 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 fits in 64 bits
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = C1 * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*) (logical right shift; C has p decimal digits,
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
if (x_sign)
res = -Cstar;
@@ -1814,12 +1813,12 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -1831,27 +1830,27 @@ __bid64_to_int32_int (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_xint (int *pres,
+bid64_to_int32_xint (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -1864,17 +1863,17 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -1888,50 +1887,50 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n <= -2^31 - 1 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31+1
// <=> 0.c(0)c(1)...c(q-1) * 10^11 > 0x50000000a, 1<=q<=16
// <=> C * 10^(11-q) >= 0x50000000a, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x50000000a has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x50000000aull) {
// set invalid flag
@@ -1942,12 +1941,12 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x50000000a <=>
// C >= 0x50000000a * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1 up)
// Note: 0x50000000a*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x50000000aull * __bid_ten2k64[q - 11];
+ tmp64 = 0x50000000aull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1958,14 +1957,14 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x500000000, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000000, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000000 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x500000000ull) {
// set invalid flag
@@ -1976,12 +1975,12 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x500000000 <=>
// C >= 0x500000000 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1 up)
// Note: 0x500000000*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000000ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000000ull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -1996,39 +1995,39 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1 < n < 2^31
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
+ if ((q + exp) <= 0) { // n = +/-0.[0...0]c(0)c(1)...c(q-1)
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
// return 0
res = 0x00000000;
BID_RETURN (res);
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1 < x <= -1 or 1 <= x < 2^31 so x can be rounded
// to nearest to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 fits in 64 bits
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = C1 * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*) (logical right shift; C has p decimal digits,
// correct by Property 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* < 10^(-x)) then
@@ -2036,19 +2035,19 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else // if (f* > T*) then
// the result is inexact
if (ind - 1 <= 2) {
- if (fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ if (fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ } // else the result is exact
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
+ } // else the result is exact
}
if (x_sign)
@@ -2062,12 +2061,12 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -2079,27 +2078,27 @@ __bid64_to_int32_xint (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_rninta (int *pres,
+bid64_to_int32_rninta (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 P128;
// check for NaN or Infinity
@@ -2111,17 +2110,17 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -2135,50 +2134,50 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n <= -2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31+1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x500000005, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000005, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x500000005ull) {
// set invalid flag
@@ -2189,12 +2188,12 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x500000005 <=>
// C >= 0x500000005 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1/2 up)
// Note: 0x500000005*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000005ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000005ull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -2205,14 +2204,14 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31-1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x4fffffffb, 1<=q<=16
// <=> C * 10^(11-q) >= 0x4fffffffb, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x4fffffffbull) {
// set invalid flag
@@ -2223,12 +2222,12 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x4fffffffb <=>
// C >= 0x4fffffffb * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1/2 up)
// Note: 0x4fffffffb*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x4fffffffbull * __bid_ten2k64[q - 11];
+ tmp64 = 0x4fffffffbull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -2243,49 +2242,49 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1/2 < n < 2^31 - 1/2
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
+ if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
// return 0
res = 0x00000000;
BID_RETURN (res);
- } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
+ } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
// if 0.c(0)c(1)...c(q-1) <= 0.5 <=> c(0)c(1)...c(q-1) <= 5 * 10^(q-1)
// res = 0
// else
// res = +/-1
ind = q - 1;
- if (C1 < __bid_midpoint64[ind]) {
- res = 0x00000000; // return 0
- } else if (x_sign) { // n < 0
- res = 0xffffffff; // return -1
- } else { // n > 0
- res = 0x00000001; // return +1
+ if (C1 < midpoint64[ind]) {
+ res = 0x00000000; // return 0
+ } else if (x_sign) { // n < 0
+ res = 0xffffffff; // return -1
+ } else { // n > 0
+ res = 0x00000001; // return +1
}
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1/2 <= x <= -1 or 1 <= x < 2^31-1/2 so x can be rounded
// to nearest away to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 = C1 + 1/2 * 10^ind where the result C1 fits in 64 bits
- C1 = C1 + __bid_midpoint64[ind - 1];
+ C1 = C1 + midpoint64[ind - 1];
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = (C1 + 1/2 * 10^x) * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*)-1 (logical right shift; C* has p decimal digits,
// correct by Pr. 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// if the result was a midpoint it was rounded away from zero
@@ -2300,12 +2299,12 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);
@@ -2317,27 +2316,27 @@ __bid64_to_int32_rninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
#if DECIMAL_CALL_BY_REFERENCE
void
-__bid64_to_int32_xrninta (int *pres,
+bid64_to_int32_xrninta (int *pres,
UINT64 *
px _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
UINT64 x = *px;
#else
int
-__bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
+bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
_EXC_INFO_PARAM) {
#endif
int res;
UINT64 x_sign;
UINT64 x_exp;
- int exp; // unbiased exponent
+ int exp; // unbiased exponent
// Note: C1 represents x_significand (UINT64)
UINT64 tmp64;
BID_UI64DOUBLE tmp1;
unsigned int x_nr_bits;
int q, ind, shift;
UINT64 C1;
- UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
+ UINT64 Cstar; // C* represents up to 16 decimal digits ~ 54 bits
UINT128 fstar;
UINT128 P128;
@@ -2350,17 +2349,17 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
BID_RETURN (res);
}
// unpack x
- x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
+ x_sign = x & MASK_SIGN; // 0 for positive, MASK_SIGN for negative
// if steering bits are 11 (condition will be 0), then exponent is G[0:w+1] =>
if ((x & MASK_STEERING_BITS) == MASK_STEERING_BITS) {
- x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT2) >> 51; // biased
C1 = (x & MASK_BINARY_SIG2) | MASK_BINARY_OR2;
- if (C1 > 9999999999999999ull) { // non-canonical
+ if (C1 > 9999999999999999ull) { // non-canonical
x_exp = 0;
C1 = 0;
}
} else {
- x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
+ x_exp = (x & MASK_BINARY_EXPONENT1) >> 53; // biased
C1 = x & MASK_BINARY_SIG1;
}
@@ -2374,50 +2373,50 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// q = nr. of decimal digits in x (1 <= q <= 54)
// determine first the nr. of bits in x
- if (C1 >= 0x0020000000000000ull) { // x >= 2^53
+ if (C1 >= 0x0020000000000000ull) { // x >= 2^53
// split the 64-bit value in two 32-bit halves to avoid rounding errors
- if (C1 >= 0x0000000100000000ull) { // x >= 2^32
- tmp1.d = (double) (C1 >> 32); // exact conversion
+ if (C1 >= 0x0000000100000000ull) { // x >= 2^32
+ tmp1.d = (double) (C1 >> 32); // exact conversion
x_nr_bits =
33 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
- } else { // x < 2^32
- tmp1.d = (double) C1; // exact conversion
+ } else { // x < 2^32
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- } else { // if x < 2^53
- tmp1.d = (double) C1; // exact conversion
+ } else { // if x < 2^53
+ tmp1.d = (double) C1; // exact conversion
x_nr_bits =
1 + ((((unsigned int) (tmp1.ui64 >> 52)) & 0x7ff) - 0x3ff);
}
- q = __bid_nr_digits[x_nr_bits - 1].digits;
+ q = nr_digits[x_nr_bits - 1].digits;
if (q == 0) {
- q = __bid_nr_digits[x_nr_bits - 1].digits1;
- if (C1 >= __bid_nr_digits[x_nr_bits - 1].threshold_lo)
+ q = nr_digits[x_nr_bits - 1].digits1;
+ if (C1 >= nr_digits[x_nr_bits - 1].threshold_lo)
q++;
}
- exp = x_exp - 398; // unbiased exponent
+ exp = x_exp - 398; // unbiased exponent
- if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
+ if ((q + exp) > 10) { // x >= 10^10 ~= 2^33.2... (cannot fit in 32 bits)
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
// return Integer Indefinite
res = 0x80000000;
BID_RETURN (res);
- } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
+ } else if ((q + exp) == 10) { // x = c(0)c(1)...c(9).c(10)...c(q-1)
// in this case 2^29.89... ~= 10^9 <= x < 10^10 ~= 2^33.2...
// so x rounded to an integer may or may not fit in a signed 32-bit int
// the cases that do not fit are identified here; the ones that fit
// fall through and will be handled with other cases further,
// under '1 <= q + exp <= 10'
- if (x_sign) { // if n < 0 and q + exp = 10
+ if (x_sign) { // if n < 0 and q + exp = 10
// if n <= -2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31+1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x500000005, 1<=q<=16
// <=> C * 10^(11-q) >= 0x500000005, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x500000005ull) {
// set invalid flag
@@ -2428,12 +2427,12 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x500000005 <=>
// C >= 0x500000005 * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31+1/2 up)
// Note: 0x500000005*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x500000005ull * __bid_ten2k64[q - 11];
+ tmp64 = 0x500000005ull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -2444,14 +2443,14 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
}
- } else { // if n > 0 and q + exp = 10
+ } else { // if n > 0 and q + exp = 10
// if n >= 2^31 - 1/2 then n is too large
// too large if c(0)c(1)...c(9).c(10)...c(q-1) >= 2^31-1/2
// <=> 0.c(0)c(1)...c(q-1) * 10^11 >= 0x4fffffffb, 1<=q<=16
// <=> C * 10^(11-q) >= 0x4fffffffb, 1<=q<=16
if (q <= 11) {
// Note: C * 10^(11-q) has 10 or 11 digits; 0x500000005 has 11 digits
- tmp64 = C1 * __bid_ten2k64[11 - q]; // C scaled up to 11-digit int
+ tmp64 = C1 * ten2k64[11 - q]; // C scaled up to 11-digit int
// c(0)c(1)...c(9)c(10) or c(0)c(1)...c(q-1)0...0 (11 digits)
if (tmp64 >= 0x4fffffffbull) {
// set invalid flag
@@ -2462,12 +2461,12 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// else cases that can be rounded to a 32-bit int fall through
// to '1 <= q + exp <= 10'
- } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
+ } else { // if (q > 11), i.e. 12 <= q <= 16 and so -15 <= exp <= -2
// C * 10^(11-q) >= 0x4fffffffb <=>
// C >= 0x4fffffffb * 10^(q-11) where 1 <= q - 11 <= 5
// (scale 2^31-1/2 up)
// Note: 0x4fffffffb*10^(q-11) has q-1 or q digits, where q <= 16
- tmp64 = 0x4fffffffbull * __bid_ten2k64[q - 11];
+ tmp64 = 0x4fffffffbull * ten2k64[q - 11];
if (C1 >= tmp64) {
// set invalid flag
*pfpsf |= INVALID_EXCEPTION;
@@ -2482,55 +2481,55 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
}
// n is not too large to be converted to int32: -2^31 - 1/2 < n < 2^31 - 1/2
// Note: some of the cases tested for above fall through to this point
- if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
+ if ((q + exp) < 0) { // n = +/-0.0...c(0)c(1)...c(q-1)
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
// return 0
res = 0x00000000;
BID_RETURN (res);
- } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
+ } else if ((q + exp) == 0) { // n = +/-0.c(0)c(1)...c(q-1)
// if 0.c(0)c(1)...c(q-1) <= 0.5 <=> c(0)c(1)...c(q-1) <= 5 * 10^(q-1)
// res = 0
// else
// res = +/-1
ind = q - 1;
- if (C1 < __bid_midpoint64[ind]) {
- res = 0x00000000; // return 0
- } else if (x_sign) { // n < 0
- res = 0xffffffff; // return -1
- } else { // n > 0
- res = 0x00000001; // return +1
+ if (C1 < midpoint64[ind]) {
+ res = 0x00000000; // return 0
+ } else if (x_sign) { // n < 0
+ res = 0xffffffff; // return -1
+ } else { // n > 0
+ res = 0x00000001; // return +1
}
// set inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
+ } else { // if (1 <= q + exp <= 10, 1 <= q <= 16, -15 <= exp <= 9)
// -2^31-1/2 <= x <= -1 or 1 <= x < 2^31-1/2 so x can be rounded
// to nearest away to a 32-bit signed integer
- if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
- ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
+ if (exp < 0) { // 2 <= q <= 16, -15 <= exp <= -1, 1 <= q + exp <= 10
+ ind = -exp; // 1 <= ind <= 15; ind is a synonym for 'x'
// chop off ind digits from the lower part of C1
// C1 = C1 + 1/2 * 10^ind where the result C1 fits in 64 bits
- C1 = C1 + __bid_midpoint64[ind - 1];
+ C1 = C1 + midpoint64[ind - 1];
// calculate C* and f*
// C* is actually floor(C*) in this case
// C* and f* need shifting and masking, as shown by
- // __bid_shiftright128[] and __bid_maskhigh128[]
+ // shiftright128[] and maskhigh128[]
// 1 <= x <= 15
- // kx = 10^(-x) = __bid_ten2mk64[ind - 1]
+ // kx = 10^(-x) = ten2mk64[ind - 1]
// C* = (C1 + 1/2 * 10^x) * 10^(-x)
// the approximation of 10^(-x) was rounded up to 54 bits
- __mul_64x64_to_128MACH (P128, C1, __bid_ten2mk64[ind - 1]);
+ __mul_64x64_to_128MACH (P128, C1, ten2mk64[ind - 1]);
Cstar = P128.w[1];
- fstar.w[1] = P128.w[1] & __bid_maskhigh128[ind - 1];
+ fstar.w[1] = P128.w[1] & maskhigh128[ind - 1];
fstar.w[0] = P128.w[0];
- // the top Ex bits of 10^(-x) are T* = __bid_ten2mk128trunc[ind].w[0], e.g.
- // if x=1, T*=__bid_ten2mk128trunc[0].w[0]=0x1999999999999999
+ // the top Ex bits of 10^(-x) are T* = ten2mk128trunc[ind].w[0], e.g.
+ // if x=1, T*=ten2mk128trunc[0].w[0]=0x1999999999999999
// C* = floor(C*)-1 (logical right shift; C* has p decimal digits,
// correct by Pr. 1)
// n = C* * 10^(e+x)
- // shift right C* by Ex-64 = __bid_shiftright128[ind]
- shift = __bid_shiftright128[ind - 1]; // 0 <= shift <= 39
+ // shift right C* by Ex-64 = shiftright128[ind]
+ shift = shiftright128[ind - 1]; // 0 <= shift <= 39
Cstar = Cstar >> shift;
// determine inexactness of the rounding of C*
// if (0 < f* - 1/2 < 10^(-x)) then
@@ -2540,31 +2539,30 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
if (ind - 1 <= 2) {
if (fstar.w[0] > 0x8000000000000000ull) {
// f* > 1/2 and the result may be exact
- tmp64 = fstar.w[0] - 0x8000000000000000ull; // f* - 1/2
- if ((tmp64 > __bid_ten2mk128trunc[ind - 1].w[1])) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ tmp64 = fstar.w[0] - 0x8000000000000000ull; // f* - 1/2
+ if ((tmp64 > ten2mk128trunc[ind - 1].w[1])) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // the result is inexact; f2* <= 1/2
+ } // else the result is exact
+ } else { // the result is inexact; f2* <= 1/2
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
}
- } else { // if 3 <= ind - 1 <= 14
- if (fstar.w[1] > __bid_one_half128[ind - 1]
- || (fstar.w[1] == __bid_one_half128[ind - 1]
- && fstar.w[0])) {
+ } else { // if 3 <= ind - 1 <= 14
+ if (fstar.w[1] > onehalf128[ind - 1] ||
+ (fstar.w[1] == onehalf128[ind - 1] && fstar.w[0])) {
// f2* > 1/2 and the result may be exact
// Calculate f2* - 1/2
- tmp64 = fstar.w[1] - __bid_one_half128[ind - 1];
- if (tmp64 || fstar.w[0] > __bid_ten2mk128trunc[ind - 1].w[1]) {
- // __bid_ten2mk128trunc[ind -1].w[1] is identical to
- // __bid_ten2mk128[ind -1].w[1]
+ tmp64 = fstar.w[1] - onehalf128[ind - 1];
+ if (tmp64 || fstar.w[0] > ten2mk128trunc[ind - 1].w[1]) {
+ // ten2mk128trunc[ind -1].w[1] is identical to
+ // ten2mk128[ind -1].w[1]
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
- } // else the result is exact
- } else { // the result is inexact; f2* <= 1/2
+ } // else the result is exact
+ } else { // the result is inexact; f2* <= 1/2
// set the inexact flag
*pfpsf |= INEXACT_EXCEPTION;
}
@@ -2582,12 +2580,12 @@ __bid64_to_int32_xrninta (UINT64 x _EXC_FLAGS_PARAM _EXC_MASKS_PARAM
res = -C1;
else
res = C1;
- } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
+ } else { // if (exp > 0) => 1 <= exp <= 9, 1 <= q < 9, 2 <= q + exp <= 10
// res = +/-C * 10^exp (exact)
if (x_sign)
- res = -C1 * __bid_ten2k64[exp];
+ res = -C1 * ten2k64[exp];
else
- res = C1 * __bid_ten2k64[exp];
+ res = C1 * ten2k64[exp];
}
}
BID_RETURN (res);