Set nonzero bits from bitwise and operator in range-ops.
Now that nonzero bits are first class citizens in the range, we can keep better track of them in range-ops, especially the bitwise and operator. This patch sets the nonzero mask for the trivial case. In doing so, I've removed some old dead code that was an attempt to keep better track of masks. I'm sure there are tons of optimizations throughout range-ops that could be implemented, especially the op1_range methods, but those always make my head hurt. I'll leave them to the smarter hackers out there. I've removed the restriction that nonzero bits can't be queried from legacy. This was causing special casing all over the place, and it's not like we can generate incorrect code. We just silently drop nonzero bits to -1 in some of the legacy code. The end result is that VRP1, and other users of legacy, may not benefit from these improvements. Tested and benchmarked on x86-64 Linux. gcc/ChangeLog: * range-op.cc (unsigned_singleton_p): Remove. (operator_bitwise_and::remove_impossible_ranges): Remove. (operator_bitwise_and::fold_range): Set nonzero bits. * * value-range.cc (irange::get_nonzero_bits): Remove legacy_mode_p assert. (irange::dump_bitmasks): Remove legacy_mode_p check.
This commit is contained in:
parent
32a753506b
commit
cab411a2b4
@ -2604,72 +2604,8 @@ private:
|
||||
void simple_op1_range_solver (irange &r, tree type,
|
||||
const irange &lhs,
|
||||
const irange &op2) const;
|
||||
void remove_impossible_ranges (irange &r, const irange &rh) const;
|
||||
} op_bitwise_and;
|
||||
|
||||
static bool
|
||||
unsigned_singleton_p (const irange &op)
|
||||
{
|
||||
tree mask;
|
||||
if (op.singleton_p (&mask))
|
||||
{
|
||||
wide_int x = wi::to_wide (mask);
|
||||
return wi::ge_p (x, 0, TYPE_SIGN (op.type ()));
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
// Remove any ranges from R that are known to be impossible when an
|
||||
// range is ANDed with MASK.
|
||||
|
||||
void
|
||||
operator_bitwise_and::remove_impossible_ranges (irange &r,
|
||||
const irange &rmask) const
|
||||
{
|
||||
if (r.undefined_p () || !unsigned_singleton_p (rmask))
|
||||
return;
|
||||
|
||||
wide_int mask = rmask.lower_bound ();
|
||||
tree type = r.type ();
|
||||
int prec = TYPE_PRECISION (type);
|
||||
int leading_zeros = wi::clz (mask);
|
||||
int_range_max impossible_ranges;
|
||||
|
||||
/* We know that starting at the most significant bit, any 0 in the
|
||||
mask means the resulting range cannot contain a 1 in that same
|
||||
position. This means the following ranges are impossible:
|
||||
|
||||
x & 0b1001 1010
|
||||
IMPOSSIBLE RANGES
|
||||
01xx xxxx [0100 0000, 0111 1111]
|
||||
001x xxxx [0010 0000, 0011 1111]
|
||||
0000 01xx [0000 0100, 0000 0111]
|
||||
0000 0001 [0000 0001, 0000 0001]
|
||||
*/
|
||||
wide_int one = wi::one (prec);
|
||||
for (int i = 0; i < prec - leading_zeros - 1; ++i)
|
||||
if (wi::bit_and (mask, wi::lshift (one, wi::uhwi (i, prec))) == 0)
|
||||
{
|
||||
tree lb = fold_build2 (LSHIFT_EXPR, type,
|
||||
build_one_cst (type),
|
||||
build_int_cst (type, i));
|
||||
tree ub_left = fold_build1 (BIT_NOT_EXPR, type,
|
||||
fold_build2 (LSHIFT_EXPR, type,
|
||||
build_minus_one_cst (type),
|
||||
build_int_cst (type, i)));
|
||||
tree ub_right = fold_build2 (LSHIFT_EXPR, type,
|
||||
build_one_cst (type),
|
||||
build_int_cst (type, i));
|
||||
tree ub = fold_build2 (BIT_IOR_EXPR, type, ub_left, ub_right);
|
||||
impossible_ranges.union_ (int_range<1> (lb, ub));
|
||||
}
|
||||
if (!impossible_ranges.undefined_p ())
|
||||
{
|
||||
impossible_ranges.invert ();
|
||||
r.intersect (impossible_ranges);
|
||||
}
|
||||
}
|
||||
|
||||
bool
|
||||
operator_bitwise_and::fold_range (irange &r, tree type,
|
||||
const irange &lh,
|
||||
@ -2678,9 +2614,9 @@ operator_bitwise_and::fold_range (irange &r, tree type,
|
||||
{
|
||||
if (range_operator::fold_range (r, type, lh, rh))
|
||||
{
|
||||
// FIXME: This is temporarily disabled because, though it
|
||||
// generates better ranges, it's noticeably slower for evrp.
|
||||
// remove_impossible_ranges (r, rh);
|
||||
if (!lh.undefined_p () && !rh.undefined_p ())
|
||||
r.set_nonzero_bits (wi::bit_and (lh.get_nonzero_bits (),
|
||||
rh.get_nonzero_bits ()));
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
|
@ -2388,10 +2388,6 @@ wide_int
|
||||
irange::get_nonzero_bits () const
|
||||
{
|
||||
gcc_checking_assert (!undefined_p ());
|
||||
// Nonzero bits are unsupported in legacy mode. The mask may be set
|
||||
// as a consequence of propagation or reading global ranges, but no
|
||||
// one from legacy land should be querying this.
|
||||
gcc_checking_assert (!legacy_mode_p ());
|
||||
|
||||
// Calculate the nonzero bits inherent in the range.
|
||||
wide_int min = lower_bound ();
|
||||
@ -2509,7 +2505,7 @@ irange::dump (FILE *file) const
|
||||
void
|
||||
irange::dump_bitmasks (FILE *file) const
|
||||
{
|
||||
if (m_nonzero_mask && !legacy_mode_p ())
|
||||
if (m_nonzero_mask)
|
||||
{
|
||||
wide_int nz = get_nonzero_bits ();
|
||||
if (nz != -1)
|
||||
|
Loading…
Reference in New Issue
Block a user