tcg/optimize: Propagate sign info for shifting
For constant shifts, we can simply shift the s_mask. For variable shifts, we know that sar does not reduce the s_mask, which helps for sequences like ext32s_i64 t, in sar_i64 t, t, v ext32s_i64 out, t allowing the final extend to be eliminated. Reviewed-by: Alex Bennée <alex.bennee@linaro.org> Reviewed-by: Luis Pires <luis.pires@eldorado.org.br> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
This commit is contained in:
parent
2b9d0c59ed
commit
93a967fbb5
@ -85,6 +85,18 @@ static uint64_t smask_from_zmask(uint64_t zmask)
|
|||||||
return ~(~0ull >> rep);
|
return ~(~0ull >> rep);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Recreate a properly left-aligned smask after manipulation.
|
||||||
|
* Some bit-shuffling, particularly shifts and rotates, may
|
||||||
|
* retain sign bits on the left, but may scatter disconnected
|
||||||
|
* sign bits on the right. Retain only what remains to the left.
|
||||||
|
*/
|
||||||
|
static uint64_t smask_from_smask(int64_t smask)
|
||||||
|
{
|
||||||
|
/* Only the 1 bits are significant for smask */
|
||||||
|
return smask_from_zmask(~smask);
|
||||||
|
}
|
||||||
|
|
||||||
static inline TempOptInfo *ts_info(TCGTemp *ts)
|
static inline TempOptInfo *ts_info(TCGTemp *ts)
|
||||||
{
|
{
|
||||||
return ts->state_ptr;
|
return ts->state_ptr;
|
||||||
@ -1843,18 +1855,50 @@ static bool fold_sextract(OptContext *ctx, TCGOp *op)
|
|||||||
|
|
||||||
static bool fold_shift(OptContext *ctx, TCGOp *op)
|
static bool fold_shift(OptContext *ctx, TCGOp *op)
|
||||||
{
|
{
|
||||||
|
uint64_t s_mask, z_mask, sign;
|
||||||
|
|
||||||
if (fold_const2(ctx, op) ||
|
if (fold_const2(ctx, op) ||
|
||||||
fold_ix_to_i(ctx, op, 0) ||
|
fold_ix_to_i(ctx, op, 0) ||
|
||||||
fold_xi_to_x(ctx, op, 0)) {
|
fold_xi_to_x(ctx, op, 0)) {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
s_mask = arg_info(op->args[1])->s_mask;
|
||||||
|
z_mask = arg_info(op->args[1])->z_mask;
|
||||||
|
|
||||||
if (arg_is_const(op->args[2])) {
|
if (arg_is_const(op->args[2])) {
|
||||||
ctx->z_mask = do_constant_folding(op->opc, ctx->type,
|
int sh = arg_info(op->args[2])->val;
|
||||||
arg_info(op->args[1])->z_mask,
|
|
||||||
arg_info(op->args[2])->val);
|
ctx->z_mask = do_constant_folding(op->opc, ctx->type, z_mask, sh);
|
||||||
|
|
||||||
|
s_mask = do_constant_folding(op->opc, ctx->type, s_mask, sh);
|
||||||
|
ctx->s_mask = smask_from_smask(s_mask);
|
||||||
|
|
||||||
return fold_masks(ctx, op);
|
return fold_masks(ctx, op);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
switch (op->opc) {
|
||||||
|
CASE_OP_32_64(sar):
|
||||||
|
/*
|
||||||
|
* Arithmetic right shift will not reduce the number of
|
||||||
|
* input sign repetitions.
|
||||||
|
*/
|
||||||
|
ctx->s_mask = s_mask;
|
||||||
|
break;
|
||||||
|
CASE_OP_32_64(shr):
|
||||||
|
/*
|
||||||
|
* If the sign bit is known zero, then logical right shift
|
||||||
|
* will not reduced the number of input sign repetitions.
|
||||||
|
*/
|
||||||
|
sign = (s_mask & -s_mask) >> 1;
|
||||||
|
if (!(z_mask & sign)) {
|
||||||
|
ctx->s_mask = s_mask;
|
||||||
|
}
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user