[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PULL 17/72] tcg/optimize: Compute sign mask in fold_deposit
From: |
Richard Henderson |
Subject: |
[PULL 17/72] tcg/optimize: Compute sign mask in fold_deposit |
Date: |
Tue, 24 Dec 2024 12:04:26 -0800 |
The input which overlaps the sign bit of the output can
have its input s_mask propagated to the output s_mask.
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
tcg/optimize.c | 14 ++++++++++++--
1 file changed, 12 insertions(+), 2 deletions(-)
diff --git a/tcg/optimize.c b/tcg/optimize.c
index c0f0390431..b774c96f49 100644
--- a/tcg/optimize.c
+++ b/tcg/optimize.c
@@ -1629,8 +1629,9 @@ static bool fold_deposit(OptContext *ctx, TCGOp *op)
TempOptInfo *t2 = arg_info(op->args[2]);
int ofs = op->args[3];
int len = op->args[4];
+ int width;
TCGOpcode and_opc;
- uint64_t z_mask;
+ uint64_t z_mask, s_mask;
if (ti_is_const(t1) && ti_is_const(t2)) {
return tcg_opt_gen_movi(ctx, op, op->args[0],
@@ -1641,9 +1642,11 @@ static bool fold_deposit(OptContext *ctx, TCGOp *op)
switch (ctx->type) {
case TCG_TYPE_I32:
and_opc = INDEX_op_and_i32;
+ width = 32;
break;
case TCG_TYPE_I64:
and_opc = INDEX_op_and_i64;
+ width = 64;
break;
default:
g_assert_not_reached();
@@ -1668,8 +1671,15 @@ static bool fold_deposit(OptContext *ctx, TCGOp *op)
return fold_and(ctx, op);
}
+ /* The s_mask from the top portion of the deposit is still valid. */
+ if (ofs + len == width) {
+ s_mask = t2->s_mask << ofs;
+ } else {
+ s_mask = t1->s_mask & ~MAKE_64BIT_MASK(0, ofs + len);
+ }
+
z_mask = deposit64(t1->z_mask, ofs, len, t2->z_mask);
- return fold_masks_z(ctx, op, z_mask);
+ return fold_masks_zs(ctx, op, z_mask, s_mask);
}
static bool fold_divide(OptContext *ctx, TCGOp *op)
--
2.43.0
- [PULL 12/72] tcg/optimize: Use fold_masks_zs in fold_andc, (continued)
- [PULL 12/72] tcg/optimize: Use fold_masks_zs in fold_andc, Richard Henderson, 2024/12/24
- [PULL 14/72] tcg/optimize: Use fold_masks_zs in fold_count_zeros, Richard Henderson, 2024/12/24
- [PULL 16/72] tcg/optimize: Use fold_and and fold_masks_z in fold_deposit, Richard Henderson, 2024/12/24
- [PULL 19/72] tcg/optimize: Use finish_folding in fold_dup, fold_dup2, Richard Henderson, 2024/12/24
- [PULL 25/72] tcg/optimize: Use fold_masks_zs in fold_movcond, Richard Henderson, 2024/12/24
- [PULL 03/72] tcg/optimize: Split out finish_bb, finish_ebb, Richard Henderson, 2024/12/24
- [PULL 08/72] tcg/optimize: Change representation of s_mask, Richard Henderson, 2024/12/24
- [PULL 05/72] tcg/optimize: Copy mask writeback to fold_masks, Richard Henderson, 2024/12/24
- [PULL 13/72] tcg/optimize: Use fold_masks_zs in fold_bswap, Richard Henderson, 2024/12/24
- [PULL 15/72] tcg/optimize: Use fold_masks_z in fold_ctpop, Richard Henderson, 2024/12/24
- [PULL 17/72] tcg/optimize: Compute sign mask in fold_deposit,
Richard Henderson <=
- [PULL 18/72] tcg/optimize: Use finish_folding in fold_divide, Richard Henderson, 2024/12/24
- [PULL 21/72] tcg/optimize: Use fold_masks_z in fold_extract, Richard Henderson, 2024/12/24
- [PULL 24/72] tcg/optimize: Use fold_masks_z in fold_extu, Richard Henderson, 2024/12/24
- [PULL 26/72] tcg/optimize: Use finish_folding in fold_mul*, Richard Henderson, 2024/12/24
- [PULL 29/72] tcg/optimize: Use fold_masks_s in fold_nor, Richard Henderson, 2024/12/24
- [PULL 30/72] tcg/optimize: Use fold_masks_s in fold_not, Richard Henderson, 2024/12/24
- [PULL 32/72] tcg/optimize: Use fold_masks_zs in fold_orc, Richard Henderson, 2024/12/24
- [PULL 31/72] tcg/optimize: Use fold_masks_zs in fold_or, Richard Henderson, 2024/12/24
- [PULL 34/72] tcg/optimize: Return true from fold_qemu_st, fold_tcg_st, Richard Henderson, 2024/12/24
- [PULL 35/72] tcg/optimize: Use finish_folding in fold_remainder, Richard Henderson, 2024/12/24