[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PATCH v2 07/51] tcg/optimize: Use fold_masks_zs in fold_and
From: |
Richard Henderson |
Subject: |
[PATCH v2 07/51] tcg/optimize: Use fold_masks_zs in fold_and |
Date: |
Thu, 19 Dec 2024 20:10:19 -0800 |
Avoid the use of the OptContext slots. Find TempOptInfo once.
Sink mask computation below fold_affected_mask early exit.
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
tcg/optimize.c | 30 ++++++++++++++++--------------
1 file changed, 16 insertions(+), 14 deletions(-)
diff --git a/tcg/optimize.c b/tcg/optimize.c
index 41fd69365b..52c785472f 100644
--- a/tcg/optimize.c
+++ b/tcg/optimize.c
@@ -1313,7 +1313,8 @@ static bool fold_add2(OptContext *ctx, TCGOp *op)
static bool fold_and(OptContext *ctx, TCGOp *op)
{
- uint64_t z1, z2;
+ uint64_t z1, z2, z_mask, s_mask;
+ TempOptInfo *t1, *t2;
if (fold_const2_commutative(ctx, op) ||
fold_xi_to_i(ctx, op, 0) ||
@@ -1322,27 +1323,28 @@ static bool fold_and(OptContext *ctx, TCGOp *op)
return true;
}
- z1 = arg_info(op->args[1])->z_mask;
- z2 = arg_info(op->args[2])->z_mask;
- ctx->z_mask = z1 & z2;
-
- /*
- * Sign repetitions are perforce all identical, whether they are 1 or 0.
- * Bitwise operations preserve the relative quantity of the repetitions.
- */
- ctx->s_mask = arg_info(op->args[1])->s_mask
- & arg_info(op->args[2])->s_mask;
+ t1 = arg_info(op->args[1]);
+ t2 = arg_info(op->args[2]);
+ z1 = t1->z_mask;
+ z2 = t2->z_mask;
/*
* Known-zeros does not imply known-ones. Therefore unless
* arg2 is constant, we can't infer affected bits from it.
*/
- if (arg_is_const(op->args[2]) &&
- fold_affected_mask(ctx, op, z1 & ~z2)) {
+ if (t2->is_const && fold_affected_mask(ctx, op, z1 & ~z2)) {
return true;
}
- return fold_masks(ctx, op);
+ z_mask = z1 & z2;
+
+ /*
+ * Sign repetitions are perforce all identical, whether they are 1 or 0.
+ * Bitwise operations preserve the relative quantity of the repetitions.
+ */
+ s_mask = t1->s_mask & t2->s_mask;
+
+ return fold_masks_zs(ctx, op, z_mask, s_mask);
}
static bool fold_andc(OptContext *ctx, TCGOp *op)
--
2.43.0
- [PATCH v2 00/51] tcg: Remove in-flight mask data from OptContext, Richard Henderson, 2024/12/19
- [PATCH v2 04/51] tcg/optimize: Split out fold_masks_zs, Richard Henderson, 2024/12/19
- [PATCH v2 03/51] tcg/optimize: Copy mask writeback to fold_masks, Richard Henderson, 2024/12/19
- [PATCH v2 02/51] tcg/optimize: Split out fold_affected_mask, Richard Henderson, 2024/12/19
- [PATCH v2 01/51] tcg/optimize: Split out finish_bb, finish_ebb, Richard Henderson, 2024/12/19
- [PATCH v2 05/51] tcg/optimize: Augment s_mask from z_mask in fold_masks_zs, Richard Henderson, 2024/12/19
- [PATCH v2 06/51] tcg/optimize: Use finish_folding in fold_add, fold_add_vec, fold_addsub2, Richard Henderson, 2024/12/19
- [PATCH v2 07/51] tcg/optimize: Use fold_masks_zs in fold_and,
Richard Henderson <=
- [PATCH v2 11/51] tcg/optimize: Use fold_masks_z in fold_ctpop, Richard Henderson, 2024/12/19
- [PATCH v2 08/51] tcg/optimize: Use fold_masks_zs in fold_andc, Richard Henderson, 2024/12/19
- [PATCH v2 12/51] tcg/optimize: Use fold_and and fold_masks_z in fold_deposit, Richard Henderson, 2024/12/19
- [PATCH v2 09/51] tcg/optimize: Use fold_masks_zs in fold_bswap, Richard Henderson, 2024/12/19
- [PATCH v2 14/51] tcg/optimize: Use finish_folding in fold_divide, Richard Henderson, 2024/12/19
- [PATCH v2 17/51] tcg/optimize: Use fold_masks_z in fold_extract, Richard Henderson, 2024/12/19
- [PATCH v2 20/51] tcg/optimize: Use fold_masks_z in fold_extu, Richard Henderson, 2024/12/19
- [PATCH v2 13/51] tcg/optimize: Compute sign mask in fold_deposit, Richard Henderson, 2024/12/19
- [PATCH v2 16/51] tcg/optimize: Use fold_masks_s in fold_eqv, Richard Henderson, 2024/12/19