[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PATCH v2 21/48] tcg/optimize: Split out fold_addsub2_i32
From: |
Richard Henderson |
Subject: |
[PATCH v2 21/48] tcg/optimize: Split out fold_addsub2_i32 |
Date: |
Thu, 7 Oct 2021 12:54:29 -0700 |
Add two additional helpers, fold_add2_i32 and fold_sub2_i32
which will not be simple wrappers forever.
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
tcg/optimize.c | 70 +++++++++++++++++++++++++++++++-------------------
1 file changed, 44 insertions(+), 26 deletions(-)
diff --git a/tcg/optimize.c b/tcg/optimize.c
index 01a0925c89..1cec06aa81 100644
--- a/tcg/optimize.c
+++ b/tcg/optimize.c
@@ -704,6 +704,39 @@ static bool fold_add(OptContext *ctx, TCGOp *op)
return fold_const2(ctx, op);
}
+static bool fold_addsub2_i32(OptContext *ctx, TCGOp *op, bool add)
+{
+ if (arg_is_const(op->args[2]) && arg_is_const(op->args[3]) &&
+ arg_is_const(op->args[4]) && arg_is_const(op->args[5])) {
+ uint32_t al = arg_info(op->args[2])->val;
+ uint32_t ah = arg_info(op->args[3])->val;
+ uint32_t bl = arg_info(op->args[4])->val;
+ uint32_t bh = arg_info(op->args[5])->val;
+ uint64_t a = ((uint64_t)ah << 32) | al;
+ uint64_t b = ((uint64_t)bh << 32) | bl;
+ TCGArg rl, rh;
+ TCGOp *op2 = tcg_op_insert_before(ctx->tcg, op, INDEX_op_mov_i32);
+
+ if (add) {
+ a += b;
+ } else {
+ a -= b;
+ }
+
+ rl = op->args[0];
+ rh = op->args[1];
+ tcg_opt_gen_movi(ctx, op, rl, (int32_t)a);
+ tcg_opt_gen_movi(ctx, op2, rh, (int32_t)(a >> 32));
+ return true;
+ }
+ return false;
+}
+
+static bool fold_add2_i32(OptContext *ctx, TCGOp *op)
+{
+ return fold_addsub2_i32(ctx, op, true);
+}
+
static bool fold_and(OptContext *ctx, TCGOp *op)
{
return fold_const2(ctx, op);
@@ -1043,6 +1076,11 @@ static bool fold_sub(OptContext *ctx, TCGOp *op)
return fold_const2(ctx, op);
}
+static bool fold_sub2_i32(OptContext *ctx, TCGOp *op)
+{
+ return fold_addsub2_i32(ctx, op, false);
+}
+
static bool fold_xor(OptContext *ctx, TCGOp *op)
{
return fold_const2(ctx, op);
@@ -1688,32 +1726,6 @@ void tcg_optimize(TCGContext *s)
}
break;
- case INDEX_op_add2_i32:
- case INDEX_op_sub2_i32:
- if (arg_is_const(op->args[2]) && arg_is_const(op->args[3])
- && arg_is_const(op->args[4]) && arg_is_const(op->args[5])) {
- uint32_t al = arg_info(op->args[2])->val;
- uint32_t ah = arg_info(op->args[3])->val;
- uint32_t bl = arg_info(op->args[4])->val;
- uint32_t bh = arg_info(op->args[5])->val;
- uint64_t a = ((uint64_t)ah << 32) | al;
- uint64_t b = ((uint64_t)bh << 32) | bl;
- TCGArg rl, rh;
- TCGOp *op2 = tcg_op_insert_before(s, op, INDEX_op_mov_i32);
-
- if (opc == INDEX_op_add2_i32) {
- a += b;
- } else {
- a -= b;
- }
-
- rl = op->args[0];
- rh = op->args[1];
- tcg_opt_gen_movi(&ctx, op, rl, (int32_t)a);
- tcg_opt_gen_movi(&ctx, op2, rh, (int32_t)(a >> 32));
- continue;
- }
- break;
default:
break;
@@ -1724,6 +1736,9 @@ void tcg_optimize(TCGContext *s)
CASE_OP_32_64_VEC(add):
done = fold_add(&ctx, op);
break;
+ case INDEX_op_add2_i32:
+ done = fold_add2_i32(&ctx, op);
+ break;
CASE_OP_32_64_VEC(and):
done = fold_and(&ctx, op);
break;
@@ -1818,6 +1833,9 @@ void tcg_optimize(TCGContext *s)
CASE_OP_32_64_VEC(sub):
done = fold_sub(&ctx, op);
break;
+ case INDEX_op_sub2_i32:
+ done = fold_sub2_i32(&ctx, op);
+ break;
CASE_OP_32_64_VEC(xor):
done = fold_xor(&ctx, op);
break;
--
2.25.1
- [PATCH v2 22/48] tcg/optimize: Split out fold_movcond, (continued)
- [PATCH v2 22/48] tcg/optimize: Split out fold_movcond, Richard Henderson, 2021/10/07
- [PATCH v2 25/48] tcg/optimize: Split out fold_deposit, Richard Henderson, 2021/10/07
- [PATCH v2 12/48] tcg/optimize: Split out finish_folding, Richard Henderson, 2021/10/07
- [PATCH v2 33/48] tcg/optimize: Add type to OptContext, Richard Henderson, 2021/10/07
- [PATCH v2 39/48] tcg/optimize: Expand fold_mulu2_i32 to all 4-arg multiplies, Richard Henderson, 2021/10/07
- [PATCH v2 11/48] tcg/optimize: Return true from tcg_opt_gen_{mov, movi}, Richard Henderson, 2021/10/07
- [PATCH v2 19/48] tcg/optimize: Split out fold_setcond, Richard Henderson, 2021/10/07
- [PATCH v2 21/48] tcg/optimize: Split out fold_addsub2_i32,
Richard Henderson <=
- [PATCH v2 23/48] tcg/optimize: Split out fold_extract2, Richard Henderson, 2021/10/07
- [PATCH v2 29/48] tcg/optimize: Split out fold_mov, Richard Henderson, 2021/10/07
- [PATCH v2 20/48] tcg/optimize: Split out fold_mulu2_i32, Richard Henderson, 2021/10/07
- [PATCH v2 36/48] tcg/optimize: Split out fold_xi_to_x, Richard Henderson, 2021/10/07
- [PATCH v2 37/48] tcg/optimize: Split out fold_ix_to_i, Richard Henderson, 2021/10/07
- [PATCH v2 38/48] tcg/optimize: Split out fold_masks, Richard Henderson, 2021/10/07
- [PATCH v2 31/48] tcg/optimize: Split out fold_xx_to_x, Richard Henderson, 2021/10/07
- [PATCH v2 13/48] tcg/optimize: Use a boolean to avoid a mass of continues, Richard Henderson, 2021/10/07
- [PATCH v2 42/48] tcg/optimize: Add more simplifications for orc, Richard Henderson, 2021/10/07