* [PATCH] tcg/optimize: lower some ANDs to two shifts
@ 2024-02-28 11:06 Paolo Bonzini
2024-02-28 12:21 ` Paolo Bonzini
0 siblings, 1 reply; 2+ messages in thread
From: Paolo Bonzini @ 2024-02-28 11:06 UTC (permalink / raw)
To: qemu-devel
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
---
tcg/optimize.c | 60 +++++++++++++++++++++++++++++++++++++++++++-------
1 file changed, 52 insertions(+), 8 deletions(-)
diff --git a/tcg/optimize.c b/tcg/optimize.c
index 3995bc047db..8ea1f287788 100644
--- a/tcg/optimize.c
+++ b/tcg/optimize.c
@@ -1281,6 +1281,43 @@ static bool fold_add2(OptContext *ctx, TCGOp *op)
return fold_addsub2(ctx, op, true);
}
+static bool fold_and_to_shifts(OptContext *ctx, uint64_t c, TCGOp *op)
+{
+ TCGOpcode shl_opc = tcg->type == TCG_TYPE_I32 ? INDEX_op_shl_i32 : INDEX_op_shl_i64;
+ TCGOpcode shr_opc = tcg->type == TCG_TYPE_I32 ? INDEX_op_shr_i32 : INDEX_op_shr_i64;
+
+ TCGOpcode first, second;
+ int count;
+ TCGOp *op2;
+
+ unsigned_c = tcg->type == TCG_TYPE_I32 ? (uint32_t) c : c;
+ if (is_power_of_2(-c) &&
+ !tcg_op_imm_match(op->opc, c)) {
+ /* AND with 11...11000, shift right then left. */
+ count = ctz64(c);
+ first = shr_opc;
+ } else if (is_power_of_2(c + 1) &&
+ !tcg_op_imm_match(INDEX_op_and_i64, c)) {
+ /* AND with 00...00111, shift left then right. */
+ int bits = tcg->type == TCG_TYPE_I32 ? 32 : 64;
+ count = bits - cto64(c);
+ first = shl_opc;
+ } else {
+ return false;
+ }
+
+
+ op->opc = first;
+ op->args[2] = arg_new_constant(ctx, count);
+
+ second = shl_opc ^ shr_opc ^ first;
+ op2 = tcg_op_insert_after(ctx->tcg, op, second, 3);
+ op2->args[0] = op->args[0];
+ op2->args[1] = op->args[0];
+ op2->args[2] = arg_new_constant(ctx, count);
+ return true;
+}
+
static bool fold_and(OptContext *ctx, TCGOp *op)
{
uint64_t z1, z2;
@@ -1294,6 +1331,18 @@ static bool fold_and(OptContext *ctx, TCGOp *op)
z1 = arg_info(op->args[1])->z_mask;
z2 = arg_info(op->args[2])->z_mask;
+
+ /*
+ * Known-zeros does not imply known-ones. Therefore unless
+ * arg2 is constant, we can't infer affected bits from it.
+ */
+ if (arg_is_const(op->args[2])) {
+ if (fold_and_to_shifts(ctx, z2, op)) {
+ return true;
+ }
+ ctx->a_mask = z1 & ~z2;
+ }
+
ctx->z_mask = z1 & z2;
/*
@@ -1303,14 +1352,6 @@ static bool fold_and(OptContext *ctx, TCGOp *op)
ctx->s_mask = arg_info(op->args[1])->s_mask
& arg_info(op->args[2])->s_mask;
- /*
- * Known-zeros does not imply known-ones. Therefore unless
- * arg2 is constant, we can't infer affected bits from it.
- */
- if (arg_is_const(op->args[2])) {
- ctx->a_mask = z1 & ~z2;
- }
-
return fold_masks(ctx, op);
}
@@ -1333,6 +1374,9 @@ static bool fold_andc(OptContext *ctx, TCGOp *op)
*/
if (arg_is_const(op->args[2])) {
uint64_t z2 = ~arg_info(op->args[2])->z_mask;
+ if (fold_and_to_shifts(ctx, z2, op)) {
+ return true;
+ }
ctx->a_mask = z1 & ~z2;
z1 &= z2;
}
--
2.43.2
^ permalink raw reply related [flat|nested] 2+ messages in thread* Re: [PATCH] tcg/optimize: lower some ANDs to two shifts
2024-02-28 11:06 [PATCH] tcg/optimize: lower some ANDs to two shifts Paolo Bonzini
@ 2024-02-28 12:21 ` Paolo Bonzini
0 siblings, 0 replies; 2+ messages in thread
From: Paolo Bonzini @ 2024-02-28 12:21 UTC (permalink / raw)
To: qemu-devel
Sorry, that was sent incorrectly.
Paolo
On Wed, Feb 28, 2024 at 12:06 PM Paolo Bonzini <pbonzini@redhat.com> wrote:
>
> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
> ---
> tcg/optimize.c | 60 +++++++++++++++++++++++++++++++++++++++++++-------
> 1 file changed, 52 insertions(+), 8 deletions(-)
>
> diff --git a/tcg/optimize.c b/tcg/optimize.c
> index 3995bc047db..8ea1f287788 100644
> --- a/tcg/optimize.c
> +++ b/tcg/optimize.c
> @@ -1281,6 +1281,43 @@ static bool fold_add2(OptContext *ctx, TCGOp *op)
> return fold_addsub2(ctx, op, true);
> }
>
> +static bool fold_and_to_shifts(OptContext *ctx, uint64_t c, TCGOp *op)
> +{
> + TCGOpcode shl_opc = tcg->type == TCG_TYPE_I32 ? INDEX_op_shl_i32 : INDEX_op_shl_i64;
> + TCGOpcode shr_opc = tcg->type == TCG_TYPE_I32 ? INDEX_op_shr_i32 : INDEX_op_shr_i64;
> +
> + TCGOpcode first, second;
> + int count;
> + TCGOp *op2;
> +
> + unsigned_c = tcg->type == TCG_TYPE_I32 ? (uint32_t) c : c;
> + if (is_power_of_2(-c) &&
> + !tcg_op_imm_match(op->opc, c)) {
> + /* AND with 11...11000, shift right then left. */
> + count = ctz64(c);
> + first = shr_opc;
> + } else if (is_power_of_2(c + 1) &&
> + !tcg_op_imm_match(INDEX_op_and_i64, c)) {
> + /* AND with 00...00111, shift left then right. */
> + int bits = tcg->type == TCG_TYPE_I32 ? 32 : 64;
> + count = bits - cto64(c);
> + first = shl_opc;
> + } else {
> + return false;
> + }
> +
> +
> + op->opc = first;
> + op->args[2] = arg_new_constant(ctx, count);
> +
> + second = shl_opc ^ shr_opc ^ first;
> + op2 = tcg_op_insert_after(ctx->tcg, op, second, 3);
> + op2->args[0] = op->args[0];
> + op2->args[1] = op->args[0];
> + op2->args[2] = arg_new_constant(ctx, count);
> + return true;
> +}
> +
> static bool fold_and(OptContext *ctx, TCGOp *op)
> {
> uint64_t z1, z2;
> @@ -1294,6 +1331,18 @@ static bool fold_and(OptContext *ctx, TCGOp *op)
>
> z1 = arg_info(op->args[1])->z_mask;
> z2 = arg_info(op->args[2])->z_mask;
> +
> + /*
> + * Known-zeros does not imply known-ones. Therefore unless
> + * arg2 is constant, we can't infer affected bits from it.
> + */
> + if (arg_is_const(op->args[2])) {
> + if (fold_and_to_shifts(ctx, z2, op)) {
> + return true;
> + }
> + ctx->a_mask = z1 & ~z2;
> + }
> +
> ctx->z_mask = z1 & z2;
>
> /*
> @@ -1303,14 +1352,6 @@ static bool fold_and(OptContext *ctx, TCGOp *op)
> ctx->s_mask = arg_info(op->args[1])->s_mask
> & arg_info(op->args[2])->s_mask;
>
> - /*
> - * Known-zeros does not imply known-ones. Therefore unless
> - * arg2 is constant, we can't infer affected bits from it.
> - */
> - if (arg_is_const(op->args[2])) {
> - ctx->a_mask = z1 & ~z2;
> - }
> -
> return fold_masks(ctx, op);
> }
>
> @@ -1333,6 +1374,9 @@ static bool fold_andc(OptContext *ctx, TCGOp *op)
> */
> if (arg_is_const(op->args[2])) {
> uint64_t z2 = ~arg_info(op->args[2])->z_mask;
> + if (fold_and_to_shifts(ctx, z2, op)) {
> + return true;
> + }
> ctx->a_mask = z1 & ~z2;
> z1 &= z2;
> }
> --
> 2.43.2
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2024-02-28 12:21 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-02-28 11:06 [PATCH] tcg/optimize: lower some ANDs to two shifts Paolo Bonzini
2024-02-28 12:21 ` Paolo Bonzini
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).