From: christoffer.dall@linaro.org (Christoffer Dall)
To: linux-arm-kernel@lists.infradead.org
Subject: [PATCH v4 07/19] arm64: insn: Add encoder for bitwise operations using literals
Date: Mon, 15 Jan 2018 12:26:19 +0100 [thread overview]
Message-ID: <20180115112619.GH21403@cbox> (raw)
In-Reply-To: <20180104184334.16571-8-marc.zyngier@arm.com>
On Thu, Jan 04, 2018 at 06:43:22PM +0000, Marc Zyngier wrote:
> We lack a way to encode operations such as AND, ORR, EOR that take
> an immediate value. Doing so is quite involved, and is all about
> reverse engineering the decoding algorithm described in the
> pseudocode function DecodeBitMasks().
Black magic.
>
> This has been tested by feeding it all the possible literal values
> and comparing the output with that of GAS.
That's comforting.
I didn't attempt at verifying the functionality or every hard-coded
value or dirty bit trick in this patch, but I did glance over the parts
I could vaguely understand and didn't see any issues.
I suppose that's a weak sort of:
Acked-by: Christoffer Dall <christoffer.dall@linaro.org>
>
> Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
> ---
> arch/arm64/include/asm/insn.h | 9 +++
> arch/arm64/kernel/insn.c | 136 ++++++++++++++++++++++++++++++++++++++++++
> 2 files changed, 145 insertions(+)
>
> diff --git a/arch/arm64/include/asm/insn.h b/arch/arm64/include/asm/insn.h
> index 21fffdd290a3..815b35bc53ed 100644
> --- a/arch/arm64/include/asm/insn.h
> +++ b/arch/arm64/include/asm/insn.h
> @@ -315,6 +315,10 @@ __AARCH64_INSN_FUNCS(eor, 0x7F200000, 0x4A000000)
> __AARCH64_INSN_FUNCS(eon, 0x7F200000, 0x4A200000)
> __AARCH64_INSN_FUNCS(ands, 0x7F200000, 0x6A000000)
> __AARCH64_INSN_FUNCS(bics, 0x7F200000, 0x6A200000)
> +__AARCH64_INSN_FUNCS(and_imm, 0x7F800000, 0x12000000)
> +__AARCH64_INSN_FUNCS(orr_imm, 0x7F800000, 0x32000000)
> +__AARCH64_INSN_FUNCS(eor_imm, 0x7F800000, 0x52000000)
> +__AARCH64_INSN_FUNCS(ands_imm, 0x7F800000, 0x72000000)
> __AARCH64_INSN_FUNCS(b, 0xFC000000, 0x14000000)
> __AARCH64_INSN_FUNCS(bl, 0xFC000000, 0x94000000)
> __AARCH64_INSN_FUNCS(cbz, 0x7F000000, 0x34000000)
> @@ -424,6 +428,11 @@ u32 aarch64_insn_gen_logical_shifted_reg(enum aarch64_insn_register dst,
> int shift,
> enum aarch64_insn_variant variant,
> enum aarch64_insn_logic_type type);
> +u32 aarch64_insn_gen_logical_immediate(enum aarch64_insn_logic_type type,
> + enum aarch64_insn_variant variant,
> + enum aarch64_insn_register Rn,
> + enum aarch64_insn_register Rd,
> + u64 imm);
> u32 aarch64_insn_gen_prefetch(enum aarch64_insn_register base,
> enum aarch64_insn_prfm_type type,
> enum aarch64_insn_prfm_target target,
> diff --git a/arch/arm64/kernel/insn.c b/arch/arm64/kernel/insn.c
> index 7e432662d454..72cb1721c63f 100644
> --- a/arch/arm64/kernel/insn.c
> +++ b/arch/arm64/kernel/insn.c
> @@ -1485,3 +1485,139 @@ pstate_check_t * const aarch32_opcode_cond_checks[16] = {
> __check_hi, __check_ls, __check_ge, __check_lt,
> __check_gt, __check_le, __check_al, __check_al
> };
> +
> +static bool range_of_ones(u64 val)
> +{
> + /* Doesn't handle full ones or full zeroes */
> + u64 sval = val >> __ffs64(val);
> +
> + /* One of Sean Eron Anderson's bithack tricks */
> + return ((sval + 1) & (sval)) == 0;
> +}
> +
> +static u32 aarch64_encode_immediate(u64 imm,
> + enum aarch64_insn_variant variant,
> + u32 insn)
> +{
> + unsigned int immr, imms, n, ones, ror, esz, tmp;
> + u64 mask = ~0UL;
> +
> + /* Can't encode full zeroes or full ones */
> + if (!imm || !~imm)
> + return AARCH64_BREAK_FAULT;
> +
> + switch (variant) {
> + case AARCH64_INSN_VARIANT_32BIT:
> + if (upper_32_bits(imm))
> + return AARCH64_BREAK_FAULT;
> + esz = 32;
> + break;
> + case AARCH64_INSN_VARIANT_64BIT:
> + insn |= AARCH64_INSN_SF_BIT;
> + esz = 64;
> + break;
> + default:
> + pr_err("%s: unknown variant encoding %d\n", __func__, variant);
> + return AARCH64_BREAK_FAULT;
> + }
> +
> + /*
> + * Inverse of Replicate(). Try to spot a repeating pattern
> + * with a pow2 stride.
> + */
> + for (tmp = esz / 2; tmp >= 2; tmp /= 2) {
> + u64 emask = BIT(tmp) - 1;
> +
> + if ((imm & emask) != ((imm >> (tmp / 2)) & emask))
> + break;
> +
> + esz = tmp;
> + mask = emask;
> + }
> +
> + /* N is only set if we're encoding a 64bit value */
> + n = esz == 64;
> +
> + /* Trim imm to the element size */
> + imm &= mask;
> +
> + /* That's how many ones we need to encode */
> + ones = hweight64(imm);
> +
> + /*
> + * imms is set to (ones - 1), prefixed with a string of ones
> + * and a zero if they fit. Cap it to 6 bits.
> + */
> + imms = ones - 1;
> + imms |= 0xf << ffs(esz);
> + imms &= BIT(6) - 1;
> +
> + /* Compute the rotation */
> + if (range_of_ones(imm)) {
> + /*
> + * Pattern: 0..01..10..0
> + *
> + * Compute how many rotate we need to align it right
> + */
> + ror = __ffs64(imm);
> + } else {
> + /*
> + * Pattern: 0..01..10..01..1
> + *
> + * Fill the unused top bits with ones, and check if
> + * the result is a valid immediate (all ones with a
> + * contiguous ranges of zeroes).
> + */
> + imm |= ~mask;
> + if (!range_of_ones(~imm))
> + return AARCH64_BREAK_FAULT;
> +
> + /*
> + * Compute the rotation to get a continuous set of
> + * ones, with the first bit set at position 0
> + */
> + ror = fls(~imm);
> + }
> +
> + /*
> + * immr is the number of bits we need to rotate back to the
> + * original set of ones. Note that this is relative to the
> + * element size...
> + */
> + immr = (esz - ror) % esz;
> +
> + insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_N, insn, n);
> + insn = aarch64_insn_encode_immediate(AARCH64_INSN_IMM_R, insn, immr);
> + return aarch64_insn_encode_immediate(AARCH64_INSN_IMM_S, insn, imms);
> +}
> +
> +u32 aarch64_insn_gen_logical_immediate(enum aarch64_insn_logic_type type,
> + enum aarch64_insn_variant variant,
> + enum aarch64_insn_register Rn,
> + enum aarch64_insn_register Rd,
> + u64 imm)
> +{
> + u32 insn;
> +
> + switch (type) {
> + case AARCH64_INSN_LOGIC_AND:
> + insn = aarch64_insn_get_and_imm_value();
> + break;
> + case AARCH64_INSN_LOGIC_ORR:
> + insn = aarch64_insn_get_orr_imm_value();
> + break;
> + case AARCH64_INSN_LOGIC_EOR:
> + insn = aarch64_insn_get_eor_imm_value();
> + break;
> + case AARCH64_INSN_LOGIC_AND_SETFLAGS:
> + insn = aarch64_insn_get_ands_imm_value();
> + break;
> + default:
> + pr_err("%s: unknown logical encoding %d\n", __func__, type);
> + return AARCH64_BREAK_FAULT;
> + }
> +
> + insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RD, insn, Rd);
> + insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, Rn);
> + return aarch64_encode_immediate(imm, variant, insn);
> +}
> --
> 2.14.2
>
next prev parent reply other threads:[~2018-01-15 11:26 UTC|newest]
Thread overview: 52+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-01-04 18:43 [PATCH v4 00/19] KVM/arm64: Randomise EL2 mappings Marc Zyngier
2018-01-04 18:43 ` [PATCH v4 01/19] arm64: asm-offsets: Avoid clashing DMA definitions Marc Zyngier
2018-01-04 18:43 ` [PATCH v4 02/19] arm64: asm-offsets: Remove unused definitions Marc Zyngier
2018-01-04 18:43 ` [PATCH v4 03/19] arm64: asm-offsets: Remove potential circular dependency Marc Zyngier
2018-01-15 8:34 ` Christoffer Dall
2018-01-15 8:42 ` Marc Zyngier
2018-01-15 9:46 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 04/19] arm64: alternatives: Enforce alignment of struct alt_instr Marc Zyngier
2018-01-15 9:11 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 05/19] arm64: alternatives: Add dynamic patching feature Marc Zyngier
2018-01-15 11:26 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 06/19] arm64: insn: Add N immediate encoding Marc Zyngier
2018-01-15 11:26 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 07/19] arm64: insn: Add encoder for bitwise operations using literals Marc Zyngier
2018-01-15 11:26 ` Christoffer Dall [this message]
2018-01-04 18:43 ` [PATCH v4 08/19] arm64: KVM: Dynamically patch the kernel/hyp VA mask Marc Zyngier
2018-01-15 11:47 ` Christoffer Dall
2018-02-15 13:11 ` Marc Zyngier
2018-02-16 9:02 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 09/19] arm64: cpufeatures: Drop the ARM64_HYP_OFFSET_LOW feature flag Marc Zyngier
2018-01-15 11:48 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 10/19] KVM: arm/arm64: Do not use kern_hyp_va() with kvm_vgic_global_state Marc Zyngier
2018-01-15 15:36 ` Christoffer Dall
2018-02-15 13:22 ` Marc Zyngier
2018-02-16 9:05 ` Christoffer Dall
2018-02-16 9:33 ` Marc Zyngier
2018-02-19 14:39 ` Christoffer Dall
2018-02-20 11:40 ` Marc Zyngier
2018-01-04 18:43 ` [PATCH v4 11/19] KVM: arm/arm64: Demote HYP VA range display to being a debug feature Marc Zyngier
2018-01-15 15:54 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 12/19] KVM: arm/arm64: Move ioremap calls to create_hyp_io_mappings Marc Zyngier
2018-01-15 18:07 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 13/19] KVM: arm/arm64: Keep GICv2 HYP VAs in kvm_vgic_global_state Marc Zyngier
2018-01-18 14:39 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 14/19] KVM: arm/arm64: Move HYP IO VAs to the "idmap" range Marc Zyngier
2018-01-18 14:39 ` Christoffer Dall
2018-02-15 13:52 ` Marc Zyngier
2018-02-16 9:25 ` Christoffer Dall
2018-02-16 15:20 ` Marc Zyngier
2018-01-04 18:43 ` [PATCH v4 15/19] arm64; insn: Add encoder for the EXTR instruction Marc Zyngier
2018-01-18 20:27 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 16/19] arm64: insn: Allow ADD/SUB (immediate) with LSL #12 Marc Zyngier
2018-01-18 20:28 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 17/19] arm64: KVM: Dynamically compute the HYP VA mask Marc Zyngier
2018-01-18 20:28 ` Christoffer Dall
2018-02-15 13:58 ` Marc Zyngier
2018-01-04 18:43 ` [PATCH v4 18/19] arm64: KVM: Introduce EL2 VA randomisation Marc Zyngier
2018-01-18 20:28 ` Christoffer Dall
2018-02-15 15:32 ` Marc Zyngier
2018-02-16 9:33 ` Christoffer Dall
2018-01-04 18:43 ` [PATCH v4 19/19] arm64: Update the KVM memory map documentation Marc Zyngier
2018-01-18 20:28 ` Christoffer Dall
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180115112619.GH21403@cbox \
--to=christoffer.dall@linaro.org \
--cc=linux-arm-kernel@lists.infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).