* [PATCH 0/5] crypto: chacha20 - Alignment fixes
@ 2017-11-22 19:51 Eric Biggers
2017-11-22 19:51 ` [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants Eric Biggers
` (5 more replies)
0 siblings, 6 replies; 14+ messages in thread
From: Eric Biggers @ 2017-11-22 19:51 UTC (permalink / raw)
To: linux-crypto, Herbert Xu
Cc: Theodore Ts'o, Jason A . Donenfeld, Martin Willi,
Ard Biesheuvel, linux-kernel, Eric Biggers
From: Eric Biggers <ebiggers@google.com>
This series fixes potentially unaligned memory accesses when loading the
initial state, key, and IV for ChaCha20, and when outputting each
keystream block.
It also removes the cra_alignmask from the generic and x86 ChaCha20
implementations, once it is no longer needed.
Eric Biggers (5):
crypto: chacha20 - Fix unaligned access when loading constants
crypto: chacha20 - Use unaligned access macros when loading key and IV
crypto: chacha20 - Remove cra_alignmask
crypto: x86/chacha20 - Remove cra_alignmask
crypto: chacha20 - Fix keystream alignment for chacha20_block()
arch/x86/crypto/chacha20_glue.c | 1 -
crypto/chacha20_generic.c | 33 +++++++++++++--------------------
drivers/char/random.c | 24 ++++++++++++------------
include/crypto/chacha20.h | 3 ++-
lib/chacha20.c | 2 +-
5 files changed, 28 insertions(+), 35 deletions(-)
--
2.15.0.448.gf294e3d99a-goog
^ permalink raw reply [flat|nested] 14+ messages in thread* [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers @ 2017-11-22 19:51 ` Eric Biggers 2017-11-22 20:26 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 2/5] crypto: chacha20 - Use unaligned access macros when loading key and IV Eric Biggers ` (4 subsequent siblings) 5 siblings, 1 reply; 14+ messages in thread From: Eric Biggers @ 2017-11-22 19:51 UTC (permalink / raw) To: linux-crypto, Herbert Xu Cc: Theodore Ts'o, Jason A . Donenfeld, Martin Willi, Ard Biesheuvel, linux-kernel, Eric Biggers From: Eric Biggers <ebiggers@google.com> The four 32-bit constants for the initial state of ChaCha20 were loaded from a char array which is not guaranteed to have the needed alignment. Fix it by just assigning the constants directly instead. Signed-off-by: Eric Biggers <ebiggers@google.com> --- crypto/chacha20_generic.c | 10 ++++------ 1 file changed, 4 insertions(+), 6 deletions(-) diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c index 4a45fa4890c0..ec84e7837aac 100644 --- a/crypto/chacha20_generic.c +++ b/crypto/chacha20_generic.c @@ -41,12 +41,10 @@ static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src, void crypto_chacha20_init(u32 *state, struct chacha20_ctx *ctx, u8 *iv) { - static const char constant[16] = "expand 32-byte k"; - - state[0] = le32_to_cpuvp(constant + 0); - state[1] = le32_to_cpuvp(constant + 4); - state[2] = le32_to_cpuvp(constant + 8); - state[3] = le32_to_cpuvp(constant + 12); + state[0] = 0x61707865; /* "expa" */ + state[1] = 0x3320646e; /* "nd 3" */ + state[2] = 0x79622d32; /* "2-by" */ + state[3] = 0x6b206574; /* "te k" */ state[4] = ctx->key[0]; state[5] = ctx->key[1]; state[6] = ctx->key[2]; -- 2.15.0.448.gf294e3d99a-goog ^ permalink raw reply related [flat|nested] 14+ messages in thread
* Re: [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants 2017-11-22 19:51 ` [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants Eric Biggers @ 2017-11-22 20:26 ` Ard Biesheuvel 0 siblings, 0 replies; 14+ messages in thread From: Ard Biesheuvel @ 2017-11-22 20:26 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: > From: Eric Biggers <ebiggers@google.com> > > The four 32-bit constants for the initial state of ChaCha20 were loaded > from a char array which is not guaranteed to have the needed alignment. > > Fix it by just assigning the constants directly instead. > > Signed-off-by: Eric Biggers <ebiggers@google.com> I'm not thrilled about the open coded hex numbers but I don't care enough to object. Acked-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> > --- > crypto/chacha20_generic.c | 10 ++++------ > 1 file changed, 4 insertions(+), 6 deletions(-) > > diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c > index 4a45fa4890c0..ec84e7837aac 100644 > --- a/crypto/chacha20_generic.c > +++ b/crypto/chacha20_generic.c > @@ -41,12 +41,10 @@ static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src, > > void crypto_chacha20_init(u32 *state, struct chacha20_ctx *ctx, u8 *iv) > { > - static const char constant[16] = "expand 32-byte k"; > - > - state[0] = le32_to_cpuvp(constant + 0); > - state[1] = le32_to_cpuvp(constant + 4); > - state[2] = le32_to_cpuvp(constant + 8); > - state[3] = le32_to_cpuvp(constant + 12); > + state[0] = 0x61707865; /* "expa" */ > + state[1] = 0x3320646e; /* "nd 3" */ > + state[2] = 0x79622d32; /* "2-by" */ > + state[3] = 0x6b206574; /* "te k" */ > state[4] = ctx->key[0]; > state[5] = ctx->key[1]; > state[6] = ctx->key[2]; > -- > 2.15.0.448.gf294e3d99a-goog > ^ permalink raw reply [flat|nested] 14+ messages in thread
* [PATCH 2/5] crypto: chacha20 - Use unaligned access macros when loading key and IV 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers 2017-11-22 19:51 ` [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants Eric Biggers @ 2017-11-22 19:51 ` Eric Biggers 2017-11-22 20:27 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 3/5] crypto: chacha20 - Remove cra_alignmask Eric Biggers ` (3 subsequent siblings) 5 siblings, 1 reply; 14+ messages in thread From: Eric Biggers @ 2017-11-22 19:51 UTC (permalink / raw) To: linux-crypto, Herbert Xu Cc: Theodore Ts'o, Jason A . Donenfeld, Martin Willi, Ard Biesheuvel, linux-kernel, Eric Biggers From: Eric Biggers <ebiggers@google.com> The generic ChaCha20 implementation has a cra_alignmask of 3, which ensures that the key passed into crypto_chacha20_setkey() and the IV passed into crypto_chacha20_init() are 4-byte aligned. However, these functions are also called from the ARM and ARM64 implementations of ChaCha20, which intentionally do not have a cra_alignmask set. This is broken because 32-bit words are being loaded from potentially-unaligned buffers without the unaligned access macros. Fix it by using the unaligned access macros when loading the key and IV. Signed-off-by: Eric Biggers <ebiggers@google.com> --- crypto/chacha20_generic.c | 16 ++++++---------- 1 file changed, 6 insertions(+), 10 deletions(-) diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c index ec84e7837aac..b5a10ebf1b82 100644 --- a/crypto/chacha20_generic.c +++ b/crypto/chacha20_generic.c @@ -9,16 +9,12 @@ * (at your option) any later version. */ +#include <asm/unaligned.h> #include <crypto/algapi.h> #include <crypto/chacha20.h> #include <crypto/internal/skcipher.h> #include <linux/module.h> -static inline u32 le32_to_cpuvp(const void *p) -{ - return le32_to_cpup(p); -} - static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src, unsigned int bytes) { @@ -53,10 +49,10 @@ void crypto_chacha20_init(u32 *state, struct chacha20_ctx *ctx, u8 *iv) state[9] = ctx->key[5]; state[10] = ctx->key[6]; state[11] = ctx->key[7]; - state[12] = le32_to_cpuvp(iv + 0); - state[13] = le32_to_cpuvp(iv + 4); - state[14] = le32_to_cpuvp(iv + 8); - state[15] = le32_to_cpuvp(iv + 12); + state[12] = get_unaligned_le32(iv + 0); + state[13] = get_unaligned_le32(iv + 4); + state[14] = get_unaligned_le32(iv + 8); + state[15] = get_unaligned_le32(iv + 12); } EXPORT_SYMBOL_GPL(crypto_chacha20_init); @@ -70,7 +66,7 @@ int crypto_chacha20_setkey(struct crypto_skcipher *tfm, const u8 *key, return -EINVAL; for (i = 0; i < ARRAY_SIZE(ctx->key); i++) - ctx->key[i] = le32_to_cpuvp(key + i * sizeof(u32)); + ctx->key[i] = get_unaligned_le32(key + i * sizeof(u32)); return 0; } -- 2.15.0.448.gf294e3d99a-goog ^ permalink raw reply related [flat|nested] 14+ messages in thread
* Re: [PATCH 2/5] crypto: chacha20 - Use unaligned access macros when loading key and IV 2017-11-22 19:51 ` [PATCH 2/5] crypto: chacha20 - Use unaligned access macros when loading key and IV Eric Biggers @ 2017-11-22 20:27 ` Ard Biesheuvel 0 siblings, 0 replies; 14+ messages in thread From: Ard Biesheuvel @ 2017-11-22 20:27 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: > From: Eric Biggers <ebiggers@google.com> > > The generic ChaCha20 implementation has a cra_alignmask of 3, which > ensures that the key passed into crypto_chacha20_setkey() and the IV > passed into crypto_chacha20_init() are 4-byte aligned. However, these > functions are also called from the ARM and ARM64 implementations of > ChaCha20, which intentionally do not have a cra_alignmask set. This is > broken because 32-bit words are being loaded from potentially-unaligned > buffers without the unaligned access macros. > > Fix it by using the unaligned access macros when loading the key and IV. > > Signed-off-by: Eric Biggers <ebiggers@google.com> Acked-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> > --- > crypto/chacha20_generic.c | 16 ++++++---------- > 1 file changed, 6 insertions(+), 10 deletions(-) > > diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c > index ec84e7837aac..b5a10ebf1b82 100644 > --- a/crypto/chacha20_generic.c > +++ b/crypto/chacha20_generic.c > @@ -9,16 +9,12 @@ > * (at your option) any later version. > */ > > +#include <asm/unaligned.h> > #include <crypto/algapi.h> > #include <crypto/chacha20.h> > #include <crypto/internal/skcipher.h> > #include <linux/module.h> > > -static inline u32 le32_to_cpuvp(const void *p) > -{ > - return le32_to_cpup(p); > -} > - > static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src, > unsigned int bytes) > { > @@ -53,10 +49,10 @@ void crypto_chacha20_init(u32 *state, struct chacha20_ctx *ctx, u8 *iv) > state[9] = ctx->key[5]; > state[10] = ctx->key[6]; > state[11] = ctx->key[7]; > - state[12] = le32_to_cpuvp(iv + 0); > - state[13] = le32_to_cpuvp(iv + 4); > - state[14] = le32_to_cpuvp(iv + 8); > - state[15] = le32_to_cpuvp(iv + 12); > + state[12] = get_unaligned_le32(iv + 0); > + state[13] = get_unaligned_le32(iv + 4); > + state[14] = get_unaligned_le32(iv + 8); > + state[15] = get_unaligned_le32(iv + 12); > } > EXPORT_SYMBOL_GPL(crypto_chacha20_init); > > @@ -70,7 +66,7 @@ int crypto_chacha20_setkey(struct crypto_skcipher *tfm, const u8 *key, > return -EINVAL; > > for (i = 0; i < ARRAY_SIZE(ctx->key); i++) > - ctx->key[i] = le32_to_cpuvp(key + i * sizeof(u32)); > + ctx->key[i] = get_unaligned_le32(key + i * sizeof(u32)); > > return 0; > } > -- > 2.15.0.448.gf294e3d99a-goog > ^ permalink raw reply [flat|nested] 14+ messages in thread
* [PATCH 3/5] crypto: chacha20 - Remove cra_alignmask 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers 2017-11-22 19:51 ` [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants Eric Biggers 2017-11-22 19:51 ` [PATCH 2/5] crypto: chacha20 - Use unaligned access macros when loading key and IV Eric Biggers @ 2017-11-22 19:51 ` Eric Biggers 2017-11-22 20:30 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 4/5] crypto: x86/chacha20 " Eric Biggers ` (2 subsequent siblings) 5 siblings, 1 reply; 14+ messages in thread From: Eric Biggers @ 2017-11-22 19:51 UTC (permalink / raw) To: linux-crypto, Herbert Xu Cc: Theodore Ts'o, Jason A . Donenfeld, Martin Willi, Ard Biesheuvel, linux-kernel, Eric Biggers From: Eric Biggers <ebiggers@google.com> Now that crypto_chacha20_setkey() and crypto_chacha20_init() use the unaligned access macros and crypto_xor() also accepts unaligned buffers, there is no need to have a cra_alignmask set for chacha20-generic. Signed-off-by: Eric Biggers <ebiggers@google.com> --- crypto/chacha20_generic.c | 1 - 1 file changed, 1 deletion(-) diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c index b5a10ebf1b82..bb4affbd591c 100644 --- a/crypto/chacha20_generic.c +++ b/crypto/chacha20_generic.c @@ -105,7 +105,6 @@ static struct skcipher_alg alg = { .base.cra_priority = 100, .base.cra_blocksize = 1, .base.cra_ctxsize = sizeof(struct chacha20_ctx), - .base.cra_alignmask = sizeof(u32) - 1, .base.cra_module = THIS_MODULE, .min_keysize = CHACHA20_KEY_SIZE, -- 2.15.0.448.gf294e3d99a-goog ^ permalink raw reply related [flat|nested] 14+ messages in thread
* Re: [PATCH 3/5] crypto: chacha20 - Remove cra_alignmask 2017-11-22 19:51 ` [PATCH 3/5] crypto: chacha20 - Remove cra_alignmask Eric Biggers @ 2017-11-22 20:30 ` Ard Biesheuvel 0 siblings, 0 replies; 14+ messages in thread From: Ard Biesheuvel @ 2017-11-22 20:30 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: > From: Eric Biggers <ebiggers@google.com> > > Now that crypto_chacha20_setkey() and crypto_chacha20_init() use the > unaligned access macros and crypto_xor() also accepts unaligned buffers, > there is no need to have a cra_alignmask set for chacha20-generic. > > Signed-off-by: Eric Biggers <ebiggers@google.com> Acked-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> > --- > crypto/chacha20_generic.c | 1 - > 1 file changed, 1 deletion(-) > > diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c > index b5a10ebf1b82..bb4affbd591c 100644 > --- a/crypto/chacha20_generic.c > +++ b/crypto/chacha20_generic.c > @@ -105,7 +105,6 @@ static struct skcipher_alg alg = { > .base.cra_priority = 100, > .base.cra_blocksize = 1, > .base.cra_ctxsize = sizeof(struct chacha20_ctx), > - .base.cra_alignmask = sizeof(u32) - 1, > .base.cra_module = THIS_MODULE, > > .min_keysize = CHACHA20_KEY_SIZE, > -- > 2.15.0.448.gf294e3d99a-goog > ^ permalink raw reply [flat|nested] 14+ messages in thread
* [PATCH 4/5] crypto: x86/chacha20 - Remove cra_alignmask 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers ` (2 preceding siblings ...) 2017-11-22 19:51 ` [PATCH 3/5] crypto: chacha20 - Remove cra_alignmask Eric Biggers @ 2017-11-22 19:51 ` Eric Biggers 2017-11-22 20:31 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() Eric Biggers 2017-11-29 6:39 ` [PATCH 0/5] crypto: chacha20 - Alignment fixes Herbert Xu 5 siblings, 1 reply; 14+ messages in thread From: Eric Biggers @ 2017-11-22 19:51 UTC (permalink / raw) To: linux-crypto, Herbert Xu Cc: Theodore Ts'o, Jason A . Donenfeld, Martin Willi, Ard Biesheuvel, linux-kernel, Eric Biggers From: Eric Biggers <ebiggers@google.com> Now that the generic ChaCha20 implementation no longer needs a cra_alignmask, the x86 one doesn't either -- given that the x86 implementation doesn't need the alignment itself. Signed-off-by: Eric Biggers <ebiggers@google.com> --- arch/x86/crypto/chacha20_glue.c | 1 - 1 file changed, 1 deletion(-) diff --git a/arch/x86/crypto/chacha20_glue.c b/arch/x86/crypto/chacha20_glue.c index 1e6af1b35f7b..dce7c5d39c2f 100644 --- a/arch/x86/crypto/chacha20_glue.c +++ b/arch/x86/crypto/chacha20_glue.c @@ -107,7 +107,6 @@ static struct skcipher_alg alg = { .base.cra_priority = 300, .base.cra_blocksize = 1, .base.cra_ctxsize = sizeof(struct chacha20_ctx), - .base.cra_alignmask = sizeof(u32) - 1, .base.cra_module = THIS_MODULE, .min_keysize = CHACHA20_KEY_SIZE, -- 2.15.0.448.gf294e3d99a-goog ^ permalink raw reply related [flat|nested] 14+ messages in thread
* Re: [PATCH 4/5] crypto: x86/chacha20 - Remove cra_alignmask 2017-11-22 19:51 ` [PATCH 4/5] crypto: x86/chacha20 " Eric Biggers @ 2017-11-22 20:31 ` Ard Biesheuvel 0 siblings, 0 replies; 14+ messages in thread From: Ard Biesheuvel @ 2017-11-22 20:31 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: > From: Eric Biggers <ebiggers@google.com> > > Now that the generic ChaCha20 implementation no longer needs a > cra_alignmask, the x86 one doesn't either -- given that the x86 > implementation doesn't need the alignment itself. > > Signed-off-by: Eric Biggers <ebiggers@google.com> Acked-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> > --- > arch/x86/crypto/chacha20_glue.c | 1 - > 1 file changed, 1 deletion(-) > > diff --git a/arch/x86/crypto/chacha20_glue.c b/arch/x86/crypto/chacha20_glue.c > index 1e6af1b35f7b..dce7c5d39c2f 100644 > --- a/arch/x86/crypto/chacha20_glue.c > +++ b/arch/x86/crypto/chacha20_glue.c > @@ -107,7 +107,6 @@ static struct skcipher_alg alg = { > .base.cra_priority = 300, > .base.cra_blocksize = 1, > .base.cra_ctxsize = sizeof(struct chacha20_ctx), > - .base.cra_alignmask = sizeof(u32) - 1, > .base.cra_module = THIS_MODULE, > > .min_keysize = CHACHA20_KEY_SIZE, > -- > 2.15.0.448.gf294e3d99a-goog > ^ permalink raw reply [flat|nested] 14+ messages in thread
* [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers ` (3 preceding siblings ...) 2017-11-22 19:51 ` [PATCH 4/5] crypto: x86/chacha20 " Eric Biggers @ 2017-11-22 19:51 ` Eric Biggers 2017-11-22 20:51 ` Ard Biesheuvel 2017-11-29 6:39 ` [PATCH 0/5] crypto: chacha20 - Alignment fixes Herbert Xu 5 siblings, 1 reply; 14+ messages in thread From: Eric Biggers @ 2017-11-22 19:51 UTC (permalink / raw) To: linux-crypto, Herbert Xu Cc: Theodore Ts'o, Jason A . Donenfeld, Martin Willi, Ard Biesheuvel, linux-kernel, Eric Biggers From: Eric Biggers <ebiggers@google.com> When chacha20_block() outputs the keystream block, it uses 'u32' stores directly. However, the callers (crypto/chacha20_generic.c and drivers/char/random.c) declare the keystream buffer as a 'u8' array, which is not guaranteed to have the needed alignment. Fix it by having both callers declare the keystream as a 'u32' array. For now this is preferable to switching over to the unaligned access macros because chacha20_block() is only being used in cases where we can easily control the alignment (stack buffers). Signed-off-by: Eric Biggers <ebiggers@google.com> --- crypto/chacha20_generic.c | 6 +++--- drivers/char/random.c | 24 ++++++++++++------------ include/crypto/chacha20.h | 3 ++- lib/chacha20.c | 2 +- 4 files changed, 18 insertions(+), 17 deletions(-) diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c index bb4affbd591c..e451c3cb6a56 100644 --- a/crypto/chacha20_generic.c +++ b/crypto/chacha20_generic.c @@ -18,20 +18,20 @@ static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src, unsigned int bytes) { - u8 stream[CHACHA20_BLOCK_SIZE]; + u32 stream[CHACHA20_BLOCK_WORDS]; if (dst != src) memcpy(dst, src, bytes); while (bytes >= CHACHA20_BLOCK_SIZE) { chacha20_block(state, stream); - crypto_xor(dst, stream, CHACHA20_BLOCK_SIZE); + crypto_xor(dst, (const u8 *)stream, CHACHA20_BLOCK_SIZE); bytes -= CHACHA20_BLOCK_SIZE; dst += CHACHA20_BLOCK_SIZE; } if (bytes) { chacha20_block(state, stream); - crypto_xor(dst, stream, bytes); + crypto_xor(dst, (const u8 *)stream, bytes); } } diff --git a/drivers/char/random.c b/drivers/char/random.c index ec42c8bb9b0d..11304bbc78cc 100644 --- a/drivers/char/random.c +++ b/drivers/char/random.c @@ -431,9 +431,9 @@ static int crng_init = 0; static int crng_init_cnt = 0; #define CRNG_INIT_CNT_THRESH (2*CHACHA20_KEY_SIZE) static void _extract_crng(struct crng_state *crng, - __u8 out[CHACHA20_BLOCK_SIZE]); + __u32 out[CHACHA20_BLOCK_WORDS]); static void _crng_backtrack_protect(struct crng_state *crng, - __u8 tmp[CHACHA20_BLOCK_SIZE], int used); + __u32 tmp[CHACHA20_BLOCK_WORDS], int used); static void process_random_ready_list(void); static void _get_random_bytes(void *buf, int nbytes); @@ -817,7 +817,7 @@ static void crng_reseed(struct crng_state *crng, struct entropy_store *r) unsigned long flags; int i, num; union { - __u8 block[CHACHA20_BLOCK_SIZE]; + __u32 block[CHACHA20_BLOCK_WORDS]; __u32 key[8]; } buf; @@ -851,7 +851,7 @@ static void crng_reseed(struct crng_state *crng, struct entropy_store *r) } static void _extract_crng(struct crng_state *crng, - __u8 out[CHACHA20_BLOCK_SIZE]) + __u32 out[CHACHA20_BLOCK_WORDS]) { unsigned long v, flags; @@ -867,7 +867,7 @@ static void _extract_crng(struct crng_state *crng, spin_unlock_irqrestore(&crng->lock, flags); } -static void extract_crng(__u8 out[CHACHA20_BLOCK_SIZE]) +static void extract_crng(__u32 out[CHACHA20_BLOCK_WORDS]) { struct crng_state *crng = NULL; @@ -885,7 +885,7 @@ static void extract_crng(__u8 out[CHACHA20_BLOCK_SIZE]) * enough) to mutate the CRNG key to provide backtracking protection. */ static void _crng_backtrack_protect(struct crng_state *crng, - __u8 tmp[CHACHA20_BLOCK_SIZE], int used) + __u32 tmp[CHACHA20_BLOCK_WORDS], int used) { unsigned long flags; __u32 *s, *d; @@ -897,14 +897,14 @@ static void _crng_backtrack_protect(struct crng_state *crng, used = 0; } spin_lock_irqsave(&crng->lock, flags); - s = (__u32 *) &tmp[used]; + s = &tmp[used / sizeof(__u32)]; d = &crng->state[4]; for (i=0; i < 8; i++) *d++ ^= *s++; spin_unlock_irqrestore(&crng->lock, flags); } -static void crng_backtrack_protect(__u8 tmp[CHACHA20_BLOCK_SIZE], int used) +static void crng_backtrack_protect(__u32 tmp[CHACHA20_BLOCK_WORDS], int used) { struct crng_state *crng = NULL; @@ -920,7 +920,7 @@ static void crng_backtrack_protect(__u8 tmp[CHACHA20_BLOCK_SIZE], int used) static ssize_t extract_crng_user(void __user *buf, size_t nbytes) { ssize_t ret = 0, i = CHACHA20_BLOCK_SIZE; - __u8 tmp[CHACHA20_BLOCK_SIZE]; + __u32 tmp[CHACHA20_BLOCK_WORDS]; int large_request = (nbytes > 256); while (nbytes) { @@ -1507,7 +1507,7 @@ static void _warn_unseeded_randomness(const char *func_name, void *caller, */ static void _get_random_bytes(void *buf, int nbytes) { - __u8 tmp[CHACHA20_BLOCK_SIZE]; + __u32 tmp[CHACHA20_BLOCK_WORDS]; trace_get_random_bytes(nbytes, _RET_IP_); @@ -2114,7 +2114,7 @@ u64 get_random_u64(void) if (use_lock) read_lock_irqsave(&batched_entropy_reset_lock, flags); if (batch->position % ARRAY_SIZE(batch->entropy_u64) == 0) { - extract_crng((u8 *)batch->entropy_u64); + extract_crng((__u32 *)batch->entropy_u64); batch->position = 0; } ret = batch->entropy_u64[batch->position++]; @@ -2144,7 +2144,7 @@ u32 get_random_u32(void) if (use_lock) read_lock_irqsave(&batched_entropy_reset_lock, flags); if (batch->position % ARRAY_SIZE(batch->entropy_u32) == 0) { - extract_crng((u8 *)batch->entropy_u32); + extract_crng(batch->entropy_u32); batch->position = 0; } ret = batch->entropy_u32[batch->position++]; diff --git a/include/crypto/chacha20.h b/include/crypto/chacha20.h index caaa470389e0..b83d66073db0 100644 --- a/include/crypto/chacha20.h +++ b/include/crypto/chacha20.h @@ -13,12 +13,13 @@ #define CHACHA20_IV_SIZE 16 #define CHACHA20_KEY_SIZE 32 #define CHACHA20_BLOCK_SIZE 64 +#define CHACHA20_BLOCK_WORDS (CHACHA20_BLOCK_SIZE / sizeof(u32)) struct chacha20_ctx { u32 key[8]; }; -void chacha20_block(u32 *state, void *stream); +void chacha20_block(u32 *state, u32 *stream); void crypto_chacha20_init(u32 *state, struct chacha20_ctx *ctx, u8 *iv); int crypto_chacha20_setkey(struct crypto_skcipher *tfm, const u8 *key, unsigned int keysize); diff --git a/lib/chacha20.c b/lib/chacha20.c index 250ceed9ec9a..29d3801dee24 100644 --- a/lib/chacha20.c +++ b/lib/chacha20.c @@ -21,7 +21,7 @@ static inline u32 rotl32(u32 v, u8 n) return (v << n) | (v >> (sizeof(v) * 8 - n)); } -extern void chacha20_block(u32 *state, void *stream) +void chacha20_block(u32 *state, u32 *stream) { u32 x[16], *out = stream; int i; -- 2.15.0.448.gf294e3d99a-goog ^ permalink raw reply related [flat|nested] 14+ messages in thread
* Re: [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() 2017-11-22 19:51 ` [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() Eric Biggers @ 2017-11-22 20:51 ` Ard Biesheuvel 2017-11-22 21:29 ` Eric Biggers 0 siblings, 1 reply; 14+ messages in thread From: Ard Biesheuvel @ 2017-11-22 20:51 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: > From: Eric Biggers <ebiggers@google.com> > > When chacha20_block() outputs the keystream block, it uses 'u32' stores > directly. However, the callers (crypto/chacha20_generic.c and > drivers/char/random.c) declare the keystream buffer as a 'u8' array, > which is not guaranteed to have the needed alignment. > > Fix it by having both callers declare the keystream as a 'u32' array. > For now this is preferable to switching over to the unaligned access > macros because chacha20_block() is only being used in cases where we can > easily control the alignment (stack buffers). > Given this paragraph, I think we agree the correct way to fix this would be to make chacha20_block() adhere to its prototype, so if we deviate from that, there should be a good reason. On which architecture that cares about alignment is this expected to result in a measurable performance benefit? > Signed-off-by: Eric Biggers <ebiggers@google.com> > --- > crypto/chacha20_generic.c | 6 +++--- > drivers/char/random.c | 24 ++++++++++++------------ > include/crypto/chacha20.h | 3 ++- > lib/chacha20.c | 2 +- > 4 files changed, 18 insertions(+), 17 deletions(-) > > diff --git a/crypto/chacha20_generic.c b/crypto/chacha20_generic.c > index bb4affbd591c..e451c3cb6a56 100644 > --- a/crypto/chacha20_generic.c > +++ b/crypto/chacha20_generic.c > @@ -18,20 +18,20 @@ > static void chacha20_docrypt(u32 *state, u8 *dst, const u8 *src, > unsigned int bytes) > { > - u8 stream[CHACHA20_BLOCK_SIZE]; > + u32 stream[CHACHA20_BLOCK_WORDS]; > > if (dst != src) > memcpy(dst, src, bytes); > > while (bytes >= CHACHA20_BLOCK_SIZE) { > chacha20_block(state, stream); > - crypto_xor(dst, stream, CHACHA20_BLOCK_SIZE); > + crypto_xor(dst, (const u8 *)stream, CHACHA20_BLOCK_SIZE); > bytes -= CHACHA20_BLOCK_SIZE; > dst += CHACHA20_BLOCK_SIZE; > } > if (bytes) { > chacha20_block(state, stream); > - crypto_xor(dst, stream, bytes); > + crypto_xor(dst, (const u8 *)stream, bytes); > } > } > > diff --git a/drivers/char/random.c b/drivers/char/random.c > index ec42c8bb9b0d..11304bbc78cc 100644 > --- a/drivers/char/random.c > +++ b/drivers/char/random.c > @@ -431,9 +431,9 @@ static int crng_init = 0; > static int crng_init_cnt = 0; > #define CRNG_INIT_CNT_THRESH (2*CHACHA20_KEY_SIZE) > static void _extract_crng(struct crng_state *crng, > - __u8 out[CHACHA20_BLOCK_SIZE]); > + __u32 out[CHACHA20_BLOCK_WORDS]); > static void _crng_backtrack_protect(struct crng_state *crng, > - __u8 tmp[CHACHA20_BLOCK_SIZE], int used); > + __u32 tmp[CHACHA20_BLOCK_WORDS], int used); > static void process_random_ready_list(void); > static void _get_random_bytes(void *buf, int nbytes); > > @@ -817,7 +817,7 @@ static void crng_reseed(struct crng_state *crng, struct entropy_store *r) > unsigned long flags; > int i, num; > union { > - __u8 block[CHACHA20_BLOCK_SIZE]; > + __u32 block[CHACHA20_BLOCK_WORDS]; > __u32 key[8]; > } buf; > > @@ -851,7 +851,7 @@ static void crng_reseed(struct crng_state *crng, struct entropy_store *r) > } > > static void _extract_crng(struct crng_state *crng, > - __u8 out[CHACHA20_BLOCK_SIZE]) > + __u32 out[CHACHA20_BLOCK_WORDS]) > { > unsigned long v, flags; > > @@ -867,7 +867,7 @@ static void _extract_crng(struct crng_state *crng, > spin_unlock_irqrestore(&crng->lock, flags); > } > > -static void extract_crng(__u8 out[CHACHA20_BLOCK_SIZE]) > +static void extract_crng(__u32 out[CHACHA20_BLOCK_WORDS]) > { > struct crng_state *crng = NULL; > > @@ -885,7 +885,7 @@ static void extract_crng(__u8 out[CHACHA20_BLOCK_SIZE]) > * enough) to mutate the CRNG key to provide backtracking protection. > */ > static void _crng_backtrack_protect(struct crng_state *crng, > - __u8 tmp[CHACHA20_BLOCK_SIZE], int used) > + __u32 tmp[CHACHA20_BLOCK_WORDS], int used) > { > unsigned long flags; > __u32 *s, *d; > @@ -897,14 +897,14 @@ static void _crng_backtrack_protect(struct crng_state *crng, > used = 0; > } > spin_lock_irqsave(&crng->lock, flags); > - s = (__u32 *) &tmp[used]; > + s = &tmp[used / sizeof(__u32)]; > d = &crng->state[4]; > for (i=0; i < 8; i++) > *d++ ^= *s++; > spin_unlock_irqrestore(&crng->lock, flags); > } > > -static void crng_backtrack_protect(__u8 tmp[CHACHA20_BLOCK_SIZE], int used) > +static void crng_backtrack_protect(__u32 tmp[CHACHA20_BLOCK_WORDS], int used) > { > struct crng_state *crng = NULL; > > @@ -920,7 +920,7 @@ static void crng_backtrack_protect(__u8 tmp[CHACHA20_BLOCK_SIZE], int used) > static ssize_t extract_crng_user(void __user *buf, size_t nbytes) > { > ssize_t ret = 0, i = CHACHA20_BLOCK_SIZE; > - __u8 tmp[CHACHA20_BLOCK_SIZE]; > + __u32 tmp[CHACHA20_BLOCK_WORDS]; > int large_request = (nbytes > 256); > > while (nbytes) { > @@ -1507,7 +1507,7 @@ static void _warn_unseeded_randomness(const char *func_name, void *caller, > */ > static void _get_random_bytes(void *buf, int nbytes) > { > - __u8 tmp[CHACHA20_BLOCK_SIZE]; > + __u32 tmp[CHACHA20_BLOCK_WORDS]; > > trace_get_random_bytes(nbytes, _RET_IP_); > > @@ -2114,7 +2114,7 @@ u64 get_random_u64(void) > if (use_lock) > read_lock_irqsave(&batched_entropy_reset_lock, flags); > if (batch->position % ARRAY_SIZE(batch->entropy_u64) == 0) { > - extract_crng((u8 *)batch->entropy_u64); > + extract_crng((__u32 *)batch->entropy_u64); > batch->position = 0; > } > ret = batch->entropy_u64[batch->position++]; > @@ -2144,7 +2144,7 @@ u32 get_random_u32(void) > if (use_lock) > read_lock_irqsave(&batched_entropy_reset_lock, flags); > if (batch->position % ARRAY_SIZE(batch->entropy_u32) == 0) { > - extract_crng((u8 *)batch->entropy_u32); > + extract_crng(batch->entropy_u32); > batch->position = 0; > } > ret = batch->entropy_u32[batch->position++]; > diff --git a/include/crypto/chacha20.h b/include/crypto/chacha20.h > index caaa470389e0..b83d66073db0 100644 > --- a/include/crypto/chacha20.h > +++ b/include/crypto/chacha20.h > @@ -13,12 +13,13 @@ > #define CHACHA20_IV_SIZE 16 > #define CHACHA20_KEY_SIZE 32 > #define CHACHA20_BLOCK_SIZE 64 > +#define CHACHA20_BLOCK_WORDS (CHACHA20_BLOCK_SIZE / sizeof(u32)) > > struct chacha20_ctx { > u32 key[8]; > }; > > -void chacha20_block(u32 *state, void *stream); > +void chacha20_block(u32 *state, u32 *stream); > void crypto_chacha20_init(u32 *state, struct chacha20_ctx *ctx, u8 *iv); > int crypto_chacha20_setkey(struct crypto_skcipher *tfm, const u8 *key, > unsigned int keysize); > diff --git a/lib/chacha20.c b/lib/chacha20.c > index 250ceed9ec9a..29d3801dee24 100644 > --- a/lib/chacha20.c > +++ b/lib/chacha20.c > @@ -21,7 +21,7 @@ static inline u32 rotl32(u32 v, u8 n) > return (v << n) | (v >> (sizeof(v) * 8 - n)); > } > > -extern void chacha20_block(u32 *state, void *stream) > +void chacha20_block(u32 *state, u32 *stream) > { > u32 x[16], *out = stream; > int i; > -- > 2.15.0.448.gf294e3d99a-goog > ^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() 2017-11-22 20:51 ` Ard Biesheuvel @ 2017-11-22 21:29 ` Eric Biggers 2017-11-22 22:06 ` Ard Biesheuvel 0 siblings, 1 reply; 14+ messages in thread From: Eric Biggers @ 2017-11-22 21:29 UTC (permalink / raw) To: Ard Biesheuvel Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On Wed, Nov 22, 2017 at 08:51:57PM +0000, Ard Biesheuvel wrote: > On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: > > From: Eric Biggers <ebiggers@google.com> > > > > When chacha20_block() outputs the keystream block, it uses 'u32' stores > > directly. However, the callers (crypto/chacha20_generic.c and > > drivers/char/random.c) declare the keystream buffer as a 'u8' array, > > which is not guaranteed to have the needed alignment. > > > > Fix it by having both callers declare the keystream as a 'u32' array. > > For now this is preferable to switching over to the unaligned access > > macros because chacha20_block() is only being used in cases where we can > > easily control the alignment (stack buffers). > > > > Given this paragraph, I think we agree the correct way to fix this > would be to make chacha20_block() adhere to its prototype, so if we > deviate from that, there should be a good reason. On which > architecture that cares about alignment is this expected to result in > a measurable performance benefit? > Well, variables on the stack tend to be 4 or even 8-byte aligned anyway, so this change probably doesn't make a difference in practice currently. But it still should be fixed, in case it does become a problem. We could certainly leave the type as u8 array and use put_unaligned_le32() instead; that would be a simpler change. But that would be slower on architectures where a potentially-unaligned access requires multiple instructions. Eric ^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() 2017-11-22 21:29 ` Eric Biggers @ 2017-11-22 22:06 ` Ard Biesheuvel 0 siblings, 0 replies; 14+ messages in thread From: Ard Biesheuvel @ 2017-11-22 22:06 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto@vger.kernel.org, Herbert Xu, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, linux-kernel@vger.kernel.org, Eric Biggers On 22 November 2017 at 21:29, Eric Biggers <ebiggers3@gmail.com> wrote: > On Wed, Nov 22, 2017 at 08:51:57PM +0000, Ard Biesheuvel wrote: >> On 22 November 2017 at 19:51, Eric Biggers <ebiggers3@gmail.com> wrote: >> > From: Eric Biggers <ebiggers@google.com> >> > >> > When chacha20_block() outputs the keystream block, it uses 'u32' stores >> > directly. However, the callers (crypto/chacha20_generic.c and >> > drivers/char/random.c) declare the keystream buffer as a 'u8' array, >> > which is not guaranteed to have the needed alignment. >> > >> > Fix it by having both callers declare the keystream as a 'u32' array. >> > For now this is preferable to switching over to the unaligned access >> > macros because chacha20_block() is only being used in cases where we can >> > easily control the alignment (stack buffers). >> > >> >> Given this paragraph, I think we agree the correct way to fix this >> would be to make chacha20_block() adhere to its prototype, so if we >> deviate from that, there should be a good reason. On which >> architecture that cares about alignment is this expected to result in >> a measurable performance benefit? >> > > Well, variables on the stack tend to be 4 or even 8-byte aligned anyway, so this > change probably doesn't make a difference in practice currently. But it still > should be fixed, in case it does become a problem. > Agreed. > We could certainly leave the type as u8 array and use put_unaligned_le32() > instead; that would be a simpler change. But that would be slower on > architectures where a potentially-unaligned access requires multiple > instructions. > The access itself would be slower, yes. But given the amount of work performed in chacha20_block(), I seriously doubt that would actually matter in practice. ^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [PATCH 0/5] crypto: chacha20 - Alignment fixes 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers ` (4 preceding siblings ...) 2017-11-22 19:51 ` [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() Eric Biggers @ 2017-11-29 6:39 ` Herbert Xu 5 siblings, 0 replies; 14+ messages in thread From: Herbert Xu @ 2017-11-29 6:39 UTC (permalink / raw) To: Eric Biggers Cc: linux-crypto, Theodore Ts'o, Jason A . Donenfeld, Martin Willi, Ard Biesheuvel, linux-kernel, Eric Biggers On Wed, Nov 22, 2017 at 11:51:34AM -0800, Eric Biggers wrote: > From: Eric Biggers <ebiggers@google.com> > > This series fixes potentially unaligned memory accesses when loading the > initial state, key, and IV for ChaCha20, and when outputting each > keystream block. > > It also removes the cra_alignmask from the generic and x86 ChaCha20 > implementations, once it is no longer needed. > > Eric Biggers (5): > crypto: chacha20 - Fix unaligned access when loading constants > crypto: chacha20 - Use unaligned access macros when loading key and IV > crypto: chacha20 - Remove cra_alignmask > crypto: x86/chacha20 - Remove cra_alignmask > crypto: chacha20 - Fix keystream alignment for chacha20_block() All applied. Thanks. -- Email: Herbert Xu <herbert@gondor.apana.org.au> Home Page: http://gondor.apana.org.au/~herbert/ PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt ^ permalink raw reply [flat|nested] 14+ messages in thread
end of thread, other threads:[~2017-11-29 6:40 UTC | newest] Thread overview: 14+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2017-11-22 19:51 [PATCH 0/5] crypto: chacha20 - Alignment fixes Eric Biggers 2017-11-22 19:51 ` [PATCH 1/5] crypto: chacha20 - Fix unaligned access when loading constants Eric Biggers 2017-11-22 20:26 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 2/5] crypto: chacha20 - Use unaligned access macros when loading key and IV Eric Biggers 2017-11-22 20:27 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 3/5] crypto: chacha20 - Remove cra_alignmask Eric Biggers 2017-11-22 20:30 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 4/5] crypto: x86/chacha20 " Eric Biggers 2017-11-22 20:31 ` Ard Biesheuvel 2017-11-22 19:51 ` [PATCH 5/5] crypto: chacha20 - Fix keystream alignment for chacha20_block() Eric Biggers 2017-11-22 20:51 ` Ard Biesheuvel 2017-11-22 21:29 ` Eric Biggers 2017-11-22 22:06 ` Ard Biesheuvel 2017-11-29 6:39 ` [PATCH 0/5] crypto: chacha20 - Alignment fixes Herbert Xu
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).