linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2] crypto: ctr - avoid VLA use
@ 2018-03-15 11:18 Salvatore Mesoraca
  2018-03-23 15:36 ` Herbert Xu
  0 siblings, 1 reply; 3+ messages in thread
From: Salvatore Mesoraca @ 2018-03-15 11:18 UTC (permalink / raw)
  To: linux-kernel
  Cc: kernel-hardening, linux-crypto, David S. Miller, Herbert Xu,
	Kees Cook, Salvatore Mesoraca, Eric Biggers

All ciphers implemented in Linux have a block size less than or
equal to 16 bytes and the most demanding hw require 16 bytes
alignment for the block buffer.
We avoid 2 VLAs[1] by always allocating 16 bytes with 16 bytes
alignment, unless the architecture supports efficient unaligned
accesses.
We also check the selected cipher at instance creation time, if
it doesn't comply with these limits, we fail the creation.

[1] https://lkml.org/lkml/2018/3/7/621

Signed-off-by: Salvatore Mesoraca <s.mesoraca16@gmail.com>
---
 crypto/ctr.c | 20 ++++++++++++++++++--
 1 file changed, 18 insertions(+), 2 deletions(-)

diff --git a/crypto/ctr.c b/crypto/ctr.c
index 854d924..2c9f80f 100644
--- a/crypto/ctr.c
+++ b/crypto/ctr.c
@@ -21,6 +21,14 @@
 #include <linux/scatterlist.h>
 #include <linux/slab.h>
 
+#define MAX_BLOCKSIZE 16
+
+#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
+#define MAX_ALIGNMASK 15
+#else
+#define MAX_ALIGNMASK 0
+#endif
+
 struct crypto_ctr_ctx {
 	struct crypto_cipher *child;
 };
@@ -58,7 +66,7 @@ static void crypto_ctr_crypt_final(struct blkcipher_walk *walk,
 	unsigned int bsize = crypto_cipher_blocksize(tfm);
 	unsigned long alignmask = crypto_cipher_alignmask(tfm);
 	u8 *ctrblk = walk->iv;
-	u8 tmp[bsize + alignmask];
+	u8 tmp[MAX_BLOCKSIZE + MAX_ALIGNMASK];
 	u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1);
 	u8 *src = walk->src.virt.addr;
 	u8 *dst = walk->dst.virt.addr;
@@ -106,7 +114,7 @@ static int crypto_ctr_crypt_inplace(struct blkcipher_walk *walk,
 	unsigned int nbytes = walk->nbytes;
 	u8 *ctrblk = walk->iv;
 	u8 *src = walk->src.virt.addr;
-	u8 tmp[bsize + alignmask];
+	u8 tmp[MAX_BLOCKSIZE + MAX_ALIGNMASK];
 	u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1);
 
 	do {
@@ -206,6 +214,14 @@ static struct crypto_instance *crypto_ctr_alloc(struct rtattr **tb)
 	if (alg->cra_blocksize < 4)
 		goto out_put_alg;
 
+	/* Block size must be <= MAX_BLOCKSIZE. */
+	if (alg->cra_blocksize > MAX_BLOCKSIZE)
+		goto out_put_alg;
+
+	/* Alignmask must be <= MAX_ALIGNMASK. */
+	if (alg->cra_alignmask > MAX_ALIGNMASK)
+		goto out_put_alg;
+
 	/* If this is false we'd fail the alignment of crypto_inc. */
 	if (alg->cra_blocksize % 4)
 		goto out_put_alg;
-- 
1.9.1

^ permalink raw reply related	[flat|nested] 3+ messages in thread

* Re: [PATCH v2] crypto: ctr - avoid VLA use
  2018-03-15 11:18 [PATCH v2] crypto: ctr - avoid VLA use Salvatore Mesoraca
@ 2018-03-23 15:36 ` Herbert Xu
  2018-03-24 13:21   ` Salvatore Mesoraca
  0 siblings, 1 reply; 3+ messages in thread
From: Herbert Xu @ 2018-03-23 15:36 UTC (permalink / raw)
  To: Salvatore Mesoraca
  Cc: linux-kernel, kernel-hardening, linux-crypto, David S. Miller,
	Kees Cook, Eric Biggers

On Thu, Mar 15, 2018 at 12:18:58PM +0100, Salvatore Mesoraca wrote:
>
> +#define MAX_BLOCKSIZE 16
> +
> +#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
> +#define MAX_ALIGNMASK 15
> +#else
> +#define MAX_ALIGNMASK 0
> +#endif
> +

Hmm, this won't work.  Just because you have efficient unaligned
access in general doesn't mean that every implementation can live
with unaligned access.  In particular, on x86 there are quite a
few implementations that require alignment or they will fault.

So please just make it 15 unconditionally.

Thanks,
-- 
Email: Herbert Xu <herbert@gondor.apana.org.au>
Home Page: http://gondor.apana.org.au/~herbert/
PGP Key: http://gondor.apana.org.au/~herbert/pubkey.txt

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [PATCH v2] crypto: ctr - avoid VLA use
  2018-03-23 15:36 ` Herbert Xu
@ 2018-03-24 13:21   ` Salvatore Mesoraca
  0 siblings, 0 replies; 3+ messages in thread
From: Salvatore Mesoraca @ 2018-03-24 13:21 UTC (permalink / raw)
  To: Herbert Xu
  Cc: linux-kernel, Kernel Hardening, linux-crypto, David S. Miller,
	Kees Cook, Eric Biggers

2018-03-23 16:36 GMT+01:00 Herbert Xu <herbert@gondor.apana.org.au>:
> On Thu, Mar 15, 2018 at 12:18:58PM +0100, Salvatore Mesoraca wrote:
>>
>> +#define MAX_BLOCKSIZE 16
>> +
>> +#ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
>> +#define MAX_ALIGNMASK 15
>> +#else
>> +#define MAX_ALIGNMASK 0
>> +#endif
>> +
>
> Hmm, this won't work.  Just because you have efficient unaligned
> access in general doesn't mean that every implementation can live
> with unaligned access.  In particular, on x86 there are quite a
> few implementations that require alignment or they will fault.
>
> So please just make it 15 unconditionally.

Oh, thank you for pointing it out. I'll fix this in v3.

Salvatore

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2018-03-24 13:21 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2018-03-15 11:18 [PATCH v2] crypto: ctr - avoid VLA use Salvatore Mesoraca
2018-03-23 15:36 ` Herbert Xu
2018-03-24 13:21   ` Salvatore Mesoraca

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).