From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.0 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id CA6F3C433E0 for ; Wed, 23 Dec 2020 09:06:15 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 82843224D3 for ; Wed, 23 Dec 2020 09:06:15 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726463AbgLWJGO (ORCPT ); Wed, 23 Dec 2020 04:06:14 -0500 Received: from mail.kernel.org ([198.145.29.99]:59988 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728068AbgLWJGD (ORCPT ); Wed, 23 Dec 2020 04:06:03 -0500 Received: by mail.kernel.org (Postfix) with ESMTPSA id 0FD4A20760 for ; Wed, 23 Dec 2020 09:05:22 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1608714322; bh=P9wkT64INpDRhGhcH68OtjZd84cKNiQw1aLqJnrISnw=; h=References:In-Reply-To:From:Date:Subject:To:Cc:From; b=LTevQSrrBnWxq59X5f74rwAPLtixCd9Jsxg+882qwPSzkd+IWiZSbq0pXHFWEpt76 vEPt4UJLhiBDpkPJVIUJ9kWIDq1GhI906ePnlHT8cfQZW59WyI9kXeVKEjPRzqq/R4 d8iaY8OYYCIQxLbAE2SIuQZnAKDH0Po/KnYmpFgEeu5hTIPyMZ+UqfoNUweb24Id96 w0Gm2GXRKvBNGdikTEVrIqWML+yI7rE0YYCni8ukyv2PW6874fwTPNI/9hBWEJltS0 rP4bXQ5H3iZhhjTYwnJEm5bnHA475H+uMuRIKHZkJL94T5CKjTeCfPPfoLPkmzBw+3 DLbp9yZy4x4DQ== Received: by mail-oi1-f175.google.com with SMTP id x13so17645787oic.5 for ; Wed, 23 Dec 2020 01:05:22 -0800 (PST) X-Gm-Message-State: AOAM5338mI9S2KZB0dqz5+ADp7Ee2ropx9oNkQ3YBqMsJONGe+J4ROPg nUIi2EZkwRFx/8EbC2ZdZdya0a/GoBo7TfuxSOE= X-Google-Smtp-Source: ABdhPJzTVMdROfoNZLogJFvWAulUUnAKa7OhLDK5Xwg9vprAFPLfTjDfm3BDYUbZY21mEHo8KwLov2p3Bb2IlBFsJZY= X-Received: by 2002:aca:dd03:: with SMTP id u3mr17294306oig.47.1608714321262; Wed, 23 Dec 2020 01:05:21 -0800 (PST) MIME-Version: 1.0 References: <20201223081003.373663-1-ebiggers@kernel.org> <20201223081003.373663-5-ebiggers@kernel.org> In-Reply-To: <20201223081003.373663-5-ebiggers@kernel.org> From: Ard Biesheuvel Date: Wed, 23 Dec 2020 10:05:10 +0100 X-Gmail-Original-Message-ID: Message-ID: Subject: Re: [PATCH v3 04/14] crypto: blake2s - move update and final logic to internal/blake2s.h To: Eric Biggers Cc: Linux Crypto Mailing List , Linux ARM , Herbert Xu , David Sterba , "Jason A . Donenfeld" , Paul Crowley Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-crypto@vger.kernel.org On Wed, 23 Dec 2020 at 09:12, Eric Biggers wrote: > > From: Eric Biggers > > Move most of blake2s_update() and blake2s_final() into new inline > functions __blake2s_update() and __blake2s_final() in > include/crypto/internal/blake2s.h so that this logic can be shared by > the shash helper functions. This will avoid duplicating this logic > between the library and shash implementations. > > Signed-off-by: Eric Biggers Acked-by: Ard Biesheuvel > --- > include/crypto/internal/blake2s.h | 41 ++++++++++++++++++++++++++ > lib/crypto/blake2s.c | 48 ++++++------------------------- > 2 files changed, 49 insertions(+), 40 deletions(-) > > diff --git a/include/crypto/internal/blake2s.h b/include/crypto/internal/blake2s.h > index 6e376ae6b6b58..42deba4b8ceef 100644 > --- a/include/crypto/internal/blake2s.h > +++ b/include/crypto/internal/blake2s.h > @@ -4,6 +4,7 @@ > #define BLAKE2S_INTERNAL_H > > #include > +#include > > struct blake2s_tfm_ctx { > u8 key[BLAKE2S_KEY_SIZE]; > @@ -23,4 +24,44 @@ static inline void blake2s_set_lastblock(struct blake2s_state *state) > state->f[0] = -1; > } > > +typedef void (*blake2s_compress_t)(struct blake2s_state *state, > + const u8 *block, size_t nblocks, u32 inc); > + > +static inline void __blake2s_update(struct blake2s_state *state, > + const u8 *in, size_t inlen, > + blake2s_compress_t compress) > +{ > + const size_t fill = BLAKE2S_BLOCK_SIZE - state->buflen; > + > + if (unlikely(!inlen)) > + return; > + if (inlen > fill) { > + memcpy(state->buf + state->buflen, in, fill); > + (*compress)(state, state->buf, 1, BLAKE2S_BLOCK_SIZE); > + state->buflen = 0; > + in += fill; > + inlen -= fill; > + } > + if (inlen > BLAKE2S_BLOCK_SIZE) { > + const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2S_BLOCK_SIZE); > + /* Hash one less (full) block than strictly possible */ > + (*compress)(state, in, nblocks - 1, BLAKE2S_BLOCK_SIZE); > + in += BLAKE2S_BLOCK_SIZE * (nblocks - 1); > + inlen -= BLAKE2S_BLOCK_SIZE * (nblocks - 1); > + } > + memcpy(state->buf + state->buflen, in, inlen); > + state->buflen += inlen; > +} > + > +static inline void __blake2s_final(struct blake2s_state *state, u8 *out, > + blake2s_compress_t compress) > +{ > + blake2s_set_lastblock(state); > + memset(state->buf + state->buflen, 0, > + BLAKE2S_BLOCK_SIZE - state->buflen); /* Padding */ > + (*compress)(state, state->buf, 1, state->buflen); > + cpu_to_le32_array(state->h, ARRAY_SIZE(state->h)); > + memcpy(out, state->h, state->outlen); > +} > + > #endif /* BLAKE2S_INTERNAL_H */ > diff --git a/lib/crypto/blake2s.c b/lib/crypto/blake2s.c > index 6a4b6b78d630f..c64ac8bfb6a97 100644 > --- a/lib/crypto/blake2s.c > +++ b/lib/crypto/blake2s.c > @@ -15,55 +15,23 @@ > #include > #include > #include > -#include > + > +#if IS_ENABLED(CONFIG_CRYPTO_ARCH_HAVE_LIB_BLAKE2S) > +# define blake2s_compress blake2s_compress_arch > +#else > +# define blake2s_compress blake2s_compress_generic > +#endif > > void blake2s_update(struct blake2s_state *state, const u8 *in, size_t inlen) > { > - const size_t fill = BLAKE2S_BLOCK_SIZE - state->buflen; > - > - if (unlikely(!inlen)) > - return; > - if (inlen > fill) { > - memcpy(state->buf + state->buflen, in, fill); > - if (IS_ENABLED(CONFIG_CRYPTO_ARCH_HAVE_LIB_BLAKE2S)) > - blake2s_compress_arch(state, state->buf, 1, > - BLAKE2S_BLOCK_SIZE); > - else > - blake2s_compress_generic(state, state->buf, 1, > - BLAKE2S_BLOCK_SIZE); > - state->buflen = 0; > - in += fill; > - inlen -= fill; > - } > - if (inlen > BLAKE2S_BLOCK_SIZE) { > - const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2S_BLOCK_SIZE); > - /* Hash one less (full) block than strictly possible */ > - if (IS_ENABLED(CONFIG_CRYPTO_ARCH_HAVE_LIB_BLAKE2S)) > - blake2s_compress_arch(state, in, nblocks - 1, > - BLAKE2S_BLOCK_SIZE); > - else > - blake2s_compress_generic(state, in, nblocks - 1, > - BLAKE2S_BLOCK_SIZE); > - in += BLAKE2S_BLOCK_SIZE * (nblocks - 1); > - inlen -= BLAKE2S_BLOCK_SIZE * (nblocks - 1); > - } > - memcpy(state->buf + state->buflen, in, inlen); > - state->buflen += inlen; > + __blake2s_update(state, in, inlen, blake2s_compress); > } > EXPORT_SYMBOL(blake2s_update); > > void blake2s_final(struct blake2s_state *state, u8 *out) > { > WARN_ON(IS_ENABLED(DEBUG) && !out); > - blake2s_set_lastblock(state); > - memset(state->buf + state->buflen, 0, > - BLAKE2S_BLOCK_SIZE - state->buflen); /* Padding */ > - if (IS_ENABLED(CONFIG_CRYPTO_ARCH_HAVE_LIB_BLAKE2S)) > - blake2s_compress_arch(state, state->buf, 1, state->buflen); > - else > - blake2s_compress_generic(state, state->buf, 1, state->buflen); > - cpu_to_le32_array(state->h, ARRAY_SIZE(state->h)); > - memcpy(out, state->h, state->outlen); > + __blake2s_final(state, out, blake2s_compress); > memzero_explicit(state, sizeof(*state)); > } > EXPORT_SYMBOL(blake2s_final); > -- > 2.29.2 >