From: Peter Zijlstra <peterz@infradead.org>
To: torvalds@linux-foundation.org
Cc: corbet@lwn.net, will@kernel.org, peterz@infradead.org,
boqun.feng@gmail.com, mark.rutland@arm.com,
catalin.marinas@arm.com, dennis@kernel.org, tj@kernel.org,
cl@linux.com, hca@linux.ibm.com, gor@linux.ibm.com,
agordeev@linux.ibm.com, borntraeger@linux.ibm.com,
svens@linux.ibm.com, Herbert Xu <herbert@gondor.apana.org.au>,
davem@davemloft.net, tglx@linutronix.de, mingo@redhat.com,
bp@alien8.de, dave.hansen@linux.intel.com, x86@kernel.org,
hpa@zytor.com, joro@8bytes.org, suravee.suthikulpanit@amd.com,
robin.murphy@arm.com, dwmw2@infradead.org,
baolu.lu@linux.intel.com, Arnd Bergmann <arnd@arndb.de>,
penberg@kernel.org, rientjes@google.com, iamjoonsoo.kim@lge.com,
Andrew Morton <akpm@linux-foundation.org>,
vbabka@suse.cz, roman.gushchin@linux.dev, 42.hyeyoo@gmail.com,
linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org,
linux-mm@kvack.org, linux-s390@vger.kernel.org,
linux-crypto@vger.kernel.org, iommu@lists.linux.dev,
linux-arch@vger.kernel.org
Subject: [RFC][PATCH 01/12] crypto: Remove u128 usage
Date: Mon, 19 Dec 2022 16:35:26 +0100 [thread overview]
Message-ID: <20221219154118.889543494@infradead.org> (raw)
In-Reply-To: 20221219153525.632521981@infradead.org
As seems to be the common (majority) usage in crypto, use __uint128_t
instead of u128.
This frees up u128 for definition in linux/types.h.
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
---
lib/crypto/curve25519-hacl64.c | 142 ++++++++++++++++++++---------------------
lib/crypto/poly1305-donna64.c | 22 ++----
2 files changed, 80 insertions(+), 84 deletions(-)
--- a/lib/crypto/curve25519-hacl64.c
+++ b/lib/crypto/curve25519-hacl64.c
@@ -14,8 +14,6 @@
#include <crypto/curve25519.h>
#include <linux/string.h>
-typedef __uint128_t u128;
-
static __always_inline u64 u64_eq_mask(u64 a, u64 b)
{
u64 x = a ^ b;
@@ -50,77 +48,77 @@ static __always_inline void modulo_carry
b[0] = b0_;
}
-static __always_inline void fproduct_copy_from_wide_(u64 *output, u128 *input)
+static __always_inline void fproduct_copy_from_wide_(u64 *output, __uint128_t *input)
{
{
- u128 xi = input[0];
+ __uint128_t xi = input[0];
output[0] = ((u64)(xi));
}
{
- u128 xi = input[1];
+ __uint128_t xi = input[1];
output[1] = ((u64)(xi));
}
{
- u128 xi = input[2];
+ __uint128_t xi = input[2];
output[2] = ((u64)(xi));
}
{
- u128 xi = input[3];
+ __uint128_t xi = input[3];
output[3] = ((u64)(xi));
}
{
- u128 xi = input[4];
+ __uint128_t xi = input[4];
output[4] = ((u64)(xi));
}
}
static __always_inline void
-fproduct_sum_scalar_multiplication_(u128 *output, u64 *input, u64 s)
+fproduct_sum_scalar_multiplication_(__uint128_t *output, u64 *input, u64 s)
{
- output[0] += (u128)input[0] * s;
- output[1] += (u128)input[1] * s;
- output[2] += (u128)input[2] * s;
- output[3] += (u128)input[3] * s;
- output[4] += (u128)input[4] * s;
+ output[0] += (__uint128_t)input[0] * s;
+ output[1] += (__uint128_t)input[1] * s;
+ output[2] += (__uint128_t)input[2] * s;
+ output[3] += (__uint128_t)input[3] * s;
+ output[4] += (__uint128_t)input[4] * s;
}
-static __always_inline void fproduct_carry_wide_(u128 *tmp)
+static __always_inline void fproduct_carry_wide_(__uint128_t *tmp)
{
{
u32 ctr = 0;
- u128 tctr = tmp[ctr];
- u128 tctrp1 = tmp[ctr + 1];
+ __uint128_t tctr = tmp[ctr];
+ __uint128_t tctrp1 = tmp[ctr + 1];
u64 r0 = ((u64)(tctr)) & 0x7ffffffffffffLLU;
- u128 c = ((tctr) >> (51));
- tmp[ctr] = ((u128)(r0));
+ __uint128_t c = ((tctr) >> (51));
+ tmp[ctr] = ((__uint128_t)(r0));
tmp[ctr + 1] = ((tctrp1) + (c));
}
{
u32 ctr = 1;
- u128 tctr = tmp[ctr];
- u128 tctrp1 = tmp[ctr + 1];
+ __uint128_t tctr = tmp[ctr];
+ __uint128_t tctrp1 = tmp[ctr + 1];
u64 r0 = ((u64)(tctr)) & 0x7ffffffffffffLLU;
- u128 c = ((tctr) >> (51));
- tmp[ctr] = ((u128)(r0));
+ __uint128_t c = ((tctr) >> (51));
+ tmp[ctr] = ((__uint128_t)(r0));
tmp[ctr + 1] = ((tctrp1) + (c));
}
{
u32 ctr = 2;
- u128 tctr = tmp[ctr];
- u128 tctrp1 = tmp[ctr + 1];
+ __uint128_t tctr = tmp[ctr];
+ __uint128_t tctrp1 = tmp[ctr + 1];
u64 r0 = ((u64)(tctr)) & 0x7ffffffffffffLLU;
- u128 c = ((tctr) >> (51));
- tmp[ctr] = ((u128)(r0));
+ __uint128_t c = ((tctr) >> (51));
+ tmp[ctr] = ((__uint128_t)(r0));
tmp[ctr + 1] = ((tctrp1) + (c));
}
{
u32 ctr = 3;
- u128 tctr = tmp[ctr];
- u128 tctrp1 = tmp[ctr + 1];
+ __uint128_t tctr = tmp[ctr];
+ __uint128_t tctrp1 = tmp[ctr + 1];
u64 r0 = ((u64)(tctr)) & 0x7ffffffffffffLLU;
- u128 c = ((tctr) >> (51));
- tmp[ctr] = ((u128)(r0));
+ __uint128_t c = ((tctr) >> (51));
+ tmp[ctr] = ((__uint128_t)(r0));
tmp[ctr + 1] = ((tctrp1) + (c));
}
}
@@ -154,7 +152,7 @@ static __always_inline void fmul_shift_r
output[0] = 19 * b0;
}
-static __always_inline void fmul_mul_shift_reduce_(u128 *output, u64 *input,
+static __always_inline void fmul_mul_shift_reduce_(__uint128_t *output, u64 *input,
u64 *input21)
{
u32 i;
@@ -188,21 +186,21 @@ static __always_inline void fmul_fmul(u6
{
u64 tmp[5] = { input[0], input[1], input[2], input[3], input[4] };
{
- u128 b4;
- u128 b0;
- u128 b4_;
- u128 b0_;
+ __uint128_t b4;
+ __uint128_t b0;
+ __uint128_t b4_;
+ __uint128_t b0_;
u64 i0;
u64 i1;
u64 i0_;
u64 i1_;
- u128 t[5] = { 0 };
+ __uint128_t t[5] = { 0 };
fmul_mul_shift_reduce_(t, tmp, input21);
fproduct_carry_wide_(t);
b4 = t[4];
b0 = t[0];
- b4_ = ((b4) & (((u128)(0x7ffffffffffffLLU))));
- b0_ = ((b0) + (((u128)(19) * (((u64)(((b4) >> (51))))))));
+ b4_ = ((b4) & (((__uint128_t)(0x7ffffffffffffLLU))));
+ b0_ = ((b0) + (((__uint128_t)(19) * (((u64)(((b4) >> (51))))))));
t[4] = b4_;
t[0] = b0_;
fproduct_copy_from_wide_(output, t);
@@ -215,7 +213,7 @@ static __always_inline void fmul_fmul(u6
}
}
-static __always_inline void fsquare_fsquare__(u128 *tmp, u64 *output)
+static __always_inline void fsquare_fsquare__(__uint128_t *tmp, u64 *output)
{
u64 r0 = output[0];
u64 r1 = output[1];
@@ -227,16 +225,16 @@ static __always_inline void fsquare_fsqu
u64 d2 = r2 * 2 * 19;
u64 d419 = r4 * 19;
u64 d4 = d419 * 2;
- u128 s0 = ((((((u128)(r0) * (r0))) + (((u128)(d4) * (r1))))) +
- (((u128)(d2) * (r3))));
- u128 s1 = ((((((u128)(d0) * (r1))) + (((u128)(d4) * (r2))))) +
- (((u128)(r3 * 19) * (r3))));
- u128 s2 = ((((((u128)(d0) * (r2))) + (((u128)(r1) * (r1))))) +
- (((u128)(d4) * (r3))));
- u128 s3 = ((((((u128)(d0) * (r3))) + (((u128)(d1) * (r2))))) +
- (((u128)(r4) * (d419))));
- u128 s4 = ((((((u128)(d0) * (r4))) + (((u128)(d1) * (r3))))) +
- (((u128)(r2) * (r2))));
+ __uint128_t s0 = ((((((__uint128_t)(r0) * (r0))) + (((__uint128_t)(d4) * (r1))))) +
+ (((__uint128_t)(d2) * (r3))));
+ __uint128_t s1 = ((((((__uint128_t)(d0) * (r1))) + (((__uint128_t)(d4) * (r2))))) +
+ (((__uint128_t)(r3 * 19) * (r3))));
+ __uint128_t s2 = ((((((__uint128_t)(d0) * (r2))) + (((__uint128_t)(r1) * (r1))))) +
+ (((__uint128_t)(d4) * (r3))));
+ __uint128_t s3 = ((((((__uint128_t)(d0) * (r3))) + (((__uint128_t)(d1) * (r2))))) +
+ (((__uint128_t)(r4) * (d419))));
+ __uint128_t s4 = ((((((__uint128_t)(d0) * (r4))) + (((__uint128_t)(d1) * (r3))))) +
+ (((__uint128_t)(r2) * (r2))));
tmp[0] = s0;
tmp[1] = s1;
tmp[2] = s2;
@@ -244,12 +242,12 @@ static __always_inline void fsquare_fsqu
tmp[4] = s4;
}
-static __always_inline void fsquare_fsquare_(u128 *tmp, u64 *output)
+static __always_inline void fsquare_fsquare_(__uint128_t *tmp, u64 *output)
{
- u128 b4;
- u128 b0;
- u128 b4_;
- u128 b0_;
+ __uint128_t b4;
+ __uint128_t b0;
+ __uint128_t b4_;
+ __uint128_t b0_;
u64 i0;
u64 i1;
u64 i0_;
@@ -258,8 +256,8 @@ static __always_inline void fsquare_fsqu
fproduct_carry_wide_(tmp);
b4 = tmp[4];
b0 = tmp[0];
- b4_ = ((b4) & (((u128)(0x7ffffffffffffLLU))));
- b0_ = ((b0) + (((u128)(19) * (((u64)(((b4) >> (51))))))));
+ b4_ = ((b4) & (((__uint128_t)(0x7ffffffffffffLLU))));
+ b0_ = ((b0) + (((__uint128_t)(19) * (((u64)(((b4) >> (51))))))));
tmp[4] = b4_;
tmp[0] = b0_;
fproduct_copy_from_wide_(output, tmp);
@@ -271,7 +269,7 @@ static __always_inline void fsquare_fsqu
output[1] = i1_;
}
-static __always_inline void fsquare_fsquare_times_(u64 *output, u128 *tmp,
+static __always_inline void fsquare_fsquare_times_(u64 *output, __uint128_t *tmp,
u32 count1)
{
u32 i;
@@ -283,7 +281,7 @@ static __always_inline void fsquare_fsqu
static __always_inline void fsquare_fsquare_times(u64 *output, u64 *input,
u32 count1)
{
- u128 t[5];
+ __uint128_t t[5];
memcpy(output, input, 5 * sizeof(*input));
fsquare_fsquare_times_(output, t, count1);
}
@@ -291,7 +289,7 @@ static __always_inline void fsquare_fsqu
static __always_inline void fsquare_fsquare_times_inplace(u64 *output,
u32 count1)
{
- u128 t[5];
+ __uint128_t t[5];
fsquare_fsquare_times_(output, t, count1);
}
@@ -396,36 +394,36 @@ static __always_inline void fdifference(
static __always_inline void fscalar(u64 *output, u64 *b, u64 s)
{
- u128 tmp[5];
- u128 b4;
- u128 b0;
- u128 b4_;
- u128 b0_;
+ __uint128_t tmp[5];
+ __uint128_t b4;
+ __uint128_t b0;
+ __uint128_t b4_;
+ __uint128_t b0_;
{
u64 xi = b[0];
- tmp[0] = ((u128)(xi) * (s));
+ tmp[0] = ((__uint128_t)(xi) * (s));
}
{
u64 xi = b[1];
- tmp[1] = ((u128)(xi) * (s));
+ tmp[1] = ((__uint128_t)(xi) * (s));
}
{
u64 xi = b[2];
- tmp[2] = ((u128)(xi) * (s));
+ tmp[2] = ((__uint128_t)(xi) * (s));
}
{
u64 xi = b[3];
- tmp[3] = ((u128)(xi) * (s));
+ tmp[3] = ((__uint128_t)(xi) * (s));
}
{
u64 xi = b[4];
- tmp[4] = ((u128)(xi) * (s));
+ tmp[4] = ((__uint128_t)(xi) * (s));
}
fproduct_carry_wide_(tmp);
b4 = tmp[4];
b0 = tmp[0];
- b4_ = ((b4) & (((u128)(0x7ffffffffffffLLU))));
- b0_ = ((b0) + (((u128)(19) * (((u64)(((b4) >> (51))))))));
+ b4_ = ((b4) & (((__uint128_t)(0x7ffffffffffffLLU))));
+ b0_ = ((b0) + (((__uint128_t)(19) * (((u64)(((b4) >> (51))))))));
tmp[4] = b4_;
tmp[0] = b0_;
fproduct_copy_from_wide_(output, tmp);
--- a/lib/crypto/poly1305-donna64.c
+++ b/lib/crypto/poly1305-donna64.c
@@ -10,8 +10,6 @@
#include <asm/unaligned.h>
#include <crypto/internal/poly1305.h>
-typedef __uint128_t u128;
-
void poly1305_core_setkey(struct poly1305_core_key *key,
const u8 raw_key[POLY1305_BLOCK_SIZE])
{
@@ -41,7 +39,7 @@ void poly1305_core_blocks(struct poly130
u64 s1, s2;
u64 h0, h1, h2;
u64 c;
- u128 d0, d1, d2, d;
+ __uint128_t d0, d1, d2, d;
if (!nblocks)
return;
@@ -71,20 +69,20 @@ void poly1305_core_blocks(struct poly130
h2 += (((t1 >> 24)) & 0x3ffffffffffULL) | hibit64;
/* h *= r */
- d0 = (u128)h0 * r0;
- d = (u128)h1 * s2;
+ d0 = (__uint128_t)h0 * r0;
+ d = (__uint128_t)h1 * s2;
d0 += d;
- d = (u128)h2 * s1;
+ d = (__uint128_t)h2 * s1;
d0 += d;
- d1 = (u128)h0 * r1;
- d = (u128)h1 * r0;
+ d1 = (__uint128_t)h0 * r1;
+ d = (__uint128_t)h1 * r0;
d1 += d;
- d = (u128)h2 * s2;
+ d = (__uint128_t)h2 * s2;
d1 += d;
- d2 = (u128)h0 * r2;
- d = (u128)h1 * r1;
+ d2 = (__uint128_t)h0 * r2;
+ d = (__uint128_t)h1 * r1;
d2 += d;
- d = (u128)h2 * r0;
+ d = (__uint128_t)h2 * r0;
d2 += d;
/* (partial) h %= p */
next prev parent reply other threads:[~2022-12-19 15:46 UTC|newest]
Thread overview: 57+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-12-19 15:35 [RFC][PATCH 00/12] Introduce cmpxchg128() -- aka. the demise of cmpxchg_double() Peter Zijlstra
2022-12-19 15:35 ` Peter Zijlstra [this message]
2022-12-19 15:56 ` [RFC][PATCH 01/12] crypto: Remove u128 usage Jason A. Donenfeld
2022-12-19 17:00 ` Peter Zijlstra
2022-12-19 17:03 ` Jason A. Donenfeld
2022-12-20 3:50 ` Herbert Xu
2022-12-20 4:11 ` H. Peter Anvin
2022-12-20 4:15 ` Herbert Xu
2022-12-19 15:35 ` [RFC][PATCH 02/12] crypto/ghash-clmulni: Use (struct) be128 Peter Zijlstra
2022-12-20 5:45 ` Eric Biggers
2022-12-19 15:35 ` [RFC][PATCH 03/12] cyrpto/b128ops: Remove struct u128 Peter Zijlstra
2022-12-20 5:52 ` Eric Biggers
2022-12-19 15:35 ` [RFC][PATCH 04/12] types: Introduce [us]128 Peter Zijlstra
2022-12-29 8:30 ` Pavel Machek
2022-12-19 15:35 ` [RFC][PATCH 05/12] arch: Introduce arch_{,try_}_cmpxchg128{,_local}() Peter Zijlstra
2022-12-19 20:07 ` Boqun Feng
2022-12-20 11:08 ` Peter Zijlstra
2022-12-20 14:31 ` Linus Torvalds
2022-12-20 15:09 ` Peter Zijlstra
2023-01-03 13:25 ` Mark Rutland
2023-01-03 14:03 ` Mark Rutland
2023-01-03 16:19 ` Mark Rutland
2023-01-03 16:50 ` Arnd Bergmann
2023-01-04 11:36 ` Mark Rutland
2023-01-04 13:55 ` Mark Rutland
2022-12-22 1:25 ` Boqun Feng
2022-12-22 13:16 ` Peter Zijlstra
2023-01-03 17:12 ` Heiko Carstens
2023-01-09 18:50 ` Mark Rutland
2023-01-12 10:35 ` Peter Zijlstra
2022-12-19 15:35 ` [RFC][PATCH 06/12] instrumentation: Wire up cmpxchg128() Peter Zijlstra
2022-12-19 15:35 ` [RFC][PATCH 07/12] percpu: Wire up cmpxchg128 Peter Zijlstra
2022-12-29 13:36 ` Arnd Bergmann
2023-01-04 12:09 ` Heiko Carstens
2023-01-09 16:29 ` Peter Zijlstra
2022-12-19 15:35 ` [RFC][PATCH 08/12] s390: Replace cmpxchg_double() with cmpxchg128() Peter Zijlstra
2023-01-10 7:23 ` Heiko Carstens
2023-01-10 8:32 ` Peter Zijlstra
2023-01-10 11:27 ` Mark Rutland
2023-01-10 11:46 ` Heiko Carstens
2023-01-12 11:12 ` Alexander Gordeev
2022-12-19 15:35 ` [RFC][PATCH 09/12] x86,amd_iommu: Replace cmpxchg_double() Peter Zijlstra
2022-12-19 16:47 ` Niklas Schnelle
2022-12-28 8:40 ` Vasant Hegde
2022-12-19 15:35 ` [RFC][PATCH 10/12] x86,intel_iommu: " Peter Zijlstra
2022-12-19 15:35 ` [RFC][PATCH 11/12] slub: " Peter Zijlstra
2023-01-03 15:58 ` Vlastimil Babka
2023-01-03 17:16 ` Heiko Carstens
2023-01-03 19:08 ` Linus Torvalds
2023-01-04 12:07 ` Heiko Carstens
2023-01-09 16:28 ` Peter Zijlstra
2023-01-09 22:02 ` Linus Torvalds
2023-01-09 22:22 ` H. Peter Anvin
2023-01-10 2:09 ` H. Peter Anvin
2023-01-10 10:28 ` Peter Zijlstra
2022-12-19 15:35 ` [RFC][PATCH 12/12] arch: Remove cmpxchg_double Peter Zijlstra
2022-12-22 1:21 ` [RFC][PATCH 00/12] Introduce cmpxchg128() -- aka. the demise of cmpxchg_double() Boqun Feng
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20221219154118.889543494@infradead.org \
--to=peterz@infradead.org \
--cc=42.hyeyoo@gmail.com \
--cc=agordeev@linux.ibm.com \
--cc=akpm@linux-foundation.org \
--cc=arnd@arndb.de \
--cc=baolu.lu@linux.intel.com \
--cc=boqun.feng@gmail.com \
--cc=borntraeger@linux.ibm.com \
--cc=bp@alien8.de \
--cc=catalin.marinas@arm.com \
--cc=cl@linux.com \
--cc=corbet@lwn.net \
--cc=dave.hansen@linux.intel.com \
--cc=davem@davemloft.net \
--cc=dennis@kernel.org \
--cc=dwmw2@infradead.org \
--cc=gor@linux.ibm.com \
--cc=hca@linux.ibm.com \
--cc=herbert@gondor.apana.org.au \
--cc=hpa@zytor.com \
--cc=iamjoonsoo.kim@lge.com \
--cc=iommu@lists.linux.dev \
--cc=joro@8bytes.org \
--cc=linux-arch@vger.kernel.org \
--cc=linux-crypto@vger.kernel.org \
--cc=linux-doc@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-s390@vger.kernel.org \
--cc=mark.rutland@arm.com \
--cc=mingo@redhat.com \
--cc=penberg@kernel.org \
--cc=rientjes@google.com \
--cc=robin.murphy@arm.com \
--cc=roman.gushchin@linux.dev \
--cc=suravee.suthikulpanit@amd.com \
--cc=svens@linux.ibm.com \
--cc=tglx@linutronix.de \
--cc=tj@kernel.org \
--cc=torvalds@linux-foundation.org \
--cc=vbabka@suse.cz \
--cc=will@kernel.org \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).