mirror of
https://github.com/torvalds/linux.git
synced 2025-11-03 01:59:51 +02:00
As was done with the other algorithms, reorganize the BLAKE2s code so
that the generic implementation and the arch-specific "glue" code is
consolidated into a single translation unit, so that the compiler will
inline the functions and automatically decide whether to include the
generic code in the resulting binary or not.
Similarly, also consolidate the build rules into
lib/crypto/{Makefile,Kconfig}. This removes the last uses of
lib/crypto/{arm,x86}/{Makefile,Kconfig}, so remove those too.
Don't keep the !KMSAN dependency. It was needed only for other
algorithms such as ChaCha that initialize memory from assembly code.
Reviewed-by: Ard Biesheuvel <ardb@kernel.org>
Link: https://lore.kernel.org/r/20250827151131.27733-12-ebiggers@kernel.org
Signed-off-by: Eric Biggers <ebiggers@kernel.org>
165 lines
4.3 KiB
C
165 lines
4.3 KiB
C
// SPDX-License-Identifier: GPL-2.0 OR MIT
|
|
/*
|
|
* Copyright (C) 2015-2019 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved.
|
|
*
|
|
* This is an implementation of the BLAKE2s hash and PRF functions.
|
|
*
|
|
* Information: https://blake2.net/
|
|
*
|
|
*/
|
|
|
|
#include <crypto/blake2s.h>
|
|
#include <linux/bug.h>
|
|
#include <linux/export.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
|
|
static const u8 blake2s_sigma[10][16] = {
|
|
{ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 },
|
|
{ 14, 10, 4, 8, 9, 15, 13, 6, 1, 12, 0, 2, 11, 7, 5, 3 },
|
|
{ 11, 8, 12, 0, 5, 2, 15, 13, 10, 14, 3, 6, 7, 1, 9, 4 },
|
|
{ 7, 9, 3, 1, 13, 12, 11, 14, 2, 6, 5, 10, 4, 0, 15, 8 },
|
|
{ 9, 0, 5, 7, 2, 4, 10, 15, 14, 1, 11, 12, 6, 8, 3, 13 },
|
|
{ 2, 12, 6, 10, 0, 11, 8, 3, 4, 13, 7, 5, 15, 14, 1, 9 },
|
|
{ 12, 5, 1, 15, 14, 13, 4, 10, 0, 7, 6, 3, 9, 2, 8, 11 },
|
|
{ 13, 11, 7, 14, 12, 1, 3, 9, 5, 0, 15, 4, 8, 6, 2, 10 },
|
|
{ 6, 15, 14, 9, 11, 3, 0, 8, 12, 2, 13, 7, 1, 4, 10, 5 },
|
|
{ 10, 2, 8, 4, 7, 6, 1, 5, 15, 11, 9, 14, 3, 12, 13, 0 },
|
|
};
|
|
|
|
static inline void blake2s_increment_counter(struct blake2s_state *state,
|
|
const u32 inc)
|
|
{
|
|
state->t[0] += inc;
|
|
state->t[1] += (state->t[0] < inc);
|
|
}
|
|
|
|
static void __maybe_unused
|
|
blake2s_compress_generic(struct blake2s_state *state, const u8 *block,
|
|
size_t nblocks, const u32 inc)
|
|
{
|
|
u32 m[16];
|
|
u32 v[16];
|
|
int i;
|
|
|
|
WARN_ON(IS_ENABLED(DEBUG) &&
|
|
(nblocks > 1 && inc != BLAKE2S_BLOCK_SIZE));
|
|
|
|
while (nblocks > 0) {
|
|
blake2s_increment_counter(state, inc);
|
|
memcpy(m, block, BLAKE2S_BLOCK_SIZE);
|
|
le32_to_cpu_array(m, ARRAY_SIZE(m));
|
|
memcpy(v, state->h, 32);
|
|
v[ 8] = BLAKE2S_IV0;
|
|
v[ 9] = BLAKE2S_IV1;
|
|
v[10] = BLAKE2S_IV2;
|
|
v[11] = BLAKE2S_IV3;
|
|
v[12] = BLAKE2S_IV4 ^ state->t[0];
|
|
v[13] = BLAKE2S_IV5 ^ state->t[1];
|
|
v[14] = BLAKE2S_IV6 ^ state->f[0];
|
|
v[15] = BLAKE2S_IV7 ^ state->f[1];
|
|
|
|
#define G(r, i, a, b, c, d) do { \
|
|
a += b + m[blake2s_sigma[r][2 * i + 0]]; \
|
|
d = ror32(d ^ a, 16); \
|
|
c += d; \
|
|
b = ror32(b ^ c, 12); \
|
|
a += b + m[blake2s_sigma[r][2 * i + 1]]; \
|
|
d = ror32(d ^ a, 8); \
|
|
c += d; \
|
|
b = ror32(b ^ c, 7); \
|
|
} while (0)
|
|
|
|
#define ROUND(r) do { \
|
|
G(r, 0, v[0], v[ 4], v[ 8], v[12]); \
|
|
G(r, 1, v[1], v[ 5], v[ 9], v[13]); \
|
|
G(r, 2, v[2], v[ 6], v[10], v[14]); \
|
|
G(r, 3, v[3], v[ 7], v[11], v[15]); \
|
|
G(r, 4, v[0], v[ 5], v[10], v[15]); \
|
|
G(r, 5, v[1], v[ 6], v[11], v[12]); \
|
|
G(r, 6, v[2], v[ 7], v[ 8], v[13]); \
|
|
G(r, 7, v[3], v[ 4], v[ 9], v[14]); \
|
|
} while (0)
|
|
ROUND(0);
|
|
ROUND(1);
|
|
ROUND(2);
|
|
ROUND(3);
|
|
ROUND(4);
|
|
ROUND(5);
|
|
ROUND(6);
|
|
ROUND(7);
|
|
ROUND(8);
|
|
ROUND(9);
|
|
|
|
#undef G
|
|
#undef ROUND
|
|
|
|
for (i = 0; i < 8; ++i)
|
|
state->h[i] ^= v[i] ^ v[i + 8];
|
|
|
|
block += BLAKE2S_BLOCK_SIZE;
|
|
--nblocks;
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_CRYPTO_LIB_BLAKE2S_ARCH
|
|
#include "blake2s.h" /* $(SRCARCH)/blake2s.h */
|
|
#else
|
|
#define blake2s_compress blake2s_compress_generic
|
|
#endif
|
|
|
|
static inline void blake2s_set_lastblock(struct blake2s_state *state)
|
|
{
|
|
state->f[0] = -1;
|
|
}
|
|
|
|
void blake2s_update(struct blake2s_state *state, const u8 *in, size_t inlen)
|
|
{
|
|
const size_t fill = BLAKE2S_BLOCK_SIZE - state->buflen;
|
|
|
|
if (unlikely(!inlen))
|
|
return;
|
|
if (inlen > fill) {
|
|
memcpy(state->buf + state->buflen, in, fill);
|
|
blake2s_compress(state, state->buf, 1, BLAKE2S_BLOCK_SIZE);
|
|
state->buflen = 0;
|
|
in += fill;
|
|
inlen -= fill;
|
|
}
|
|
if (inlen > BLAKE2S_BLOCK_SIZE) {
|
|
const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2S_BLOCK_SIZE);
|
|
blake2s_compress(state, in, nblocks - 1, BLAKE2S_BLOCK_SIZE);
|
|
in += BLAKE2S_BLOCK_SIZE * (nblocks - 1);
|
|
inlen -= BLAKE2S_BLOCK_SIZE * (nblocks - 1);
|
|
}
|
|
memcpy(state->buf + state->buflen, in, inlen);
|
|
state->buflen += inlen;
|
|
}
|
|
EXPORT_SYMBOL(blake2s_update);
|
|
|
|
void blake2s_final(struct blake2s_state *state, u8 *out)
|
|
{
|
|
WARN_ON(IS_ENABLED(DEBUG) && !out);
|
|
blake2s_set_lastblock(state);
|
|
memset(state->buf + state->buflen, 0,
|
|
BLAKE2S_BLOCK_SIZE - state->buflen); /* Padding */
|
|
blake2s_compress(state, state->buf, 1, state->buflen);
|
|
cpu_to_le32_array(state->h, ARRAY_SIZE(state->h));
|
|
memcpy(out, state->h, state->outlen);
|
|
memzero_explicit(state, sizeof(*state));
|
|
}
|
|
EXPORT_SYMBOL(blake2s_final);
|
|
|
|
#ifdef blake2s_mod_init_arch
|
|
static int __init blake2s_mod_init(void)
|
|
{
|
|
blake2s_mod_init_arch();
|
|
return 0;
|
|
}
|
|
subsys_initcall(blake2s_mod_init);
|
|
#endif
|
|
|
|
MODULE_DESCRIPTION("BLAKE2s hash function");
|
|
MODULE_AUTHOR("Jason A. Donenfeld <Jason@zx2c4.com>");
|