mirror of
https://github.com/torvalds/linux.git
synced 2025-12-07 20:06:24 +00:00
Consolidate the ChaCha code into a single module (excluding chacha-block-generic.c which remains always built-in for random.c), similar to various other algorithms: - Each arch now provides a header file lib/crypto/$(SRCARCH)/chacha.h, replacing lib/crypto/$(SRCARCH)/chacha*.c. The header defines chacha_crypt_arch() and hchacha_block_arch(). It is included by lib/crypto/chacha.c, and thus the code gets built into the single libchacha module, with improved inlining in some cases. - Whether arch-optimized ChaCha is buildable is now controlled centrally by lib/crypto/Kconfig instead of by lib/crypto/$(SRCARCH)/Kconfig. The conditions for enabling it remain the same as before, and it remains enabled by default. - Any additional arch-specific translation units for the optimized ChaCha code, such as assembly files, are now compiled by lib/crypto/Makefile instead of lib/crypto/$(SRCARCH)/Makefile. This removes the last use for the Makefile and Kconfig files in the arm64, mips, powerpc, riscv, and s390 subdirectories of lib/crypto/. So also remove those files and the references to them. Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Link: https://lore.kernel.org/r/20250827151131.27733-7-ebiggers@kernel.org Signed-off-by: Eric Biggers <ebiggers@kernel.org>
77 lines
1.7 KiB
C
77 lines
1.7 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* ChaCha stream cipher (P10 accelerated)
|
|
*
|
|
* Copyright 2023- IBM Corp. All rights reserved.
|
|
*/
|
|
|
|
#include <crypto/internal/simd.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/cpufeature.h>
|
|
#include <linux/sizes.h>
|
|
#include <asm/simd.h>
|
|
#include <asm/switch_to.h>
|
|
|
|
asmlinkage void chacha_p10le_8x(const struct chacha_state *state, u8 *dst,
|
|
const u8 *src, unsigned int len, int nrounds);
|
|
|
|
static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_p10);
|
|
|
|
static void vsx_begin(void)
|
|
{
|
|
preempt_disable();
|
|
enable_kernel_vsx();
|
|
}
|
|
|
|
static void vsx_end(void)
|
|
{
|
|
disable_kernel_vsx();
|
|
preempt_enable();
|
|
}
|
|
|
|
static void chacha_p10_do_8x(struct chacha_state *state, u8 *dst, const u8 *src,
|
|
unsigned int bytes, int nrounds)
|
|
{
|
|
unsigned int l = bytes & ~0x0FF;
|
|
|
|
if (l > 0) {
|
|
chacha_p10le_8x(state, dst, src, l, nrounds);
|
|
bytes -= l;
|
|
src += l;
|
|
dst += l;
|
|
state->x[12] += l / CHACHA_BLOCK_SIZE;
|
|
}
|
|
|
|
if (bytes > 0)
|
|
chacha_crypt_generic(state, dst, src, bytes, nrounds);
|
|
}
|
|
|
|
#define hchacha_block_arch hchacha_block_generic /* not implemented yet */
|
|
|
|
static void chacha_crypt_arch(struct chacha_state *state, u8 *dst,
|
|
const u8 *src, unsigned int bytes, int nrounds)
|
|
{
|
|
if (!static_branch_likely(&have_p10) || bytes <= CHACHA_BLOCK_SIZE ||
|
|
!crypto_simd_usable())
|
|
return chacha_crypt_generic(state, dst, src, bytes, nrounds);
|
|
|
|
do {
|
|
unsigned int todo = min_t(unsigned int, bytes, SZ_4K);
|
|
|
|
vsx_begin();
|
|
chacha_p10_do_8x(state, dst, src, todo, nrounds);
|
|
vsx_end();
|
|
|
|
bytes -= todo;
|
|
src += todo;
|
|
dst += todo;
|
|
} while (bytes);
|
|
}
|
|
|
|
#define chacha_mod_init_arch chacha_mod_init_arch
|
|
static void chacha_mod_init_arch(void)
|
|
{
|
|
if (cpu_has_feature(CPU_FTR_ARCH_31))
|
|
static_branch_enable(&have_p10);
|
|
}
|