mirror of
https://github.com/torvalds/linux.git
synced 2026-05-12 16:18:45 +02:00
Since commitaefbab8e77("arm64: fpsimd: Preserve/restore kernel mode NEON at context switch"), kernel-mode NEON sections have been preemptible on arm64. And since commit7dadeaa6e8("sched: Further restrict the preemption modes"), voluntary preemption is no longer supported on arm64 either. Therefore, there's no longer any need to limit the length of kernel-mode NEON sections on arm64. Simplify the SHA-512 code accordingly. Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Link: https://lore.kernel.org/r/20260401000548.133151-8-ebiggers@kernel.org Signed-off-by: Eric Biggers <ebiggers@kernel.org>
36 lines
956 B
C
36 lines
956 B
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* arm64-optimized SHA-512 block function
|
|
*
|
|
* Copyright 2025 Google LLC
|
|
*/
|
|
|
|
#include <asm/simd.h>
|
|
#include <linux/cpufeature.h>
|
|
|
|
static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_sha512_insns);
|
|
|
|
asmlinkage void sha512_block_data_order(struct sha512_block_state *state,
|
|
const u8 *data, size_t nblocks);
|
|
asmlinkage void sha512_ce_transform(struct sha512_block_state *state,
|
|
const u8 *data, size_t nblocks);
|
|
|
|
static void sha512_blocks(struct sha512_block_state *state,
|
|
const u8 *data, size_t nblocks)
|
|
{
|
|
if (static_branch_likely(&have_sha512_insns) &&
|
|
likely(may_use_simd())) {
|
|
scoped_ksimd()
|
|
sha512_ce_transform(state, data, nblocks);
|
|
} else {
|
|
sha512_block_data_order(state, data, nblocks);
|
|
}
|
|
}
|
|
|
|
#define sha512_mod_init_arch sha512_mod_init_arch
|
|
static void sha512_mod_init_arch(void)
|
|
{
|
|
if (cpu_have_named_feature(SHA512))
|
|
static_branch_enable(&have_sha512_insns);
|
|
}
|