Ard Biesheuvel 565def1542 crypto: arm64/aes-ccm - Cache round keys and unroll AES loops
The CCM code as originally written attempted to use as few NEON
registers as possible, to avoid having to eagerly preserve/restore the
entire NEON register file at every call to kernel_neon_begin/end. At
that time, this API took a number of NEON registers as a parameter, and
only preserved that many registers.

Today, the NEON register file is restored lazily, and the old API is
long gone. This means we can use as many NEON registers as we can make
meaningful use of, which means in the AES case that we can keep all
round keys in registers rather than reloading each of them for each AES
block processed.

On Cortex-A53, this results in a speedup of more than 50%. (From 4
cycles per byte to 2.6 cycles per byte)

Signed-off-by: Ard Biesheuvel <ardb@kernel.org>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
2024-01-26 16:39:32 +08:00
..
2024-01-18 17:25:39 -08:00
2024-01-18 15:16:57 -08:00
2024-01-19 09:10:23 -08:00
2024-01-19 09:10:23 -08:00
2024-01-10 18:13:44 -08:00
2024-01-19 13:30:49 -08:00
2024-01-18 15:01:28 -08:00
2024-01-17 11:27:23 -08:00
2024-01-19 09:10:23 -08:00
2024-01-18 15:01:28 -08:00
2023-09-05 10:09:31 -07:00
2024-01-18 15:01:28 -08:00
2024-01-19 09:10:23 -08:00
2024-01-20 09:24:06 -08:00
2024-01-19 08:08:53 -06:00
2024-01-18 11:37:24 -08:00
2024-01-18 17:25:39 -08:00
2024-01-18 15:01:28 -08:00
2024-01-18 15:16:57 -08:00