Even though the kernel's implementations of AES-XTS were updated to implement ciphertext stealing and can operate on inputs of any size larger than or equal to the AES block size, this feature is rarely used in practice. In fact, in the kernel, AES-XTS is only used to operate on 4096 or 512 byte blocks, which means that not only the ciphertext stealing is effectively dead code, the logic in the bit sliced NEON implementation to deal with fewer than 8 blocks at a time is also never used. Since the bit-sliced NEON driver already depends on the plain NEON version, which is slower but can operate on smaller data quantities more straightforwardly, let's fallback to the plain NEON implementation of XTS for any residual inputs that are not multiples of 128 bytes. This allows us to remove a lot of complicated logic that rarely gets exercised in practice. Signed-off-by: Ard Biesheuvel <ardb@kernel.org> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
868 lines
22 KiB
ArmAsm
868 lines
22 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Bit sliced AES using NEON instructions
|
|
*
|
|
* Copyright (C) 2016 Linaro Ltd <ard.biesheuvel@linaro.org>
|
|
*/
|
|
|
|
/*
|
|
* The algorithm implemented here is described in detail by the paper
|
|
* 'Faster and Timing-Attack Resistant AES-GCM' by Emilia Kaesper and
|
|
* Peter Schwabe (https://eprint.iacr.org/2009/129.pdf)
|
|
*
|
|
* This implementation is based primarily on the OpenSSL implementation
|
|
* for 32-bit ARM written by Andy Polyakov <appro@openssl.org>
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
#include <asm/assembler.h>
|
|
|
|
.text
|
|
|
|
rounds .req x11
|
|
bskey .req x12
|
|
|
|
.macro in_bs_ch, b0, b1, b2, b3, b4, b5, b6, b7
|
|
eor \b2, \b2, \b1
|
|
eor \b5, \b5, \b6
|
|
eor \b3, \b3, \b0
|
|
eor \b6, \b6, \b2
|
|
eor \b5, \b5, \b0
|
|
eor \b6, \b6, \b3
|
|
eor \b3, \b3, \b7
|
|
eor \b7, \b7, \b5
|
|
eor \b3, \b3, \b4
|
|
eor \b4, \b4, \b5
|
|
eor \b2, \b2, \b7
|
|
eor \b3, \b3, \b1
|
|
eor \b1, \b1, \b5
|
|
.endm
|
|
|
|
.macro out_bs_ch, b0, b1, b2, b3, b4, b5, b6, b7
|
|
eor \b0, \b0, \b6
|
|
eor \b1, \b1, \b4
|
|
eor \b4, \b4, \b6
|
|
eor \b2, \b2, \b0
|
|
eor \b6, \b6, \b1
|
|
eor \b1, \b1, \b5
|
|
eor \b5, \b5, \b3
|
|
eor \b3, \b3, \b7
|
|
eor \b7, \b7, \b5
|
|
eor \b2, \b2, \b5
|
|
eor \b4, \b4, \b7
|
|
.endm
|
|
|
|
.macro inv_in_bs_ch, b6, b1, b2, b4, b7, b0, b3, b5
|
|
eor \b1, \b1, \b7
|
|
eor \b4, \b4, \b7
|
|
eor \b7, \b7, \b5
|
|
eor \b1, \b1, \b3
|
|
eor \b2, \b2, \b5
|
|
eor \b3, \b3, \b7
|
|
eor \b6, \b6, \b1
|
|
eor \b2, \b2, \b0
|
|
eor \b5, \b5, \b3
|
|
eor \b4, \b4, \b6
|
|
eor \b0, \b0, \b6
|
|
eor \b1, \b1, \b4
|
|
.endm
|
|
|
|
.macro inv_out_bs_ch, b6, b5, b0, b3, b7, b1, b4, b2
|
|
eor \b1, \b1, \b5
|
|
eor \b2, \b2, \b7
|
|
eor \b3, \b3, \b1
|
|
eor \b4, \b4, \b5
|
|
eor \b7, \b7, \b5
|
|
eor \b3, \b3, \b4
|
|
eor \b5, \b5, \b0
|
|
eor \b3, \b3, \b7
|
|
eor \b6, \b6, \b2
|
|
eor \b2, \b2, \b1
|
|
eor \b6, \b6, \b3
|
|
eor \b3, \b3, \b0
|
|
eor \b5, \b5, \b6
|
|
.endm
|
|
|
|
.macro mul_gf4, x0, x1, y0, y1, t0, t1
|
|
eor \t0, \y0, \y1
|
|
and \t0, \t0, \x0
|
|
eor \x0, \x0, \x1
|
|
and \t1, \x1, \y0
|
|
and \x0, \x0, \y1
|
|
eor \x1, \t1, \t0
|
|
eor \x0, \x0, \t1
|
|
.endm
|
|
|
|
.macro mul_gf4_n_gf4, x0, x1, y0, y1, t0, x2, x3, y2, y3, t1
|
|
eor \t0, \y0, \y1
|
|
eor \t1, \y2, \y3
|
|
and \t0, \t0, \x0
|
|
and \t1, \t1, \x2
|
|
eor \x0, \x0, \x1
|
|
eor \x2, \x2, \x3
|
|
and \x1, \x1, \y0
|
|
and \x3, \x3, \y2
|
|
and \x0, \x0, \y1
|
|
and \x2, \x2, \y3
|
|
eor \x1, \x1, \x0
|
|
eor \x2, \x2, \x3
|
|
eor \x0, \x0, \t0
|
|
eor \x3, \x3, \t1
|
|
.endm
|
|
|
|
.macro mul_gf16_2, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
y0, y1, y2, y3, t0, t1, t2, t3
|
|
eor \t0, \x0, \x2
|
|
eor \t1, \x1, \x3
|
|
mul_gf4 \x0, \x1, \y0, \y1, \t2, \t3
|
|
eor \y0, \y0, \y2
|
|
eor \y1, \y1, \y3
|
|
mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x2, \x3, \y2, \y3, \t2
|
|
eor \x0, \x0, \t0
|
|
eor \x2, \x2, \t0
|
|
eor \x1, \x1, \t1
|
|
eor \x3, \x3, \t1
|
|
eor \t0, \x4, \x6
|
|
eor \t1, \x5, \x7
|
|
mul_gf4_n_gf4 \t0, \t1, \y0, \y1, \t3, \x6, \x7, \y2, \y3, \t2
|
|
eor \y0, \y0, \y2
|
|
eor \y1, \y1, \y3
|
|
mul_gf4 \x4, \x5, \y0, \y1, \t2, \t3
|
|
eor \x4, \x4, \t0
|
|
eor \x6, \x6, \t0
|
|
eor \x5, \x5, \t1
|
|
eor \x7, \x7, \t1
|
|
.endm
|
|
|
|
.macro inv_gf256, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
t0, t1, t2, t3, s0, s1, s2, s3
|
|
eor \t3, \x4, \x6
|
|
eor \t0, \x5, \x7
|
|
eor \t1, \x1, \x3
|
|
eor \s1, \x7, \x6
|
|
eor \s0, \x0, \x2
|
|
eor \s3, \t3, \t0
|
|
orr \t2, \t0, \t1
|
|
and \s2, \t3, \s0
|
|
orr \t3, \t3, \s0
|
|
eor \s0, \s0, \t1
|
|
and \t0, \t0, \t1
|
|
eor \t1, \x3, \x2
|
|
and \s3, \s3, \s0
|
|
and \s1, \s1, \t1
|
|
eor \t1, \x4, \x5
|
|
eor \s0, \x1, \x0
|
|
eor \t3, \t3, \s1
|
|
eor \t2, \t2, \s1
|
|
and \s1, \t1, \s0
|
|
orr \t1, \t1, \s0
|
|
eor \t3, \t3, \s3
|
|
eor \t0, \t0, \s1
|
|
eor \t2, \t2, \s2
|
|
eor \t1, \t1, \s3
|
|
eor \t0, \t0, \s2
|
|
and \s0, \x7, \x3
|
|
eor \t1, \t1, \s2
|
|
and \s1, \x6, \x2
|
|
and \s2, \x5, \x1
|
|
orr \s3, \x4, \x0
|
|
eor \t3, \t3, \s0
|
|
eor \t1, \t1, \s2
|
|
eor \s0, \t0, \s3
|
|
eor \t2, \t2, \s1
|
|
and \s2, \t3, \t1
|
|
eor \s1, \t2, \s2
|
|
eor \s3, \s0, \s2
|
|
bsl \s1, \t1, \s0
|
|
not \t0, \s0
|
|
bsl \s0, \s1, \s3
|
|
bsl \t0, \s1, \s3
|
|
bsl \s3, \t3, \t2
|
|
eor \t3, \t3, \t2
|
|
and \s2, \s0, \s3
|
|
eor \t1, \t1, \t0
|
|
eor \s2, \s2, \t3
|
|
mul_gf16_2 \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \
|
|
\s3, \s2, \s1, \t1, \s0, \t0, \t2, \t3
|
|
.endm
|
|
|
|
.macro sbox, b0, b1, b2, b3, b4, b5, b6, b7, \
|
|
t0, t1, t2, t3, s0, s1, s2, s3
|
|
in_bs_ch \b0\().16b, \b1\().16b, \b2\().16b, \b3\().16b, \
|
|
\b4\().16b, \b5\().16b, \b6\().16b, \b7\().16b
|
|
inv_gf256 \b6\().16b, \b5\().16b, \b0\().16b, \b3\().16b, \
|
|
\b7\().16b, \b1\().16b, \b4\().16b, \b2\().16b, \
|
|
\t0\().16b, \t1\().16b, \t2\().16b, \t3\().16b, \
|
|
\s0\().16b, \s1\().16b, \s2\().16b, \s3\().16b
|
|
out_bs_ch \b7\().16b, \b1\().16b, \b4\().16b, \b2\().16b, \
|
|
\b6\().16b, \b5\().16b, \b0\().16b, \b3\().16b
|
|
.endm
|
|
|
|
.macro inv_sbox, b0, b1, b2, b3, b4, b5, b6, b7, \
|
|
t0, t1, t2, t3, s0, s1, s2, s3
|
|
inv_in_bs_ch \b0\().16b, \b1\().16b, \b2\().16b, \b3\().16b, \
|
|
\b4\().16b, \b5\().16b, \b6\().16b, \b7\().16b
|
|
inv_gf256 \b5\().16b, \b1\().16b, \b2\().16b, \b6\().16b, \
|
|
\b3\().16b, \b7\().16b, \b0\().16b, \b4\().16b, \
|
|
\t0\().16b, \t1\().16b, \t2\().16b, \t3\().16b, \
|
|
\s0\().16b, \s1\().16b, \s2\().16b, \s3\().16b
|
|
inv_out_bs_ch \b3\().16b, \b7\().16b, \b0\().16b, \b4\().16b, \
|
|
\b5\().16b, \b1\().16b, \b2\().16b, \b6\().16b
|
|
.endm
|
|
|
|
.macro enc_next_rk
|
|
ldp q16, q17, [bskey], #128
|
|
ldp q18, q19, [bskey, #-96]
|
|
ldp q20, q21, [bskey, #-64]
|
|
ldp q22, q23, [bskey, #-32]
|
|
.endm
|
|
|
|
.macro dec_next_rk
|
|
ldp q16, q17, [bskey, #-128]!
|
|
ldp q18, q19, [bskey, #32]
|
|
ldp q20, q21, [bskey, #64]
|
|
ldp q22, q23, [bskey, #96]
|
|
.endm
|
|
|
|
.macro add_round_key, x0, x1, x2, x3, x4, x5, x6, x7
|
|
eor \x0\().16b, \x0\().16b, v16.16b
|
|
eor \x1\().16b, \x1\().16b, v17.16b
|
|
eor \x2\().16b, \x2\().16b, v18.16b
|
|
eor \x3\().16b, \x3\().16b, v19.16b
|
|
eor \x4\().16b, \x4\().16b, v20.16b
|
|
eor \x5\().16b, \x5\().16b, v21.16b
|
|
eor \x6\().16b, \x6\().16b, v22.16b
|
|
eor \x7\().16b, \x7\().16b, v23.16b
|
|
.endm
|
|
|
|
.macro shift_rows, x0, x1, x2, x3, x4, x5, x6, x7, mask
|
|
tbl \x0\().16b, {\x0\().16b}, \mask\().16b
|
|
tbl \x1\().16b, {\x1\().16b}, \mask\().16b
|
|
tbl \x2\().16b, {\x2\().16b}, \mask\().16b
|
|
tbl \x3\().16b, {\x3\().16b}, \mask\().16b
|
|
tbl \x4\().16b, {\x4\().16b}, \mask\().16b
|
|
tbl \x5\().16b, {\x5\().16b}, \mask\().16b
|
|
tbl \x6\().16b, {\x6\().16b}, \mask\().16b
|
|
tbl \x7\().16b, {\x7\().16b}, \mask\().16b
|
|
.endm
|
|
|
|
.macro mix_cols, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
t0, t1, t2, t3, t4, t5, t6, t7, inv
|
|
ext \t0\().16b, \x0\().16b, \x0\().16b, #12
|
|
ext \t1\().16b, \x1\().16b, \x1\().16b, #12
|
|
eor \x0\().16b, \x0\().16b, \t0\().16b
|
|
ext \t2\().16b, \x2\().16b, \x2\().16b, #12
|
|
eor \x1\().16b, \x1\().16b, \t1\().16b
|
|
ext \t3\().16b, \x3\().16b, \x3\().16b, #12
|
|
eor \x2\().16b, \x2\().16b, \t2\().16b
|
|
ext \t4\().16b, \x4\().16b, \x4\().16b, #12
|
|
eor \x3\().16b, \x3\().16b, \t3\().16b
|
|
ext \t5\().16b, \x5\().16b, \x5\().16b, #12
|
|
eor \x4\().16b, \x4\().16b, \t4\().16b
|
|
ext \t6\().16b, \x6\().16b, \x6\().16b, #12
|
|
eor \x5\().16b, \x5\().16b, \t5\().16b
|
|
ext \t7\().16b, \x7\().16b, \x7\().16b, #12
|
|
eor \x6\().16b, \x6\().16b, \t6\().16b
|
|
eor \t1\().16b, \t1\().16b, \x0\().16b
|
|
eor \x7\().16b, \x7\().16b, \t7\().16b
|
|
ext \x0\().16b, \x0\().16b, \x0\().16b, #8
|
|
eor \t2\().16b, \t2\().16b, \x1\().16b
|
|
eor \t0\().16b, \t0\().16b, \x7\().16b
|
|
eor \t1\().16b, \t1\().16b, \x7\().16b
|
|
ext \x1\().16b, \x1\().16b, \x1\().16b, #8
|
|
eor \t5\().16b, \t5\().16b, \x4\().16b
|
|
eor \x0\().16b, \x0\().16b, \t0\().16b
|
|
eor \t6\().16b, \t6\().16b, \x5\().16b
|
|
eor \x1\().16b, \x1\().16b, \t1\().16b
|
|
ext \t0\().16b, \x4\().16b, \x4\().16b, #8
|
|
eor \t4\().16b, \t4\().16b, \x3\().16b
|
|
ext \t1\().16b, \x5\().16b, \x5\().16b, #8
|
|
eor \t7\().16b, \t7\().16b, \x6\().16b
|
|
ext \x4\().16b, \x3\().16b, \x3\().16b, #8
|
|
eor \t3\().16b, \t3\().16b, \x2\().16b
|
|
ext \x5\().16b, \x7\().16b, \x7\().16b, #8
|
|
eor \t4\().16b, \t4\().16b, \x7\().16b
|
|
ext \x3\().16b, \x6\().16b, \x6\().16b, #8
|
|
eor \t3\().16b, \t3\().16b, \x7\().16b
|
|
ext \x6\().16b, \x2\().16b, \x2\().16b, #8
|
|
eor \x7\().16b, \t1\().16b, \t5\().16b
|
|
.ifb \inv
|
|
eor \x2\().16b, \t0\().16b, \t4\().16b
|
|
eor \x4\().16b, \x4\().16b, \t3\().16b
|
|
eor \x5\().16b, \x5\().16b, \t7\().16b
|
|
eor \x3\().16b, \x3\().16b, \t6\().16b
|
|
eor \x6\().16b, \x6\().16b, \t2\().16b
|
|
.else
|
|
eor \t3\().16b, \t3\().16b, \x4\().16b
|
|
eor \x5\().16b, \x5\().16b, \t7\().16b
|
|
eor \x2\().16b, \x3\().16b, \t6\().16b
|
|
eor \x3\().16b, \t0\().16b, \t4\().16b
|
|
eor \x4\().16b, \x6\().16b, \t2\().16b
|
|
mov \x6\().16b, \t3\().16b
|
|
.endif
|
|
.endm
|
|
|
|
.macro inv_mix_cols, x0, x1, x2, x3, x4, x5, x6, x7, \
|
|
t0, t1, t2, t3, t4, t5, t6, t7
|
|
ext \t0\().16b, \x0\().16b, \x0\().16b, #8
|
|
ext \t6\().16b, \x6\().16b, \x6\().16b, #8
|
|
ext \t7\().16b, \x7\().16b, \x7\().16b, #8
|
|
eor \t0\().16b, \t0\().16b, \x0\().16b
|
|
ext \t1\().16b, \x1\().16b, \x1\().16b, #8
|
|
eor \t6\().16b, \t6\().16b, \x6\().16b
|
|
ext \t2\().16b, \x2\().16b, \x2\().16b, #8
|
|
eor \t7\().16b, \t7\().16b, \x7\().16b
|
|
ext \t3\().16b, \x3\().16b, \x3\().16b, #8
|
|
eor \t1\().16b, \t1\().16b, \x1\().16b
|
|
ext \t4\().16b, \x4\().16b, \x4\().16b, #8
|
|
eor \t2\().16b, \t2\().16b, \x2\().16b
|
|
ext \t5\().16b, \x5\().16b, \x5\().16b, #8
|
|
eor \t3\().16b, \t3\().16b, \x3\().16b
|
|
eor \t4\().16b, \t4\().16b, \x4\().16b
|
|
eor \t5\().16b, \t5\().16b, \x5\().16b
|
|
eor \x0\().16b, \x0\().16b, \t6\().16b
|
|
eor \x1\().16b, \x1\().16b, \t6\().16b
|
|
eor \x2\().16b, \x2\().16b, \t0\().16b
|
|
eor \x4\().16b, \x4\().16b, \t2\().16b
|
|
eor \x3\().16b, \x3\().16b, \t1\().16b
|
|
eor \x1\().16b, \x1\().16b, \t7\().16b
|
|
eor \x2\().16b, \x2\().16b, \t7\().16b
|
|
eor \x4\().16b, \x4\().16b, \t6\().16b
|
|
eor \x5\().16b, \x5\().16b, \t3\().16b
|
|
eor \x3\().16b, \x3\().16b, \t6\().16b
|
|
eor \x6\().16b, \x6\().16b, \t4\().16b
|
|
eor \x4\().16b, \x4\().16b, \t7\().16b
|
|
eor \x5\().16b, \x5\().16b, \t7\().16b
|
|
eor \x7\().16b, \x7\().16b, \t5\().16b
|
|
mix_cols \x0, \x1, \x2, \x3, \x4, \x5, \x6, \x7, \
|
|
\t0, \t1, \t2, \t3, \t4, \t5, \t6, \t7, 1
|
|
.endm
|
|
|
|
.macro swapmove_2x, a0, b0, a1, b1, n, mask, t0, t1
|
|
ushr \t0\().2d, \b0\().2d, #\n
|
|
ushr \t1\().2d, \b1\().2d, #\n
|
|
eor \t0\().16b, \t0\().16b, \a0\().16b
|
|
eor \t1\().16b, \t1\().16b, \a1\().16b
|
|
and \t0\().16b, \t0\().16b, \mask\().16b
|
|
and \t1\().16b, \t1\().16b, \mask\().16b
|
|
eor \a0\().16b, \a0\().16b, \t0\().16b
|
|
shl \t0\().2d, \t0\().2d, #\n
|
|
eor \a1\().16b, \a1\().16b, \t1\().16b
|
|
shl \t1\().2d, \t1\().2d, #\n
|
|
eor \b0\().16b, \b0\().16b, \t0\().16b
|
|
eor \b1\().16b, \b1\().16b, \t1\().16b
|
|
.endm
|
|
|
|
.macro bitslice, x7, x6, x5, x4, x3, x2, x1, x0, t0, t1, t2, t3
|
|
movi \t0\().16b, #0x55
|
|
movi \t1\().16b, #0x33
|
|
swapmove_2x \x0, \x1, \x2, \x3, 1, \t0, \t2, \t3
|
|
swapmove_2x \x4, \x5, \x6, \x7, 1, \t0, \t2, \t3
|
|
movi \t0\().16b, #0x0f
|
|
swapmove_2x \x0, \x2, \x1, \x3, 2, \t1, \t2, \t3
|
|
swapmove_2x \x4, \x6, \x5, \x7, 2, \t1, \t2, \t3
|
|
swapmove_2x \x0, \x4, \x1, \x5, 4, \t0, \t2, \t3
|
|
swapmove_2x \x2, \x6, \x3, \x7, 4, \t0, \t2, \t3
|
|
.endm
|
|
|
|
|
|
.align 6
|
|
M0: .octa 0x0004080c0105090d02060a0e03070b0f
|
|
|
|
M0SR: .octa 0x0004080c05090d010a0e02060f03070b
|
|
SR: .octa 0x0f0e0d0c0a09080b0504070600030201
|
|
SRM0: .octa 0x01060b0c0207080d0304090e00050a0f
|
|
|
|
M0ISR: .octa 0x0004080c0d0105090a0e0206070b0f03
|
|
ISR: .octa 0x0f0e0d0c080b0a090504070602010003
|
|
ISRM0: .octa 0x0306090c00070a0d01040b0e0205080f
|
|
|
|
/*
|
|
* void aesbs_convert_key(u8 out[], u32 const rk[], int rounds)
|
|
*/
|
|
SYM_FUNC_START(aesbs_convert_key)
|
|
ld1 {v7.4s}, [x1], #16 // load round 0 key
|
|
ld1 {v17.4s}, [x1], #16 // load round 1 key
|
|
|
|
movi v8.16b, #0x01 // bit masks
|
|
movi v9.16b, #0x02
|
|
movi v10.16b, #0x04
|
|
movi v11.16b, #0x08
|
|
movi v12.16b, #0x10
|
|
movi v13.16b, #0x20
|
|
movi v14.16b, #0x40
|
|
movi v15.16b, #0x80
|
|
ldr q16, M0
|
|
|
|
sub x2, x2, #1
|
|
str q7, [x0], #16 // save round 0 key
|
|
|
|
.Lkey_loop:
|
|
tbl v7.16b ,{v17.16b}, v16.16b
|
|
ld1 {v17.4s}, [x1], #16 // load next round key
|
|
|
|
cmtst v0.16b, v7.16b, v8.16b
|
|
cmtst v1.16b, v7.16b, v9.16b
|
|
cmtst v2.16b, v7.16b, v10.16b
|
|
cmtst v3.16b, v7.16b, v11.16b
|
|
cmtst v4.16b, v7.16b, v12.16b
|
|
cmtst v5.16b, v7.16b, v13.16b
|
|
cmtst v6.16b, v7.16b, v14.16b
|
|
cmtst v7.16b, v7.16b, v15.16b
|
|
not v0.16b, v0.16b
|
|
not v1.16b, v1.16b
|
|
not v5.16b, v5.16b
|
|
not v6.16b, v6.16b
|
|
|
|
subs x2, x2, #1
|
|
stp q0, q1, [x0], #128
|
|
stp q2, q3, [x0, #-96]
|
|
stp q4, q5, [x0, #-64]
|
|
stp q6, q7, [x0, #-32]
|
|
b.ne .Lkey_loop
|
|
|
|
movi v7.16b, #0x63 // compose .L63
|
|
eor v17.16b, v17.16b, v7.16b
|
|
str q17, [x0]
|
|
ret
|
|
SYM_FUNC_END(aesbs_convert_key)
|
|
|
|
.align 4
|
|
SYM_FUNC_START_LOCAL(aesbs_encrypt8)
|
|
ldr q9, [bskey], #16 // round 0 key
|
|
ldr q8, M0SR
|
|
ldr q24, SR
|
|
|
|
eor v10.16b, v0.16b, v9.16b // xor with round0 key
|
|
eor v11.16b, v1.16b, v9.16b
|
|
tbl v0.16b, {v10.16b}, v8.16b
|
|
eor v12.16b, v2.16b, v9.16b
|
|
tbl v1.16b, {v11.16b}, v8.16b
|
|
eor v13.16b, v3.16b, v9.16b
|
|
tbl v2.16b, {v12.16b}, v8.16b
|
|
eor v14.16b, v4.16b, v9.16b
|
|
tbl v3.16b, {v13.16b}, v8.16b
|
|
eor v15.16b, v5.16b, v9.16b
|
|
tbl v4.16b, {v14.16b}, v8.16b
|
|
eor v10.16b, v6.16b, v9.16b
|
|
tbl v5.16b, {v15.16b}, v8.16b
|
|
eor v11.16b, v7.16b, v9.16b
|
|
tbl v6.16b, {v10.16b}, v8.16b
|
|
tbl v7.16b, {v11.16b}, v8.16b
|
|
|
|
bitslice v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11
|
|
|
|
sub rounds, rounds, #1
|
|
b .Lenc_sbox
|
|
|
|
.Lenc_loop:
|
|
shift_rows v0, v1, v2, v3, v4, v5, v6, v7, v24
|
|
.Lenc_sbox:
|
|
sbox v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
subs rounds, rounds, #1
|
|
b.cc .Lenc_done
|
|
|
|
enc_next_rk
|
|
|
|
mix_cols v0, v1, v4, v6, v3, v7, v2, v5, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
|
|
add_round_key v0, v1, v2, v3, v4, v5, v6, v7
|
|
|
|
b.ne .Lenc_loop
|
|
ldr q24, SRM0
|
|
b .Lenc_loop
|
|
|
|
.Lenc_done:
|
|
ldr q12, [bskey] // last round key
|
|
|
|
bitslice v0, v1, v4, v6, v3, v7, v2, v5, v8, v9, v10, v11
|
|
|
|
eor v0.16b, v0.16b, v12.16b
|
|
eor v1.16b, v1.16b, v12.16b
|
|
eor v4.16b, v4.16b, v12.16b
|
|
eor v6.16b, v6.16b, v12.16b
|
|
eor v3.16b, v3.16b, v12.16b
|
|
eor v7.16b, v7.16b, v12.16b
|
|
eor v2.16b, v2.16b, v12.16b
|
|
eor v5.16b, v5.16b, v12.16b
|
|
ret
|
|
SYM_FUNC_END(aesbs_encrypt8)
|
|
|
|
.align 4
|
|
SYM_FUNC_START_LOCAL(aesbs_decrypt8)
|
|
lsl x9, rounds, #7
|
|
add bskey, bskey, x9
|
|
|
|
ldr q9, [bskey, #-112]! // round 0 key
|
|
ldr q8, M0ISR
|
|
ldr q24, ISR
|
|
|
|
eor v10.16b, v0.16b, v9.16b // xor with round0 key
|
|
eor v11.16b, v1.16b, v9.16b
|
|
tbl v0.16b, {v10.16b}, v8.16b
|
|
eor v12.16b, v2.16b, v9.16b
|
|
tbl v1.16b, {v11.16b}, v8.16b
|
|
eor v13.16b, v3.16b, v9.16b
|
|
tbl v2.16b, {v12.16b}, v8.16b
|
|
eor v14.16b, v4.16b, v9.16b
|
|
tbl v3.16b, {v13.16b}, v8.16b
|
|
eor v15.16b, v5.16b, v9.16b
|
|
tbl v4.16b, {v14.16b}, v8.16b
|
|
eor v10.16b, v6.16b, v9.16b
|
|
tbl v5.16b, {v15.16b}, v8.16b
|
|
eor v11.16b, v7.16b, v9.16b
|
|
tbl v6.16b, {v10.16b}, v8.16b
|
|
tbl v7.16b, {v11.16b}, v8.16b
|
|
|
|
bitslice v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11
|
|
|
|
sub rounds, rounds, #1
|
|
b .Ldec_sbox
|
|
|
|
.Ldec_loop:
|
|
shift_rows v0, v1, v2, v3, v4, v5, v6, v7, v24
|
|
.Ldec_sbox:
|
|
inv_sbox v0, v1, v2, v3, v4, v5, v6, v7, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
subs rounds, rounds, #1
|
|
b.cc .Ldec_done
|
|
|
|
dec_next_rk
|
|
|
|
add_round_key v0, v1, v6, v4, v2, v7, v3, v5
|
|
|
|
inv_mix_cols v0, v1, v6, v4, v2, v7, v3, v5, v8, v9, v10, v11, v12, \
|
|
v13, v14, v15
|
|
|
|
b.ne .Ldec_loop
|
|
ldr q24, ISRM0
|
|
b .Ldec_loop
|
|
.Ldec_done:
|
|
ldr q12, [bskey, #-16] // last round key
|
|
|
|
bitslice v0, v1, v6, v4, v2, v7, v3, v5, v8, v9, v10, v11
|
|
|
|
eor v0.16b, v0.16b, v12.16b
|
|
eor v1.16b, v1.16b, v12.16b
|
|
eor v6.16b, v6.16b, v12.16b
|
|
eor v4.16b, v4.16b, v12.16b
|
|
eor v2.16b, v2.16b, v12.16b
|
|
eor v7.16b, v7.16b, v12.16b
|
|
eor v3.16b, v3.16b, v12.16b
|
|
eor v5.16b, v5.16b, v12.16b
|
|
ret
|
|
SYM_FUNC_END(aesbs_decrypt8)
|
|
|
|
/*
|
|
* aesbs_ecb_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks)
|
|
* aesbs_ecb_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks)
|
|
*/
|
|
.macro __ecb_crypt, do8, o0, o1, o2, o3, o4, o5, o6, o7
|
|
frame_push 5
|
|
|
|
mov x19, x0
|
|
mov x20, x1
|
|
mov x21, x2
|
|
mov x22, x3
|
|
mov x23, x4
|
|
|
|
99: mov x5, #1
|
|
lsl x5, x5, x23
|
|
subs w23, w23, #8
|
|
csel x23, x23, xzr, pl
|
|
csel x5, x5, xzr, mi
|
|
|
|
ld1 {v0.16b}, [x20], #16
|
|
tbnz x5, #1, 0f
|
|
ld1 {v1.16b}, [x20], #16
|
|
tbnz x5, #2, 0f
|
|
ld1 {v2.16b}, [x20], #16
|
|
tbnz x5, #3, 0f
|
|
ld1 {v3.16b}, [x20], #16
|
|
tbnz x5, #4, 0f
|
|
ld1 {v4.16b}, [x20], #16
|
|
tbnz x5, #5, 0f
|
|
ld1 {v5.16b}, [x20], #16
|
|
tbnz x5, #6, 0f
|
|
ld1 {v6.16b}, [x20], #16
|
|
tbnz x5, #7, 0f
|
|
ld1 {v7.16b}, [x20], #16
|
|
|
|
0: mov bskey, x21
|
|
mov rounds, x22
|
|
bl \do8
|
|
|
|
st1 {\o0\().16b}, [x19], #16
|
|
tbnz x5, #1, 1f
|
|
st1 {\o1\().16b}, [x19], #16
|
|
tbnz x5, #2, 1f
|
|
st1 {\o2\().16b}, [x19], #16
|
|
tbnz x5, #3, 1f
|
|
st1 {\o3\().16b}, [x19], #16
|
|
tbnz x5, #4, 1f
|
|
st1 {\o4\().16b}, [x19], #16
|
|
tbnz x5, #5, 1f
|
|
st1 {\o5\().16b}, [x19], #16
|
|
tbnz x5, #6, 1f
|
|
st1 {\o6\().16b}, [x19], #16
|
|
tbnz x5, #7, 1f
|
|
st1 {\o7\().16b}, [x19], #16
|
|
|
|
cbz x23, 1f
|
|
b 99b
|
|
|
|
1: frame_pop
|
|
ret
|
|
.endm
|
|
|
|
.align 4
|
|
SYM_FUNC_START(aesbs_ecb_encrypt)
|
|
__ecb_crypt aesbs_encrypt8, v0, v1, v4, v6, v3, v7, v2, v5
|
|
SYM_FUNC_END(aesbs_ecb_encrypt)
|
|
|
|
.align 4
|
|
SYM_FUNC_START(aesbs_ecb_decrypt)
|
|
__ecb_crypt aesbs_decrypt8, v0, v1, v6, v4, v2, v7, v3, v5
|
|
SYM_FUNC_END(aesbs_ecb_decrypt)
|
|
|
|
/*
|
|
* aesbs_cbc_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
*/
|
|
.align 4
|
|
SYM_FUNC_START(aesbs_cbc_decrypt)
|
|
frame_push 6
|
|
|
|
mov x19, x0
|
|
mov x20, x1
|
|
mov x21, x2
|
|
mov x22, x3
|
|
mov x23, x4
|
|
mov x24, x5
|
|
|
|
99: mov x6, #1
|
|
lsl x6, x6, x23
|
|
subs w23, w23, #8
|
|
csel x23, x23, xzr, pl
|
|
csel x6, x6, xzr, mi
|
|
|
|
ld1 {v0.16b}, [x20], #16
|
|
mov v25.16b, v0.16b
|
|
tbnz x6, #1, 0f
|
|
ld1 {v1.16b}, [x20], #16
|
|
mov v26.16b, v1.16b
|
|
tbnz x6, #2, 0f
|
|
ld1 {v2.16b}, [x20], #16
|
|
mov v27.16b, v2.16b
|
|
tbnz x6, #3, 0f
|
|
ld1 {v3.16b}, [x20], #16
|
|
mov v28.16b, v3.16b
|
|
tbnz x6, #4, 0f
|
|
ld1 {v4.16b}, [x20], #16
|
|
mov v29.16b, v4.16b
|
|
tbnz x6, #5, 0f
|
|
ld1 {v5.16b}, [x20], #16
|
|
mov v30.16b, v5.16b
|
|
tbnz x6, #6, 0f
|
|
ld1 {v6.16b}, [x20], #16
|
|
mov v31.16b, v6.16b
|
|
tbnz x6, #7, 0f
|
|
ld1 {v7.16b}, [x20]
|
|
|
|
0: mov bskey, x21
|
|
mov rounds, x22
|
|
bl aesbs_decrypt8
|
|
|
|
ld1 {v24.16b}, [x24] // load IV
|
|
|
|
eor v1.16b, v1.16b, v25.16b
|
|
eor v6.16b, v6.16b, v26.16b
|
|
eor v4.16b, v4.16b, v27.16b
|
|
eor v2.16b, v2.16b, v28.16b
|
|
eor v7.16b, v7.16b, v29.16b
|
|
eor v0.16b, v0.16b, v24.16b
|
|
eor v3.16b, v3.16b, v30.16b
|
|
eor v5.16b, v5.16b, v31.16b
|
|
|
|
st1 {v0.16b}, [x19], #16
|
|
mov v24.16b, v25.16b
|
|
tbnz x6, #1, 1f
|
|
st1 {v1.16b}, [x19], #16
|
|
mov v24.16b, v26.16b
|
|
tbnz x6, #2, 1f
|
|
st1 {v6.16b}, [x19], #16
|
|
mov v24.16b, v27.16b
|
|
tbnz x6, #3, 1f
|
|
st1 {v4.16b}, [x19], #16
|
|
mov v24.16b, v28.16b
|
|
tbnz x6, #4, 1f
|
|
st1 {v2.16b}, [x19], #16
|
|
mov v24.16b, v29.16b
|
|
tbnz x6, #5, 1f
|
|
st1 {v7.16b}, [x19], #16
|
|
mov v24.16b, v30.16b
|
|
tbnz x6, #6, 1f
|
|
st1 {v3.16b}, [x19], #16
|
|
mov v24.16b, v31.16b
|
|
tbnz x6, #7, 1f
|
|
ld1 {v24.16b}, [x20], #16
|
|
st1 {v5.16b}, [x19], #16
|
|
1: st1 {v24.16b}, [x24] // store IV
|
|
|
|
cbz x23, 2f
|
|
b 99b
|
|
|
|
2: frame_pop
|
|
ret
|
|
SYM_FUNC_END(aesbs_cbc_decrypt)
|
|
|
|
.macro next_tweak, out, in, const, tmp
|
|
sshr \tmp\().2d, \in\().2d, #63
|
|
and \tmp\().16b, \tmp\().16b, \const\().16b
|
|
add \out\().2d, \in\().2d, \in\().2d
|
|
ext \tmp\().16b, \tmp\().16b, \tmp\().16b, #8
|
|
eor \out\().16b, \out\().16b, \tmp\().16b
|
|
.endm
|
|
|
|
/*
|
|
* aesbs_xts_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
* aesbs_xts_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
*/
|
|
SYM_FUNC_START_LOCAL(__xts_crypt8)
|
|
movi v18.2s, #0x1
|
|
movi v19.2s, #0x87
|
|
uzp1 v18.4s, v18.4s, v19.4s
|
|
|
|
ld1 {v0.16b-v3.16b}, [x1], #64
|
|
ld1 {v4.16b-v7.16b}, [x1], #64
|
|
|
|
next_tweak v26, v25, v18, v19
|
|
next_tweak v27, v26, v18, v19
|
|
next_tweak v28, v27, v18, v19
|
|
next_tweak v29, v28, v18, v19
|
|
next_tweak v30, v29, v18, v19
|
|
next_tweak v31, v30, v18, v19
|
|
next_tweak v16, v31, v18, v19
|
|
next_tweak v17, v16, v18, v19
|
|
|
|
eor v0.16b, v0.16b, v25.16b
|
|
eor v1.16b, v1.16b, v26.16b
|
|
eor v2.16b, v2.16b, v27.16b
|
|
eor v3.16b, v3.16b, v28.16b
|
|
eor v4.16b, v4.16b, v29.16b
|
|
eor v5.16b, v5.16b, v30.16b
|
|
eor v6.16b, v6.16b, v31.16b
|
|
eor v7.16b, v7.16b, v16.16b
|
|
|
|
stp q16, q17, [sp, #16]
|
|
|
|
mov bskey, x2
|
|
mov rounds, x3
|
|
br x16
|
|
SYM_FUNC_END(__xts_crypt8)
|
|
|
|
.macro __xts_crypt, do8, o0, o1, o2, o3, o4, o5, o6, o7
|
|
stp x29, x30, [sp, #-48]!
|
|
mov x29, sp
|
|
|
|
ld1 {v25.16b}, [x5]
|
|
|
|
0: adr x16, \do8
|
|
bl __xts_crypt8
|
|
|
|
eor v16.16b, \o0\().16b, v25.16b
|
|
eor v17.16b, \o1\().16b, v26.16b
|
|
eor v18.16b, \o2\().16b, v27.16b
|
|
eor v19.16b, \o3\().16b, v28.16b
|
|
|
|
ldp q24, q25, [sp, #16]
|
|
|
|
eor v20.16b, \o4\().16b, v29.16b
|
|
eor v21.16b, \o5\().16b, v30.16b
|
|
eor v22.16b, \o6\().16b, v31.16b
|
|
eor v23.16b, \o7\().16b, v24.16b
|
|
|
|
st1 {v16.16b-v19.16b}, [x0], #64
|
|
st1 {v20.16b-v23.16b}, [x0], #64
|
|
|
|
subs x4, x4, #8
|
|
b.gt 0b
|
|
|
|
st1 {v25.16b}, [x5]
|
|
ldp x29, x30, [sp], #48
|
|
ret
|
|
.endm
|
|
|
|
SYM_FUNC_START(aesbs_xts_encrypt)
|
|
__xts_crypt aesbs_encrypt8, v0, v1, v4, v6, v3, v7, v2, v5
|
|
SYM_FUNC_END(aesbs_xts_encrypt)
|
|
|
|
SYM_FUNC_START(aesbs_xts_decrypt)
|
|
__xts_crypt aesbs_decrypt8, v0, v1, v6, v4, v2, v7, v3, v5
|
|
SYM_FUNC_END(aesbs_xts_decrypt)
|
|
|
|
.macro next_ctr, v
|
|
mov \v\().d[1], x8
|
|
adds x8, x8, #1
|
|
mov \v\().d[0], x7
|
|
adc x7, x7, xzr
|
|
rev64 \v\().16b, \v\().16b
|
|
.endm
|
|
|
|
/*
|
|
* aesbs_ctr_encrypt(u8 out[], u8 const in[], u8 const rk[],
|
|
* int rounds, int blocks, u8 iv[])
|
|
*/
|
|
SYM_FUNC_START(aesbs_ctr_encrypt)
|
|
stp x29, x30, [sp, #-16]!
|
|
mov x29, sp
|
|
|
|
ldp x7, x8, [x5]
|
|
ld1 {v0.16b}, [x5]
|
|
CPU_LE( rev x7, x7 )
|
|
CPU_LE( rev x8, x8 )
|
|
adds x8, x8, #1
|
|
adc x7, x7, xzr
|
|
|
|
0: next_ctr v1
|
|
next_ctr v2
|
|
next_ctr v3
|
|
next_ctr v4
|
|
next_ctr v5
|
|
next_ctr v6
|
|
next_ctr v7
|
|
|
|
mov bskey, x2
|
|
mov rounds, x3
|
|
bl aesbs_encrypt8
|
|
|
|
ld1 { v8.16b-v11.16b}, [x1], #64
|
|
ld1 {v12.16b-v15.16b}, [x1], #64
|
|
|
|
eor v8.16b, v0.16b, v8.16b
|
|
eor v9.16b, v1.16b, v9.16b
|
|
eor v10.16b, v4.16b, v10.16b
|
|
eor v11.16b, v6.16b, v11.16b
|
|
eor v12.16b, v3.16b, v12.16b
|
|
eor v13.16b, v7.16b, v13.16b
|
|
eor v14.16b, v2.16b, v14.16b
|
|
eor v15.16b, v5.16b, v15.16b
|
|
|
|
st1 { v8.16b-v11.16b}, [x0], #64
|
|
st1 {v12.16b-v15.16b}, [x0], #64
|
|
|
|
next_ctr v0
|
|
subs x4, x4, #8
|
|
b.gt 0b
|
|
|
|
st1 {v0.16b}, [x5]
|
|
ldp x29, x30, [sp], #16
|
|
ret
|
|
SYM_FUNC_END(aesbs_ctr_encrypt)
|