[PATCH v5 4/8] crypto: x86/aesni-xctr: Add accelerated implementation of XCTR
Eric Biggers
ebiggers at kernel.org
Sun May 1 14:31:32 PDT 2022
On Wed, Apr 27, 2022 at 12:37:55AM +0000, Nathan Huckleberry wrote:
> Add hardware accelerated versions of XCTR for x86-64 CPUs with AESNI
> support. These implementations are modified versions of the CTR
> implementations found in aesni-intel_asm.S and aes_ctrby8_avx-x86_64.S.
Just one implementation now, using aes_ctrby8_avx-x86_64.S.
> +/* Note: the "x" prefix in these aliases means "this is an xmm register". The
> + * alias prefixes have no relation to XCTR where the "X" prefix means "XOR
> + * counter".
> + */
Block comments look like:
/*
* text
*/
> + .if !\xctr
> + vpshufb xbyteswap, xcounter, xdata0
> + .set i, 1
> + .rept (by - 1)
> + club XDATA, i
> + vpaddq (ddq_add_1 + 16 * (i - 1))(%rip), xcounter, var_xdata
> + vptest ddq_low_msk(%rip), var_xdata
> + jnz 1f
> + vpaddq ddq_high_add_1(%rip), var_xdata, var_xdata
> + vpaddq ddq_high_add_1(%rip), xcounter, xcounter
> + 1:
> + vpshufb xbyteswap, var_xdata, var_xdata
> + .set i, (i +1)
> + .endr
> + .else
> + movq counter, xtmp
> + .set i, 0
> + .rept (by)
> + club XDATA, i
> + vpaddq (ddq_add_1 + 16 * i)(%rip), xtmp, var_xdata
> + .set i, (i +1)
> + .endr
> + .set i, 0
> + .rept (by)
> + club XDATA, i
> + vpxor xiv, var_xdata, var_xdata
> + .set i, (i +1)
> + .endr
> + .endif
I'm not a fan of 'if !condition ... else ...', as the else clause is
double-negated. It's more straightforward to do 'if condition ... else ...'.
> + .if !\xctr
> + vmovdqa byteswap_const(%rip), xbyteswap
> + vmovdqu (p_iv), xcounter
> + vpshufb xbyteswap, xcounter, xcounter
> + .else
> + andq $(~0xf), num_bytes
> + shr $4, counter
> + vmovdqu (p_iv), xiv
> + .endif
Isn't the 'andq $(~0xf), num_bytes' instruction unnecessary? If it is
necessary, I'd expect it to be necessary for CTR too.
Otherwise this file looks good.
Note, the macros in this file all expand to way too much code, especially due to
the separate cases for AES-128, AES-192, and AES-256, and for each one every
partial stride length 1..7. Of course, this is true for the existing CTR code
too, so I don't think you have to fix this... But maybe think about addressing
this later. Changing the handling of partial strides might be the easiest way
to save a lot of code without hurting any micro-benchmarks too much. Also maybe
some or all of the AES key sizes could be combined.
> +#ifdef CONFIG_X86_64
> +/*
> + * XCTR does not have a non-AVX implementation, so it must be enabled
> + * conditionally.
> + */
> +static struct skcipher_alg aesni_xctr = {
> + .base = {
> + .cra_name = "__xctr(aes)",
> + .cra_driver_name = "__xctr-aes-aesni",
> + .cra_priority = 400,
> + .cra_flags = CRYPTO_ALG_INTERNAL,
> + .cra_blocksize = 1,
> + .cra_ctxsize = CRYPTO_AES_CTX_SIZE,
> + .cra_module = THIS_MODULE,
> + },
> + .min_keysize = AES_MIN_KEY_SIZE,
> + .max_keysize = AES_MAX_KEY_SIZE,
> + .ivsize = AES_BLOCK_SIZE,
> + .chunksize = AES_BLOCK_SIZE,
> + .setkey = aesni_skcipher_setkey,
> + .encrypt = xctr_crypt,
> + .decrypt = xctr_crypt,
> +};
> +
> +static struct simd_skcipher_alg *aesni_simd_xctr;
> +#endif
Comment the #endif above:
#endif /* CONFIG_X86_64 */
> @@ -1180,8 +1274,19 @@ static int __init aesni_init(void)
> if (err)
> goto unregister_skciphers;
>
> +#ifdef CONFIG_X86_64
> + if (boot_cpu_has(X86_FEATURE_AVX))
> + err = simd_register_skciphers_compat(&aesni_xctr, 1,
> + &aesni_simd_xctr);
> + if (err)
> + goto unregister_aeads;
> +#endif
> +
> return 0;
>
> +unregister_aeads:
> + simd_unregister_aeads(aesni_aeads, ARRAY_SIZE(aesni_aeads),
> + aesni_simd_aeads);
This will cause a compiler warning in 32-bit builds because the
'unregister_aeads' label won't be used.
- Eric
More information about the linux-arm-kernel
mailing list