root/lib/crypto/sparc/aes.h
/* SPDX-License-Identifier: GPL-2.0-only */
/*
 * AES accelerated using the sparc64 aes opcodes
 *
 * Copyright (C) 2008, Intel Corp.
 * Copyright (c) 2010, Intel Corporation.
 * Copyright 2026 Google LLC
 */

#include <asm/fpumacro.h>
#include <asm/opcodes.h>
#include <asm/pstate.h>
#include <asm/elf.h>

static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_aes_opcodes);

EXPORT_SYMBOL_GPL(aes_sparc64_key_expand);
EXPORT_SYMBOL_GPL(aes_sparc64_load_encrypt_keys_128);
EXPORT_SYMBOL_GPL(aes_sparc64_load_encrypt_keys_192);
EXPORT_SYMBOL_GPL(aes_sparc64_load_encrypt_keys_256);
EXPORT_SYMBOL_GPL(aes_sparc64_load_decrypt_keys_128);
EXPORT_SYMBOL_GPL(aes_sparc64_load_decrypt_keys_192);
EXPORT_SYMBOL_GPL(aes_sparc64_load_decrypt_keys_256);
EXPORT_SYMBOL_GPL(aes_sparc64_ecb_encrypt_128);
EXPORT_SYMBOL_GPL(aes_sparc64_ecb_encrypt_192);
EXPORT_SYMBOL_GPL(aes_sparc64_ecb_encrypt_256);
EXPORT_SYMBOL_GPL(aes_sparc64_ecb_decrypt_128);
EXPORT_SYMBOL_GPL(aes_sparc64_ecb_decrypt_192);
EXPORT_SYMBOL_GPL(aes_sparc64_ecb_decrypt_256);
EXPORT_SYMBOL_GPL(aes_sparc64_cbc_encrypt_128);
EXPORT_SYMBOL_GPL(aes_sparc64_cbc_encrypt_192);
EXPORT_SYMBOL_GPL(aes_sparc64_cbc_encrypt_256);
EXPORT_SYMBOL_GPL(aes_sparc64_cbc_decrypt_128);
EXPORT_SYMBOL_GPL(aes_sparc64_cbc_decrypt_192);
EXPORT_SYMBOL_GPL(aes_sparc64_cbc_decrypt_256);
EXPORT_SYMBOL_GPL(aes_sparc64_ctr_crypt_128);
EXPORT_SYMBOL_GPL(aes_sparc64_ctr_crypt_192);
EXPORT_SYMBOL_GPL(aes_sparc64_ctr_crypt_256);

void aes_sparc64_encrypt_128(const u64 *key, const u32 *input, u32 *output);
void aes_sparc64_encrypt_192(const u64 *key, const u32 *input, u32 *output);
void aes_sparc64_encrypt_256(const u64 *key, const u32 *input, u32 *output);
void aes_sparc64_decrypt_128(const u64 *key, const u32 *input, u32 *output);
void aes_sparc64_decrypt_192(const u64 *key, const u32 *input, u32 *output);
void aes_sparc64_decrypt_256(const u64 *key, const u32 *input, u32 *output);

static void aes_preparekey_arch(union aes_enckey_arch *k,
                                union aes_invkey_arch *inv_k,
                                const u8 *in_key, int key_len, int nrounds)
{
        if (static_branch_likely(&have_aes_opcodes)) {
                u32 aligned_key[AES_MAX_KEY_SIZE / 4];

                if (IS_ALIGNED((uintptr_t)in_key, 4)) {
                        aes_sparc64_key_expand((const u32 *)in_key,
                                               k->sparc_rndkeys, key_len);
                } else {
                        memcpy(aligned_key, in_key, key_len);
                        aes_sparc64_key_expand(aligned_key,
                                               k->sparc_rndkeys, key_len);
                        memzero_explicit(aligned_key, key_len);
                }
                /*
                 * Note that nothing needs to be written to inv_k (if it's
                 * non-NULL) here, since the SPARC64 assembly code uses
                 * k->sparc_rndkeys for both encryption and decryption.
                 */
        } else {
                aes_expandkey_generic(k->rndkeys,
                                      inv_k ? inv_k->inv_rndkeys : NULL,
                                      in_key, key_len);
        }
}

static void aes_sparc64_encrypt(const struct aes_enckey *key,
                                const u32 *input, u32 *output)
{
        if (key->len == AES_KEYSIZE_128)
                aes_sparc64_encrypt_128(key->k.sparc_rndkeys, input, output);
        else if (key->len == AES_KEYSIZE_192)
                aes_sparc64_encrypt_192(key->k.sparc_rndkeys, input, output);
        else
                aes_sparc64_encrypt_256(key->k.sparc_rndkeys, input, output);
}

static void aes_encrypt_arch(const struct aes_enckey *key,
                             u8 out[AES_BLOCK_SIZE],
                             const u8 in[AES_BLOCK_SIZE])
{
        u32 bounce_buf[AES_BLOCK_SIZE / 4];

        if (static_branch_likely(&have_aes_opcodes)) {
                if (IS_ALIGNED((uintptr_t)in | (uintptr_t)out, 4)) {
                        aes_sparc64_encrypt(key, (const u32 *)in, (u32 *)out);
                } else {
                        memcpy(bounce_buf, in, AES_BLOCK_SIZE);
                        aes_sparc64_encrypt(key, bounce_buf, bounce_buf);
                        memcpy(out, bounce_buf, AES_BLOCK_SIZE);
                }
        } else {
                aes_encrypt_generic(key->k.rndkeys, key->nrounds, out, in);
        }
}

static void aes_sparc64_decrypt(const struct aes_key *key,
                                const u32 *input, u32 *output)
{
        if (key->len == AES_KEYSIZE_128)
                aes_sparc64_decrypt_128(key->k.sparc_rndkeys, input, output);
        else if (key->len == AES_KEYSIZE_192)
                aes_sparc64_decrypt_192(key->k.sparc_rndkeys, input, output);
        else
                aes_sparc64_decrypt_256(key->k.sparc_rndkeys, input, output);
}

static void aes_decrypt_arch(const struct aes_key *key,
                             u8 out[AES_BLOCK_SIZE],
                             const u8 in[AES_BLOCK_SIZE])
{
        u32 bounce_buf[AES_BLOCK_SIZE / 4];

        if (static_branch_likely(&have_aes_opcodes)) {
                if (IS_ALIGNED((uintptr_t)in | (uintptr_t)out, 4)) {
                        aes_sparc64_decrypt(key, (const u32 *)in, (u32 *)out);
                } else {
                        memcpy(bounce_buf, in, AES_BLOCK_SIZE);
                        aes_sparc64_decrypt(key, bounce_buf, bounce_buf);
                        memcpy(out, bounce_buf, AES_BLOCK_SIZE);
                }
        } else {
                aes_decrypt_generic(key->inv_k.inv_rndkeys, key->nrounds,
                                    out, in);
        }
}

#define aes_mod_init_arch aes_mod_init_arch
static void aes_mod_init_arch(void)
{
        unsigned long cfr;

        if (!(sparc64_elf_hwcap & HWCAP_SPARC_CRYPTO))
                return;

        __asm__ __volatile__("rd %%asr26, %0" : "=r" (cfr));
        if (!(cfr & CFR_AES))
                return;

        static_branch_enable(&have_aes_opcodes);
}