root/lib/crypto/s390/gf128hash.h
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * GHASH optimized using the CP Assist for Cryptographic Functions (CPACF)
 *
 * Copyright 2026 Google LLC
 */
#include <asm/cpacf.h>
#include <linux/cpufeature.h>

static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_ghash);

#define ghash_preparekey_arch ghash_preparekey_arch
static void ghash_preparekey_arch(struct ghash_key *key,
                                  const u8 raw_key[GHASH_BLOCK_SIZE])
{
        /* Save key in POLYVAL format for fallback */
        ghash_key_to_polyval(raw_key, &key->h);

        /* Save key in GHASH format for CPACF_KIMD_GHASH */
        memcpy(key->h_raw, raw_key, GHASH_BLOCK_SIZE);
}

#define ghash_blocks_arch ghash_blocks_arch
static void ghash_blocks_arch(struct polyval_elem *acc,
                              const struct ghash_key *key,
                              const u8 *data, size_t nblocks)
{
        if (static_branch_likely(&have_cpacf_ghash)) {
                /*
                 * CPACF_KIMD_GHASH requires the accumulator and key in a single
                 * buffer, each using the GHASH convention.
                 */
                u8 ctx[2][GHASH_BLOCK_SIZE] __aligned(8);

                polyval_acc_to_ghash(acc, ctx[0]);
                memcpy(ctx[1], key->h_raw, GHASH_BLOCK_SIZE);

                cpacf_kimd(CPACF_KIMD_GHASH, ctx, data,
                           nblocks * GHASH_BLOCK_SIZE);

                ghash_acc_to_polyval(ctx[0], acc);
                memzero_explicit(ctx, sizeof(ctx));
        } else {
                ghash_blocks_generic(acc, &key->h, data, nblocks);
        }
}

#define gf128hash_mod_init_arch gf128hash_mod_init_arch
static void gf128hash_mod_init_arch(void)
{
        if (cpu_have_feature(S390_CPU_FEATURE_MSA) &&
            cpacf_query_func(CPACF_KIMD, CPACF_KIMD_GHASH))
                static_branch_enable(&have_cpacf_ghash);
}