root/include/crypto/acompress.h
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * Asynchronous Compression operations
 *
 * Copyright (c) 2016, Intel Corporation
 * Authors: Weigang Li <weigang.li@intel.com>
 *          Giovanni Cabiddu <giovanni.cabiddu@intel.com>
 */
#ifndef _CRYPTO_ACOMP_H
#define _CRYPTO_ACOMP_H

#include <linux/atomic.h>
#include <linux/args.h>
#include <linux/compiler_types.h>
#include <linux/container_of.h>
#include <linux/crypto.h>
#include <linux/err.h>
#include <linux/scatterlist.h>
#include <linux/slab.h>
#include <linux/spinlock_types.h>
#include <linux/types.h>

/* Set this bit if source is virtual address instead of SG list. */
#define CRYPTO_ACOMP_REQ_SRC_VIRT       0x00000002

/* Set this bit for if virtual address source cannot be used for DMA. */
#define CRYPTO_ACOMP_REQ_SRC_NONDMA     0x00000004

/* Set this bit if destination is virtual address instead of SG list. */
#define CRYPTO_ACOMP_REQ_DST_VIRT       0x00000008

/* Set this bit for if virtual address destination cannot be used for DMA. */
#define CRYPTO_ACOMP_REQ_DST_NONDMA     0x00000010

/* Private flags that should not be touched by the user. */
#define CRYPTO_ACOMP_REQ_PRIVATE \
        (CRYPTO_ACOMP_REQ_SRC_VIRT | CRYPTO_ACOMP_REQ_SRC_NONDMA | \
         CRYPTO_ACOMP_REQ_DST_VIRT | CRYPTO_ACOMP_REQ_DST_NONDMA)

#define CRYPTO_ACOMP_DST_MAX            131072

#define MAX_SYNC_COMP_REQSIZE           0

#define ACOMP_REQUEST_ON_STACK(name, tfm) \
        char __##name##_req[sizeof(struct acomp_req) + \
                            MAX_SYNC_COMP_REQSIZE] CRYPTO_MINALIGN_ATTR; \
        struct acomp_req *name = acomp_request_on_stack_init( \
                __##name##_req, (tfm))

#define ACOMP_REQUEST_CLONE(name, gfp) \
        acomp_request_clone(name, sizeof(__##name##_req), gfp)

struct acomp_req;
struct folio;

struct acomp_req_chain {
        crypto_completion_t compl;
        void *data;
        struct scatterlist ssg;
        struct scatterlist dsg;
        union {
                const u8 *src;
                struct folio *sfolio;
        };
        union {
                u8 *dst;
                struct folio *dfolio;
        };
        u32 flags;
};

/**
 * struct acomp_req - asynchronous (de)compression request
 *
 * @base:       Common attributes for asynchronous crypto requests
 * @src:        Source scatterlist
 * @dst:        Destination scatterlist
 * @svirt:      Source virtual address
 * @dvirt:      Destination virtual address
 * @slen:       Size of the input buffer
 * @dlen:       Size of the output buffer and number of bytes produced
 * @chain:      Private API code data, do not use
 * @__ctx:      Start of private context data
 */
struct acomp_req {
        struct crypto_async_request base;
        union {
                struct scatterlist *src;
                const u8 *svirt;
        };
        union {
                struct scatterlist *dst;
                u8 *dvirt;
        };
        unsigned int slen;
        unsigned int dlen;

        struct acomp_req_chain chain;

        void *__ctx[] CRYPTO_MINALIGN_ATTR;
};

/**
 * struct crypto_acomp - user-instantiated objects which encapsulate
 * algorithms and core processing logic
 *
 * @compress:           Function performs a compress operation
 * @decompress:         Function performs a de-compress operation
 * @reqsize:            Context size for (de)compression requests
 * @fb:                 Synchronous fallback tfm
 * @base:               Common crypto API algorithm data structure
 */
struct crypto_acomp {
        int (*compress)(struct acomp_req *req);
        int (*decompress)(struct acomp_req *req);
        unsigned int reqsize;
        struct crypto_tfm base;
};

#define COMP_ALG_COMMON {                       \
        struct crypto_alg base;                 \
}
struct comp_alg_common COMP_ALG_COMMON;

/**
 * DOC: Asynchronous Compression API
 *
 * The Asynchronous Compression API is used with the algorithms of type
 * CRYPTO_ALG_TYPE_ACOMPRESS (listed as type "acomp" in /proc/crypto)
 */

/**
 * crypto_alloc_acomp() -- allocate ACOMPRESS tfm handle
 * @alg_name:   is the cra_name / name or cra_driver_name / driver name of the
 *              compression algorithm e.g. "deflate"
 * @type:       specifies the type of the algorithm
 * @mask:       specifies the mask for the algorithm
 *
 * Allocate a handle for a compression algorithm. The returned struct
 * crypto_acomp is the handle that is required for any subsequent
 * API invocation for the compression operations.
 *
 * Return:      allocated handle in case of success; IS_ERR() is true in case
 *              of an error, PTR_ERR() returns the error code.
 */
struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
                                        u32 mask);
/**
 * crypto_alloc_acomp_node() -- allocate ACOMPRESS tfm handle with desired NUMA node
 * @alg_name:   is the cra_name / name or cra_driver_name / driver name of the
 *              compression algorithm e.g. "deflate"
 * @type:       specifies the type of the algorithm
 * @mask:       specifies the mask for the algorithm
 * @node:       specifies the NUMA node the ZIP hardware belongs to
 *
 * Allocate a handle for a compression algorithm. Drivers should try to use
 * (de)compressors on the specified NUMA node.
 * The returned struct crypto_acomp is the handle that is required for any
 * subsequent API invocation for the compression operations.
 *
 * Return:      allocated handle in case of success; IS_ERR() is true in case
 *              of an error, PTR_ERR() returns the error code.
 */
struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
                                        u32 mask, int node);

static inline struct crypto_tfm *crypto_acomp_tfm(struct crypto_acomp *tfm)
{
        return &tfm->base;
}

static inline struct comp_alg_common *__crypto_comp_alg_common(
        struct crypto_alg *alg)
{
        return container_of(alg, struct comp_alg_common, base);
}

static inline struct crypto_acomp *__crypto_acomp_tfm(struct crypto_tfm *tfm)
{
        return container_of(tfm, struct crypto_acomp, base);
}

static inline struct comp_alg_common *crypto_comp_alg_common(
        struct crypto_acomp *tfm)
{
        return __crypto_comp_alg_common(crypto_acomp_tfm(tfm)->__crt_alg);
}

static inline unsigned int crypto_acomp_reqsize(struct crypto_acomp *tfm)
{
        return tfm->reqsize;
}

static inline void acomp_request_set_tfm(struct acomp_req *req,
                                         struct crypto_acomp *tfm)
{
        crypto_request_set_tfm(&req->base, crypto_acomp_tfm(tfm));
}

static inline bool acomp_is_async(struct crypto_acomp *tfm)
{
        return crypto_comp_alg_common(tfm)->base.cra_flags &
               CRYPTO_ALG_ASYNC;
}

static inline struct crypto_acomp *crypto_acomp_reqtfm(struct acomp_req *req)
{
        return __crypto_acomp_tfm(req->base.tfm);
}

/**
 * crypto_free_acomp() -- free ACOMPRESS tfm handle
 *
 * @tfm:        ACOMPRESS tfm handle allocated with crypto_alloc_acomp()
 *
 * If @tfm is a NULL or error pointer, this function does nothing.
 */
static inline void crypto_free_acomp(struct crypto_acomp *tfm)
{
        crypto_destroy_tfm(tfm, crypto_acomp_tfm(tfm));
}

static inline int crypto_has_acomp(const char *alg_name, u32 type, u32 mask)
{
        type &= ~CRYPTO_ALG_TYPE_MASK;
        type |= CRYPTO_ALG_TYPE_ACOMPRESS;
        mask |= CRYPTO_ALG_TYPE_ACOMPRESS_MASK;

        return crypto_has_alg(alg_name, type, mask);
}

static inline const char *crypto_acomp_alg_name(struct crypto_acomp *tfm)
{
        return crypto_tfm_alg_name(crypto_acomp_tfm(tfm));
}

static inline const char *crypto_acomp_driver_name(struct crypto_acomp *tfm)
{
        return crypto_tfm_alg_driver_name(crypto_acomp_tfm(tfm));
}

/**
 * acomp_request_alloc() -- allocates asynchronous (de)compression request
 *
 * @tfm:        ACOMPRESS tfm handle allocated with crypto_alloc_acomp()
 * @gfp:        gfp to pass to kzalloc (defaults to GFP_KERNEL)
 *
 * Return:      allocated handle in case of success or NULL in case of an error
 */
static inline struct acomp_req *acomp_request_alloc_extra_noprof(
        struct crypto_acomp *tfm, size_t extra, gfp_t gfp)
{
        struct acomp_req *req;
        size_t len;

        len = ALIGN(sizeof(*req) + crypto_acomp_reqsize(tfm), CRYPTO_MINALIGN);
        if (check_add_overflow(len, extra, &len))
                return NULL;

        req = kzalloc_noprof(len, gfp);
        if (likely(req))
                acomp_request_set_tfm(req, tfm);
        return req;
}
#define acomp_request_alloc_noprof(tfm, ...) \
        CONCATENATE(acomp_request_alloc_noprof_, COUNT_ARGS(__VA_ARGS__))( \
                tfm, ##__VA_ARGS__)
#define acomp_request_alloc_noprof_0(tfm) \
        acomp_request_alloc_noprof_1(tfm, GFP_KERNEL)
#define acomp_request_alloc_noprof_1(tfm, gfp) \
        acomp_request_alloc_extra_noprof(tfm, 0, gfp)
#define acomp_request_alloc(...)        alloc_hooks(acomp_request_alloc_noprof(__VA_ARGS__))

/**
 * acomp_request_alloc_extra() -- allocate acomp request with extra memory
 *
 * @tfm:        ACOMPRESS tfm handle allocated with crypto_alloc_acomp()
 * @extra:      amount of extra memory
 * @gfp:        gfp to pass to kzalloc
 *
 * Return:      allocated handle in case of success or NULL in case of an error
 */
#define acomp_request_alloc_extra(...)  alloc_hooks(acomp_request_alloc_extra_noprof(__VA_ARGS__))

static inline void *acomp_request_extra(struct acomp_req *req)
{
        struct crypto_acomp *tfm = crypto_acomp_reqtfm(req);
        size_t len;

        len = ALIGN(sizeof(*req) + crypto_acomp_reqsize(tfm), CRYPTO_MINALIGN);
        return (void *)((char *)req + len);
}

static inline bool acomp_req_on_stack(struct acomp_req *req)
{
        return crypto_req_on_stack(&req->base);
}

/**
 * acomp_request_free() -- zeroize and free asynchronous (de)compression
 *                         request as well as the output buffer if allocated
 *                         inside the algorithm
 *
 * @req:        request to free
 */
static inline void acomp_request_free(struct acomp_req *req)
{
        if (!req || acomp_req_on_stack(req))
                return;
        kfree_sensitive(req);
}

/**
 * acomp_request_set_callback() -- Sets an asynchronous callback
 *
 * Callback will be called when an asynchronous operation on a given
 * request is finished.
 *
 * @req:        request that the callback will be set for
 * @flgs:       specify for instance if the operation may backlog
 * @cmlp:       callback which will be called
 * @data:       private data used by the caller
 */
static inline void acomp_request_set_callback(struct acomp_req *req,
                                              u32 flgs,
                                              crypto_completion_t cmpl,
                                              void *data)
{
        flgs &= ~CRYPTO_ACOMP_REQ_PRIVATE;
        flgs |= req->base.flags & CRYPTO_ACOMP_REQ_PRIVATE;
        crypto_request_set_callback(&req->base, flgs, cmpl, data);
}

/**
 * acomp_request_set_params() -- Sets request parameters
 *
 * Sets parameters required by an acomp operation
 *
 * @req:        asynchronous compress request
 * @src:        pointer to input buffer scatterlist
 * @dst:        pointer to output buffer scatterlist. If this is NULL, the
 *              acomp layer will allocate the output memory
 * @slen:       size of the input buffer
 * @dlen:       size of the output buffer. If dst is NULL, this can be used by
 *              the user to specify the maximum amount of memory to allocate
 */
static inline void acomp_request_set_params(struct acomp_req *req,
                                            struct scatterlist *src,
                                            struct scatterlist *dst,
                                            unsigned int slen,
                                            unsigned int dlen)
{
        req->src = src;
        req->dst = dst;
        req->slen = slen;
        req->dlen = dlen;

        req->base.flags &= ~(CRYPTO_ACOMP_REQ_SRC_VIRT |
                             CRYPTO_ACOMP_REQ_SRC_NONDMA |
                             CRYPTO_ACOMP_REQ_DST_VIRT |
                             CRYPTO_ACOMP_REQ_DST_NONDMA);
}

/**
 * acomp_request_set_src_sg() -- Sets source scatterlist
 *
 * Sets source scatterlist required by an acomp operation.
 *
 * @req:        asynchronous compress request
 * @src:        pointer to input buffer scatterlist
 * @slen:       size of the input buffer
 */
static inline void acomp_request_set_src_sg(struct acomp_req *req,
                                            struct scatterlist *src,
                                            unsigned int slen)
{
        req->src = src;
        req->slen = slen;

        req->base.flags &= ~CRYPTO_ACOMP_REQ_SRC_NONDMA;
        req->base.flags &= ~CRYPTO_ACOMP_REQ_SRC_VIRT;
}

/**
 * acomp_request_set_src_dma() -- Sets DMA source virtual address
 *
 * Sets source virtual address required by an acomp operation.
 * The address must be usable for DMA.
 *
 * @req:        asynchronous compress request
 * @src:        virtual address pointer to input buffer
 * @slen:       size of the input buffer
 */
static inline void acomp_request_set_src_dma(struct acomp_req *req,
                                             const u8 *src, unsigned int slen)
{
        req->svirt = src;
        req->slen = slen;

        req->base.flags &= ~CRYPTO_ACOMP_REQ_SRC_NONDMA;
        req->base.flags |= CRYPTO_ACOMP_REQ_SRC_VIRT;
}

/**
 * acomp_request_set_src_nondma() -- Sets non-DMA source virtual address
 *
 * Sets source virtual address required by an acomp operation.
 * The address can not be used for DMA.
 *
 * @req:        asynchronous compress request
 * @src:        virtual address pointer to input buffer
 * @slen:       size of the input buffer
 */
static inline void acomp_request_set_src_nondma(struct acomp_req *req,
                                                const u8 *src,
                                                unsigned int slen)
{
        req->svirt = src;
        req->slen = slen;

        req->base.flags |= CRYPTO_ACOMP_REQ_SRC_NONDMA;
        req->base.flags |= CRYPTO_ACOMP_REQ_SRC_VIRT;
}

/**
 * acomp_request_set_src_folio() -- Sets source folio
 *
 * Sets source folio required by an acomp operation.
 *
 * @req:        asynchronous compress request
 * @folio:      pointer to input folio
 * @off:        input folio offset
 * @len:        size of the input buffer
 */
static inline void acomp_request_set_src_folio(struct acomp_req *req,
                                               struct folio *folio, size_t off,
                                               unsigned int len)
{
        sg_init_table(&req->chain.ssg, 1);
        sg_set_folio(&req->chain.ssg, folio, len, off);
        acomp_request_set_src_sg(req, &req->chain.ssg, len);
}

/**
 * acomp_request_set_dst_sg() -- Sets destination scatterlist
 *
 * Sets destination scatterlist required by an acomp operation.
 *
 * @req:        asynchronous compress request
 * @dst:        pointer to output buffer scatterlist
 * @dlen:       size of the output buffer
 */
static inline void acomp_request_set_dst_sg(struct acomp_req *req,
                                            struct scatterlist *dst,
                                            unsigned int dlen)
{
        req->dst = dst;
        req->dlen = dlen;

        req->base.flags &= ~CRYPTO_ACOMP_REQ_DST_NONDMA;
        req->base.flags &= ~CRYPTO_ACOMP_REQ_DST_VIRT;
}

/**
 * acomp_request_set_dst_dma() -- Sets DMA destination virtual address
 *
 * Sets destination virtual address required by an acomp operation.
 * The address must be usable for DMA.
 *
 * @req:        asynchronous compress request
 * @dst:        virtual address pointer to output buffer
 * @dlen:       size of the output buffer
 */
static inline void acomp_request_set_dst_dma(struct acomp_req *req,
                                             u8 *dst, unsigned int dlen)
{
        req->dvirt = dst;
        req->dlen = dlen;

        req->base.flags &= ~CRYPTO_ACOMP_REQ_DST_NONDMA;
        req->base.flags |= CRYPTO_ACOMP_REQ_DST_VIRT;
}

/**
 * acomp_request_set_dst_nondma() -- Sets non-DMA destination virtual address
 *
 * Sets destination virtual address required by an acomp operation.
 * The address can not be used for DMA.
 *
 * @req:        asynchronous compress request
 * @dst:        virtual address pointer to output buffer
 * @dlen:       size of the output buffer
 */
static inline void acomp_request_set_dst_nondma(struct acomp_req *req,
                                                u8 *dst, unsigned int dlen)
{
        req->dvirt = dst;
        req->dlen = dlen;

        req->base.flags |= CRYPTO_ACOMP_REQ_DST_NONDMA;
        req->base.flags |= CRYPTO_ACOMP_REQ_DST_VIRT;
}

/**
 * acomp_request_set_dst_folio() -- Sets destination folio
 *
 * Sets destination folio required by an acomp operation.
 *
 * @req:        asynchronous compress request
 * @folio:      pointer to input folio
 * @off:        input folio offset
 * @len:        size of the input buffer
 */
static inline void acomp_request_set_dst_folio(struct acomp_req *req,
                                               struct folio *folio, size_t off,
                                               unsigned int len)
{
        sg_init_table(&req->chain.dsg, 1);
        sg_set_folio(&req->chain.dsg, folio, len, off);
        acomp_request_set_dst_sg(req, &req->chain.dsg, len);
}

/**
 * crypto_acomp_compress() -- Invoke asynchronous compress operation
 *
 * Function invokes the asynchronous compress operation
 *
 * @req:        asynchronous compress request
 *
 * Return:      zero on success; error code in case of error
 */
int crypto_acomp_compress(struct acomp_req *req);

/**
 * crypto_acomp_decompress() -- Invoke asynchronous decompress operation
 *
 * Function invokes the asynchronous decompress operation
 *
 * @req:        asynchronous compress request
 *
 * Return:      zero on success; error code in case of error
 */
int crypto_acomp_decompress(struct acomp_req *req);

static inline struct acomp_req *acomp_request_on_stack_init(
        char *buf, struct crypto_acomp *tfm)
{
        struct acomp_req *req = (void *)buf;

        crypto_stack_request_init(&req->base, crypto_acomp_tfm(tfm));
        return req;
}

struct acomp_req *acomp_request_clone(struct acomp_req *req,
                                      size_t total, gfp_t gfp);

#endif