root/drivers/crypto/hifn_795x.c
// SPDX-License-Identifier: GPL-2.0-or-later
/*
 * 2007+ Copyright (c) Evgeniy Polyakov <johnpol@2ka.mipt.ru>
 * All rights reserved.
 */

#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/moduleparam.h>
#include <linux/mod_devicetable.h>
#include <linux/interrupt.h>
#include <linux/pci.h>
#include <linux/slab.h>
#include <linux/delay.h>
#include <linux/mm.h>
#include <linux/dma-mapping.h>
#include <linux/scatterlist.h>
#include <linux/highmem.h>
#include <linux/crypto.h>
#include <linux/hw_random.h>
#include <linux/ktime.h>

#include <crypto/algapi.h>
#include <crypto/internal/des.h>
#include <crypto/internal/skcipher.h>

static char hifn_pll_ref[sizeof("extNNN")] = "ext";
module_param_string(hifn_pll_ref, hifn_pll_ref, sizeof(hifn_pll_ref), 0444);
MODULE_PARM_DESC(hifn_pll_ref,
                 "PLL reference clock (pci[freq] or ext[freq], default ext)");

static atomic_t hifn_dev_number;

#define ACRYPTO_OP_DECRYPT      0
#define ACRYPTO_OP_ENCRYPT      1
#define ACRYPTO_OP_HMAC         2
#define ACRYPTO_OP_RNG          3

#define ACRYPTO_MODE_ECB                0
#define ACRYPTO_MODE_CBC                1
#define ACRYPTO_MODE_CFB                2
#define ACRYPTO_MODE_OFB                3

#define ACRYPTO_TYPE_AES_128    0
#define ACRYPTO_TYPE_AES_192    1
#define ACRYPTO_TYPE_AES_256    2
#define ACRYPTO_TYPE_3DES       3
#define ACRYPTO_TYPE_DES        4

#define PCI_VENDOR_ID_HIFN              0x13A3
#define PCI_DEVICE_ID_HIFN_7955         0x0020
#define PCI_DEVICE_ID_HIFN_7956         0x001d

/* I/O region sizes */

#define HIFN_BAR0_SIZE                  0x1000
#define HIFN_BAR1_SIZE                  0x2000
#define HIFN_BAR2_SIZE                  0x8000

/* DMA registres */

#define HIFN_DMA_CRA                    0x0C    /* DMA Command Ring Address */
#define HIFN_DMA_SDRA                   0x1C    /* DMA Source Data Ring Address */
#define HIFN_DMA_RRA                    0x2C    /* DMA Result Ring Address */
#define HIFN_DMA_DDRA                   0x3C    /* DMA Destination Data Ring Address */
#define HIFN_DMA_STCTL                  0x40    /* DMA Status and Control */
#define HIFN_DMA_INTREN                 0x44    /* DMA Interrupt Enable */
#define HIFN_DMA_CFG1                   0x48    /* DMA Configuration #1 */
#define HIFN_DMA_CFG2                   0x6C    /* DMA Configuration #2 */
#define HIFN_CHIP_ID                    0x98    /* Chip ID */

/*
 * Processing Unit Registers (offset from BASEREG0)
 */
#define HIFN_0_PUDATA           0x00    /* Processing Unit Data */
#define HIFN_0_PUCTRL           0x04    /* Processing Unit Control */
#define HIFN_0_PUISR            0x08    /* Processing Unit Interrupt Status */
#define HIFN_0_PUCNFG           0x0c    /* Processing Unit Configuration */
#define HIFN_0_PUIER            0x10    /* Processing Unit Interrupt Enable */
#define HIFN_0_PUSTAT           0x14    /* Processing Unit Status/Chip ID */
#define HIFN_0_FIFOSTAT         0x18    /* FIFO Status */
#define HIFN_0_FIFOCNFG         0x1c    /* FIFO Configuration */
#define HIFN_0_SPACESIZE        0x20    /* Register space size */

/* Processing Unit Control Register (HIFN_0_PUCTRL) */
#define HIFN_PUCTRL_CLRSRCFIFO  0x0010  /* clear source fifo */
#define HIFN_PUCTRL_STOP        0x0008  /* stop pu */
#define HIFN_PUCTRL_LOCKRAM     0x0004  /* lock ram */
#define HIFN_PUCTRL_DMAENA      0x0002  /* enable dma */
#define HIFN_PUCTRL_RESET       0x0001  /* Reset processing unit */

/* Processing Unit Interrupt Status Register (HIFN_0_PUISR) */
#define HIFN_PUISR_CMDINVAL     0x8000  /* Invalid command interrupt */
#define HIFN_PUISR_DATAERR      0x4000  /* Data error interrupt */
#define HIFN_PUISR_SRCFIFO      0x2000  /* Source FIFO ready interrupt */
#define HIFN_PUISR_DSTFIFO      0x1000  /* Destination FIFO ready interrupt */
#define HIFN_PUISR_DSTOVER      0x0200  /* Destination overrun interrupt */
#define HIFN_PUISR_SRCCMD       0x0080  /* Source command interrupt */
#define HIFN_PUISR_SRCCTX       0x0040  /* Source context interrupt */
#define HIFN_PUISR_SRCDATA      0x0020  /* Source data interrupt */
#define HIFN_PUISR_DSTDATA      0x0010  /* Destination data interrupt */
#define HIFN_PUISR_DSTRESULT    0x0004  /* Destination result interrupt */

/* Processing Unit Configuration Register (HIFN_0_PUCNFG) */
#define HIFN_PUCNFG_DRAMMASK    0xe000  /* DRAM size mask */
#define HIFN_PUCNFG_DSZ_256K    0x0000  /* 256k dram */
#define HIFN_PUCNFG_DSZ_512K    0x2000  /* 512k dram */
#define HIFN_PUCNFG_DSZ_1M      0x4000  /* 1m dram */
#define HIFN_PUCNFG_DSZ_2M      0x6000  /* 2m dram */
#define HIFN_PUCNFG_DSZ_4M      0x8000  /* 4m dram */
#define HIFN_PUCNFG_DSZ_8M      0xa000  /* 8m dram */
#define HIFN_PUNCFG_DSZ_16M     0xc000  /* 16m dram */
#define HIFN_PUCNFG_DSZ_32M     0xe000  /* 32m dram */
#define HIFN_PUCNFG_DRAMREFRESH 0x1800  /* DRAM refresh rate mask */
#define HIFN_PUCNFG_DRFR_512    0x0000  /* 512 divisor of ECLK */
#define HIFN_PUCNFG_DRFR_256    0x0800  /* 256 divisor of ECLK */
#define HIFN_PUCNFG_DRFR_128    0x1000  /* 128 divisor of ECLK */
#define HIFN_PUCNFG_TCALLPHASES 0x0200  /* your guess is as good as mine... */
#define HIFN_PUCNFG_TCDRVTOTEM  0x0100  /* your guess is as good as mine... */
#define HIFN_PUCNFG_BIGENDIAN   0x0080  /* DMA big endian mode */
#define HIFN_PUCNFG_BUS32       0x0040  /* Bus width 32bits */
#define HIFN_PUCNFG_BUS16       0x0000  /* Bus width 16 bits */
#define HIFN_PUCNFG_CHIPID      0x0020  /* Allow chipid from PUSTAT */
#define HIFN_PUCNFG_DRAM        0x0010  /* Context RAM is DRAM */
#define HIFN_PUCNFG_SRAM        0x0000  /* Context RAM is SRAM */
#define HIFN_PUCNFG_COMPSING    0x0004  /* Enable single compression context */
#define HIFN_PUCNFG_ENCCNFG     0x0002  /* Encryption configuration */

/* Processing Unit Interrupt Enable Register (HIFN_0_PUIER) */
#define HIFN_PUIER_CMDINVAL     0x8000  /* Invalid command interrupt */
#define HIFN_PUIER_DATAERR      0x4000  /* Data error interrupt */
#define HIFN_PUIER_SRCFIFO      0x2000  /* Source FIFO ready interrupt */
#define HIFN_PUIER_DSTFIFO      0x1000  /* Destination FIFO ready interrupt */
#define HIFN_PUIER_DSTOVER      0x0200  /* Destination overrun interrupt */
#define HIFN_PUIER_SRCCMD       0x0080  /* Source command interrupt */
#define HIFN_PUIER_SRCCTX       0x0040  /* Source context interrupt */
#define HIFN_PUIER_SRCDATA      0x0020  /* Source data interrupt */
#define HIFN_PUIER_DSTDATA      0x0010  /* Destination data interrupt */
#define HIFN_PUIER_DSTRESULT    0x0004  /* Destination result interrupt */

/* Processing Unit Status Register/Chip ID (HIFN_0_PUSTAT) */
#define HIFN_PUSTAT_CMDINVAL    0x8000  /* Invalid command interrupt */
#define HIFN_PUSTAT_DATAERR     0x4000  /* Data error interrupt */
#define HIFN_PUSTAT_SRCFIFO     0x2000  /* Source FIFO ready interrupt */
#define HIFN_PUSTAT_DSTFIFO     0x1000  /* Destination FIFO ready interrupt */
#define HIFN_PUSTAT_DSTOVER     0x0200  /* Destination overrun interrupt */
#define HIFN_PUSTAT_SRCCMD      0x0080  /* Source command interrupt */
#define HIFN_PUSTAT_SRCCTX      0x0040  /* Source context interrupt */
#define HIFN_PUSTAT_SRCDATA     0x0020  /* Source data interrupt */
#define HIFN_PUSTAT_DSTDATA     0x0010  /* Destination data interrupt */
#define HIFN_PUSTAT_DSTRESULT   0x0004  /* Destination result interrupt */
#define HIFN_PUSTAT_CHIPREV     0x00ff  /* Chip revision mask */
#define HIFN_PUSTAT_CHIPENA     0xff00  /* Chip enabled mask */
#define HIFN_PUSTAT_ENA_2       0x1100  /* Level 2 enabled */
#define HIFN_PUSTAT_ENA_1       0x1000  /* Level 1 enabled */
#define HIFN_PUSTAT_ENA_0       0x3000  /* Level 0 enabled */
#define HIFN_PUSTAT_REV_2       0x0020  /* 7751 PT6/2 */
#define HIFN_PUSTAT_REV_3       0x0030  /* 7751 PT6/3 */

/* FIFO Status Register (HIFN_0_FIFOSTAT) */
#define HIFN_FIFOSTAT_SRC       0x7f00  /* Source FIFO available */
#define HIFN_FIFOSTAT_DST       0x007f  /* Destination FIFO available */

/* FIFO Configuration Register (HIFN_0_FIFOCNFG) */
#define HIFN_FIFOCNFG_THRESHOLD 0x0400  /* must be written as 1 */

/*
 * DMA Interface Registers (offset from BASEREG1)
 */
#define HIFN_1_DMA_CRAR         0x0c    /* DMA Command Ring Address */
#define HIFN_1_DMA_SRAR         0x1c    /* DMA Source Ring Address */
#define HIFN_1_DMA_RRAR         0x2c    /* DMA Result Ring Address */
#define HIFN_1_DMA_DRAR         0x3c    /* DMA Destination Ring Address */
#define HIFN_1_DMA_CSR          0x40    /* DMA Status and Control */
#define HIFN_1_DMA_IER          0x44    /* DMA Interrupt Enable */
#define HIFN_1_DMA_CNFG         0x48    /* DMA Configuration */
#define HIFN_1_PLL              0x4c    /* 795x: PLL config */
#define HIFN_1_7811_RNGENA      0x60    /* 7811: rng enable */
#define HIFN_1_7811_RNGCFG      0x64    /* 7811: rng config */
#define HIFN_1_7811_RNGDAT      0x68    /* 7811: rng data */
#define HIFN_1_7811_RNGSTS      0x6c    /* 7811: rng status */
#define HIFN_1_7811_MIPSRST     0x94    /* 7811: MIPS reset */
#define HIFN_1_REVID            0x98    /* Revision ID */
#define HIFN_1_UNLOCK_SECRET1   0xf4
#define HIFN_1_UNLOCK_SECRET2   0xfc
#define HIFN_1_PUB_RESET        0x204   /* Public/RNG Reset */
#define HIFN_1_PUB_BASE         0x300   /* Public Base Address */
#define HIFN_1_PUB_OPLEN        0x304   /* Public Operand Length */
#define HIFN_1_PUB_OP           0x308   /* Public Operand */
#define HIFN_1_PUB_STATUS       0x30c   /* Public Status */
#define HIFN_1_PUB_IEN          0x310   /* Public Interrupt enable */
#define HIFN_1_RNG_CONFIG       0x314   /* RNG config */
#define HIFN_1_RNG_DATA         0x318   /* RNG data */
#define HIFN_1_PUB_MEM          0x400   /* start of Public key memory */
#define HIFN_1_PUB_MEMEND       0xbff   /* end of Public key memory */

/* DMA Status and Control Register (HIFN_1_DMA_CSR) */
#define HIFN_DMACSR_D_CTRLMASK  0xc0000000      /* Destinition Ring Control */
#define HIFN_DMACSR_D_CTRL_NOP  0x00000000      /* Dest. Control: no-op */
#define HIFN_DMACSR_D_CTRL_DIS  0x40000000      /* Dest. Control: disable */
#define HIFN_DMACSR_D_CTRL_ENA  0x80000000      /* Dest. Control: enable */
#define HIFN_DMACSR_D_ABORT     0x20000000      /* Destinition Ring PCIAbort */
#define HIFN_DMACSR_D_DONE      0x10000000      /* Destinition Ring Done */
#define HIFN_DMACSR_D_LAST      0x08000000      /* Destinition Ring Last */
#define HIFN_DMACSR_D_WAIT      0x04000000      /* Destinition Ring Waiting */
#define HIFN_DMACSR_D_OVER      0x02000000      /* Destinition Ring Overflow */
#define HIFN_DMACSR_R_CTRL      0x00c00000      /* Result Ring Control */
#define HIFN_DMACSR_R_CTRL_NOP  0x00000000      /* Result Control: no-op */
#define HIFN_DMACSR_R_CTRL_DIS  0x00400000      /* Result Control: disable */
#define HIFN_DMACSR_R_CTRL_ENA  0x00800000      /* Result Control: enable */
#define HIFN_DMACSR_R_ABORT     0x00200000      /* Result Ring PCI Abort */
#define HIFN_DMACSR_R_DONE      0x00100000      /* Result Ring Done */
#define HIFN_DMACSR_R_LAST      0x00080000      /* Result Ring Last */
#define HIFN_DMACSR_R_WAIT      0x00040000      /* Result Ring Waiting */
#define HIFN_DMACSR_R_OVER      0x00020000      /* Result Ring Overflow */
#define HIFN_DMACSR_S_CTRL      0x0000c000      /* Source Ring Control */
#define HIFN_DMACSR_S_CTRL_NOP  0x00000000      /* Source Control: no-op */
#define HIFN_DMACSR_S_CTRL_DIS  0x00004000      /* Source Control: disable */
#define HIFN_DMACSR_S_CTRL_ENA  0x00008000      /* Source Control: enable */
#define HIFN_DMACSR_S_ABORT     0x00002000      /* Source Ring PCI Abort */
#define HIFN_DMACSR_S_DONE      0x00001000      /* Source Ring Done */
#define HIFN_DMACSR_S_LAST      0x00000800      /* Source Ring Last */
#define HIFN_DMACSR_S_WAIT      0x00000400      /* Source Ring Waiting */
#define HIFN_DMACSR_ILLW        0x00000200      /* Illegal write (7811 only) */
#define HIFN_DMACSR_ILLR        0x00000100      /* Illegal read (7811 only) */
#define HIFN_DMACSR_C_CTRL      0x000000c0      /* Command Ring Control */
#define HIFN_DMACSR_C_CTRL_NOP  0x00000000      /* Command Control: no-op */
#define HIFN_DMACSR_C_CTRL_DIS  0x00000040      /* Command Control: disable */
#define HIFN_DMACSR_C_CTRL_ENA  0x00000080      /* Command Control: enable */
#define HIFN_DMACSR_C_ABORT     0x00000020      /* Command Ring PCI Abort */
#define HIFN_DMACSR_C_DONE      0x00000010      /* Command Ring Done */
#define HIFN_DMACSR_C_LAST      0x00000008      /* Command Ring Last */
#define HIFN_DMACSR_C_WAIT      0x00000004      /* Command Ring Waiting */
#define HIFN_DMACSR_PUBDONE     0x00000002      /* Public op done (7951 only) */
#define HIFN_DMACSR_ENGINE      0x00000001      /* Command Ring Engine IRQ */

/* DMA Interrupt Enable Register (HIFN_1_DMA_IER) */
#define HIFN_DMAIER_D_ABORT     0x20000000      /* Destination Ring PCIAbort */
#define HIFN_DMAIER_D_DONE      0x10000000      /* Destination Ring Done */
#define HIFN_DMAIER_D_LAST      0x08000000      /* Destination Ring Last */
#define HIFN_DMAIER_D_WAIT      0x04000000      /* Destination Ring Waiting */
#define HIFN_DMAIER_D_OVER      0x02000000      /* Destination Ring Overflow */
#define HIFN_DMAIER_R_ABORT     0x00200000      /* Result Ring PCI Abort */
#define HIFN_DMAIER_R_DONE      0x00100000      /* Result Ring Done */
#define HIFN_DMAIER_R_LAST      0x00080000      /* Result Ring Last */
#define HIFN_DMAIER_R_WAIT      0x00040000      /* Result Ring Waiting */
#define HIFN_DMAIER_R_OVER      0x00020000      /* Result Ring Overflow */
#define HIFN_DMAIER_S_ABORT     0x00002000      /* Source Ring PCI Abort */
#define HIFN_DMAIER_S_DONE      0x00001000      /* Source Ring Done */
#define HIFN_DMAIER_S_LAST      0x00000800      /* Source Ring Last */
#define HIFN_DMAIER_S_WAIT      0x00000400      /* Source Ring Waiting */
#define HIFN_DMAIER_ILLW        0x00000200      /* Illegal write (7811 only) */
#define HIFN_DMAIER_ILLR        0x00000100      /* Illegal read (7811 only) */
#define HIFN_DMAIER_C_ABORT     0x00000020      /* Command Ring PCI Abort */
#define HIFN_DMAIER_C_DONE      0x00000010      /* Command Ring Done */
#define HIFN_DMAIER_C_LAST      0x00000008      /* Command Ring Last */
#define HIFN_DMAIER_C_WAIT      0x00000004      /* Command Ring Waiting */
#define HIFN_DMAIER_PUBDONE     0x00000002      /* public op done (7951 only) */
#define HIFN_DMAIER_ENGINE      0x00000001      /* Engine IRQ */

/* DMA Configuration Register (HIFN_1_DMA_CNFG) */
#define HIFN_DMACNFG_BIGENDIAN  0x10000000      /* big endian mode */
#define HIFN_DMACNFG_POLLFREQ   0x00ff0000      /* Poll frequency mask */
#define HIFN_DMACNFG_UNLOCK     0x00000800
#define HIFN_DMACNFG_POLLINVAL  0x00000700      /* Invalid Poll Scalar */
#define HIFN_DMACNFG_LAST       0x00000010      /* Host control LAST bit */
#define HIFN_DMACNFG_MODE       0x00000004      /* DMA mode */
#define HIFN_DMACNFG_DMARESET   0x00000002      /* DMA Reset # */
#define HIFN_DMACNFG_MSTRESET   0x00000001      /* Master Reset # */

/* PLL configuration register */
#define HIFN_PLL_REF_CLK_HBI    0x00000000      /* HBI reference clock */
#define HIFN_PLL_REF_CLK_PLL    0x00000001      /* PLL reference clock */
#define HIFN_PLL_BP             0x00000002      /* Reference clock bypass */
#define HIFN_PLL_PK_CLK_HBI     0x00000000      /* PK engine HBI clock */
#define HIFN_PLL_PK_CLK_PLL     0x00000008      /* PK engine PLL clock */
#define HIFN_PLL_PE_CLK_HBI     0x00000000      /* PE engine HBI clock */
#define HIFN_PLL_PE_CLK_PLL     0x00000010      /* PE engine PLL clock */
#define HIFN_PLL_RESERVED_1     0x00000400      /* Reserved bit, must be 1 */
#define HIFN_PLL_ND_SHIFT       11              /* Clock multiplier shift */
#define HIFN_PLL_ND_MULT_2      0x00000000      /* PLL clock multiplier 2 */
#define HIFN_PLL_ND_MULT_4      0x00000800      /* PLL clock multiplier 4 */
#define HIFN_PLL_ND_MULT_6      0x00001000      /* PLL clock multiplier 6 */
#define HIFN_PLL_ND_MULT_8      0x00001800      /* PLL clock multiplier 8 */
#define HIFN_PLL_ND_MULT_10     0x00002000      /* PLL clock multiplier 10 */
#define HIFN_PLL_ND_MULT_12     0x00002800      /* PLL clock multiplier 12 */
#define HIFN_PLL_IS_1_8         0x00000000      /* charge pump (mult. 1-8) */
#define HIFN_PLL_IS_9_12        0x00010000      /* charge pump (mult. 9-12) */

#define HIFN_PLL_FCK_MAX        266             /* Maximum PLL frequency */

/* Public key reset register (HIFN_1_PUB_RESET) */
#define HIFN_PUBRST_RESET       0x00000001      /* reset public/rng unit */

/* Public base address register (HIFN_1_PUB_BASE) */
#define HIFN_PUBBASE_ADDR       0x00003fff      /* base address */

/* Public operand length register (HIFN_1_PUB_OPLEN) */
#define HIFN_PUBOPLEN_MOD_M     0x0000007f      /* modulus length mask */
#define HIFN_PUBOPLEN_MOD_S     0               /* modulus length shift */
#define HIFN_PUBOPLEN_EXP_M     0x0003ff80      /* exponent length mask */
#define HIFN_PUBOPLEN_EXP_S     7               /* exponent length shift */
#define HIFN_PUBOPLEN_RED_M     0x003c0000      /* reducend length mask */
#define HIFN_PUBOPLEN_RED_S     18              /* reducend length shift */

/* Public operation register (HIFN_1_PUB_OP) */
#define HIFN_PUBOP_AOFFSET_M    0x0000007f      /* A offset mask */
#define HIFN_PUBOP_AOFFSET_S    0               /* A offset shift */
#define HIFN_PUBOP_BOFFSET_M    0x00000f80      /* B offset mask */
#define HIFN_PUBOP_BOFFSET_S    7               /* B offset shift */
#define HIFN_PUBOP_MOFFSET_M    0x0003f000      /* M offset mask */
#define HIFN_PUBOP_MOFFSET_S    12              /* M offset shift */
#define HIFN_PUBOP_OP_MASK      0x003c0000      /* Opcode: */
#define HIFN_PUBOP_OP_NOP       0x00000000      /*  NOP */
#define HIFN_PUBOP_OP_ADD       0x00040000      /*  ADD */
#define HIFN_PUBOP_OP_ADDC      0x00080000      /*  ADD w/carry */
#define HIFN_PUBOP_OP_SUB       0x000c0000      /*  SUB */
#define HIFN_PUBOP_OP_SUBC      0x00100000      /*  SUB w/carry */
#define HIFN_PUBOP_OP_MODADD    0x00140000      /*  Modular ADD */
#define HIFN_PUBOP_OP_MODSUB    0x00180000      /*  Modular SUB */
#define HIFN_PUBOP_OP_INCA      0x001c0000      /*  INC A */
#define HIFN_PUBOP_OP_DECA      0x00200000      /*  DEC A */
#define HIFN_PUBOP_OP_MULT      0x00240000      /*  MULT */
#define HIFN_PUBOP_OP_MODMULT   0x00280000      /*  Modular MULT */
#define HIFN_PUBOP_OP_MODRED    0x002c0000      /*  Modular RED */
#define HIFN_PUBOP_OP_MODEXP    0x00300000      /*  Modular EXP */

/* Public status register (HIFN_1_PUB_STATUS) */
#define HIFN_PUBSTS_DONE        0x00000001      /* operation done */
#define HIFN_PUBSTS_CARRY       0x00000002      /* carry */

/* Public interrupt enable register (HIFN_1_PUB_IEN) */
#define HIFN_PUBIEN_DONE        0x00000001      /* operation done interrupt */

/* Random number generator config register (HIFN_1_RNG_CONFIG) */
#define HIFN_RNGCFG_ENA         0x00000001      /* enable rng */

#define HIFN_NAMESIZE                   32
#define HIFN_MAX_RESULT_ORDER           5

#define HIFN_D_CMD_RSIZE                (24 * 1)
#define HIFN_D_SRC_RSIZE                (80 * 1)
#define HIFN_D_DST_RSIZE                (80 * 1)
#define HIFN_D_RES_RSIZE                (24 * 1)

#define HIFN_D_DST_DALIGN               4

#define HIFN_QUEUE_LENGTH               (HIFN_D_CMD_RSIZE - 1)

#define AES_MIN_KEY_SIZE                16
#define AES_MAX_KEY_SIZE                32

#define HIFN_DES_KEY_LENGTH             8
#define HIFN_3DES_KEY_LENGTH            24
#define HIFN_MAX_CRYPT_KEY_LENGTH       AES_MAX_KEY_SIZE
#define HIFN_IV_LENGTH                  8
#define HIFN_AES_IV_LENGTH              16
#define HIFN_MAX_IV_LENGTH              HIFN_AES_IV_LENGTH

#define HIFN_MAC_KEY_LENGTH             64
#define HIFN_MD5_LENGTH                 16
#define HIFN_SHA1_LENGTH                20
#define HIFN_MAC_TRUNC_LENGTH           12

#define HIFN_MAX_COMMAND                (8 + 8 + 8 + 64 + 260)
#define HIFN_MAX_RESULT                 (8 + 4 + 4 + 20 + 4)
#define HIFN_USED_RESULT                12

struct hifn_desc {
        volatile __le32         l;
        volatile __le32         p;
};

struct hifn_dma {
        struct hifn_desc        cmdr[HIFN_D_CMD_RSIZE + 1];
        struct hifn_desc        srcr[HIFN_D_SRC_RSIZE + 1];
        struct hifn_desc        dstr[HIFN_D_DST_RSIZE + 1];
        struct hifn_desc        resr[HIFN_D_RES_RSIZE + 1];

        u8                      command_bufs[HIFN_D_CMD_RSIZE][HIFN_MAX_COMMAND];
        u8                      result_bufs[HIFN_D_CMD_RSIZE][HIFN_MAX_RESULT];

        /*
         *  Our current positions for insertion and removal from the descriptor
         *  rings.
         */
        volatile int            cmdi, srci, dsti, resi;
        volatile int            cmdu, srcu, dstu, resu;
        int                     cmdk, srck, dstk, resk;
};

#define HIFN_FLAG_CMD_BUSY      (1 << 0)
#define HIFN_FLAG_SRC_BUSY      (1 << 1)
#define HIFN_FLAG_DST_BUSY      (1 << 2)
#define HIFN_FLAG_RES_BUSY      (1 << 3)
#define HIFN_FLAG_OLD_KEY       (1 << 4)

#define HIFN_DEFAULT_ACTIVE_NUM 5

struct hifn_device {
        char                    name[HIFN_NAMESIZE];

        int                     irq;

        struct pci_dev          *pdev;
        void __iomem            *bar[3];

        void                    *desc_virt;
        dma_addr_t              desc_dma;

        u32                     dmareg;

        void                    *sa[HIFN_D_RES_RSIZE];

        spinlock_t              lock;

        u32                     flags;
        int                     active, started;
        struct delayed_work     work;
        unsigned long           reset;
        unsigned long           success;
        unsigned long           prev_success;

        u8                      snum;

        struct tasklet_struct   tasklet;

        struct crypto_queue     queue;
        struct list_head        alg_list;

        unsigned int            pk_clk_freq;

#ifdef CONFIG_CRYPTO_DEV_HIFN_795X_RNG
        unsigned int            rng_wait_time;
        ktime_t                 rngtime;
        struct hwrng            rng;
#endif
};

#define HIFN_D_LENGTH                   0x0000ffff
#define HIFN_D_NOINVALID                0x01000000
#define HIFN_D_MASKDONEIRQ              0x02000000
#define HIFN_D_DESTOVER                 0x04000000
#define HIFN_D_OVER                     0x08000000
#define HIFN_D_LAST                     0x20000000
#define HIFN_D_JUMP                     0x40000000
#define HIFN_D_VALID                    0x80000000

struct hifn_base_command {
        volatile __le16         masks;
        volatile __le16         session_num;
        volatile __le16         total_source_count;
        volatile __le16         total_dest_count;
};

#define HIFN_BASE_CMD_COMP              0x0100  /* enable compression engine */
#define HIFN_BASE_CMD_PAD               0x0200  /* enable padding engine */
#define HIFN_BASE_CMD_MAC               0x0400  /* enable MAC engine */
#define HIFN_BASE_CMD_CRYPT             0x0800  /* enable crypt engine */
#define HIFN_BASE_CMD_DECODE            0x2000
#define HIFN_BASE_CMD_SRCLEN_M          0xc000
#define HIFN_BASE_CMD_SRCLEN_S          14
#define HIFN_BASE_CMD_DSTLEN_M          0x3000
#define HIFN_BASE_CMD_DSTLEN_S          12
#define HIFN_BASE_CMD_LENMASK_HI        0x30000
#define HIFN_BASE_CMD_LENMASK_LO        0x0ffff

/*
 * Structure to help build up the command data structure.
 */
struct hifn_crypt_command {
        volatile __le16         masks;
        volatile __le16         header_skip;
        volatile __le16         source_count;
        volatile __le16         reserved;
};

#define HIFN_CRYPT_CMD_ALG_MASK         0x0003          /* algorithm: */
#define HIFN_CRYPT_CMD_ALG_DES          0x0000          /*   DES */
#define HIFN_CRYPT_CMD_ALG_3DES         0x0001          /*   3DES */
#define HIFN_CRYPT_CMD_ALG_RC4          0x0002          /*   RC4 */
#define HIFN_CRYPT_CMD_ALG_AES          0x0003          /*   AES */
#define HIFN_CRYPT_CMD_MODE_MASK        0x0018          /* Encrypt mode: */
#define HIFN_CRYPT_CMD_MODE_ECB         0x0000          /*   ECB */
#define HIFN_CRYPT_CMD_MODE_CBC         0x0008          /*   CBC */
#define HIFN_CRYPT_CMD_MODE_CFB         0x0010          /*   CFB */
#define HIFN_CRYPT_CMD_MODE_OFB         0x0018          /*   OFB */
#define HIFN_CRYPT_CMD_CLR_CTX          0x0040          /* clear context */
#define HIFN_CRYPT_CMD_KSZ_MASK         0x0600          /* AES key size: */
#define HIFN_CRYPT_CMD_KSZ_128          0x0000          /*  128 bit */
#define HIFN_CRYPT_CMD_KSZ_192          0x0200          /*  192 bit */
#define HIFN_CRYPT_CMD_KSZ_256          0x0400          /*  256 bit */
#define HIFN_CRYPT_CMD_NEW_KEY          0x0800          /* expect new key */
#define HIFN_CRYPT_CMD_NEW_IV           0x1000          /* expect new iv */
#define HIFN_CRYPT_CMD_SRCLEN_M         0xc000
#define HIFN_CRYPT_CMD_SRCLEN_S         14

#define HIFN_MAC_CMD_ALG_MASK           0x0001
#define HIFN_MAC_CMD_ALG_SHA1           0x0000
#define HIFN_MAC_CMD_ALG_MD5            0x0001
#define HIFN_MAC_CMD_MODE_MASK          0x000c
#define HIFN_MAC_CMD_MODE_HMAC          0x0000
#define HIFN_MAC_CMD_MODE_SSL_MAC       0x0004
#define HIFN_MAC_CMD_MODE_HASH          0x0008
#define HIFN_MAC_CMD_MODE_FULL          0x0004
#define HIFN_MAC_CMD_TRUNC              0x0010
#define HIFN_MAC_CMD_RESULT             0x0020
#define HIFN_MAC_CMD_APPEND             0x0040
#define HIFN_MAC_CMD_SRCLEN_M           0xc000
#define HIFN_MAC_CMD_SRCLEN_S           14

/*
 * MAC POS IPsec initiates authentication after encryption on encodes
 * and before decryption on decodes.
 */
#define HIFN_MAC_CMD_POS_IPSEC          0x0200
#define HIFN_MAC_CMD_NEW_KEY            0x0800

#define HIFN_COMP_CMD_SRCLEN_M          0xc000
#define HIFN_COMP_CMD_SRCLEN_S          14
#define HIFN_COMP_CMD_ONE               0x0100  /* must be one */
#define HIFN_COMP_CMD_CLEARHIST         0x0010  /* clear history */
#define HIFN_COMP_CMD_UPDATEHIST        0x0008  /* update history */
#define HIFN_COMP_CMD_LZS_STRIP0        0x0004  /* LZS: strip zero */
#define HIFN_COMP_CMD_MPPC_RESTART      0x0004  /* MPPC: restart */
#define HIFN_COMP_CMD_ALG_MASK          0x0001  /* compression mode: */
#define HIFN_COMP_CMD_ALG_MPPC          0x0001  /*   MPPC */
#define HIFN_COMP_CMD_ALG_LZS           0x0000  /*   LZS */

struct hifn_base_result {
        volatile __le16         flags;
        volatile __le16         session;
        volatile __le16         src_cnt;                /* 15:0 of source count */
        volatile __le16         dst_cnt;                /* 15:0 of dest count */
};

#define HIFN_BASE_RES_DSTOVERRUN        0x0200  /* destination overrun */
#define HIFN_BASE_RES_SRCLEN_M          0xc000  /* 17:16 of source count */
#define HIFN_BASE_RES_SRCLEN_S          14
#define HIFN_BASE_RES_DSTLEN_M          0x3000  /* 17:16 of dest count */
#define HIFN_BASE_RES_DSTLEN_S          12

struct hifn_comp_result {
        volatile __le16         flags;
        volatile __le16         crc;
};

#define HIFN_COMP_RES_LCB_M             0xff00  /* longitudinal check byte */
#define HIFN_COMP_RES_LCB_S             8
#define HIFN_COMP_RES_RESTART           0x0004  /* MPPC: restart */
#define HIFN_COMP_RES_ENDMARKER         0x0002  /* LZS: end marker seen */
#define HIFN_COMP_RES_SRC_NOTZERO       0x0001  /* source expired */

struct hifn_mac_result {
        volatile __le16         flags;
        volatile __le16         reserved;
        /* followed by 0, 6, 8, or 10 u16's of the MAC, then crypt */
};

#define HIFN_MAC_RES_MISCOMPARE         0x0002  /* compare failed */
#define HIFN_MAC_RES_SRC_NOTZERO        0x0001  /* source expired */

struct hifn_crypt_result {
        volatile __le16         flags;
        volatile __le16         reserved;
};

#define HIFN_CRYPT_RES_SRC_NOTZERO      0x0001  /* source expired */

#ifndef HIFN_POLL_FREQUENCY
#define HIFN_POLL_FREQUENCY     0x1
#endif

#ifndef HIFN_POLL_SCALAR
#define HIFN_POLL_SCALAR        0x0
#endif

#define HIFN_MAX_SEGLEN         0xffff          /* maximum dma segment len */
#define HIFN_MAX_DMALEN         0x3ffff         /* maximum dma length */

struct hifn_crypto_alg {
        struct list_head        entry;
        struct skcipher_alg     alg;
        struct hifn_device      *dev;
};

#define ASYNC_SCATTERLIST_CACHE 16

#define ASYNC_FLAGS_MISALIGNED  (1 << 0)

struct hifn_cipher_walk {
        struct scatterlist      cache[ASYNC_SCATTERLIST_CACHE];
        u32                     flags;
        int                     num;
};

struct hifn_context {
        u8                      key[HIFN_MAX_CRYPT_KEY_LENGTH];
        struct hifn_device      *dev;
        unsigned int            keysize;
};

struct hifn_request_context {
        u8                      *iv;
        unsigned int            ivsize;
        u8                      op, type, mode, unused;
        struct hifn_cipher_walk walk;
};

#define crypto_alg_to_hifn(a)   container_of(a, struct hifn_crypto_alg, alg)

static inline u32 hifn_read_0(struct hifn_device *dev, u32 reg)
{
        return readl(dev->bar[0] + reg);
}

static inline u32 hifn_read_1(struct hifn_device *dev, u32 reg)
{
        return readl(dev->bar[1] + reg);
}

static inline void hifn_write_0(struct hifn_device *dev, u32 reg, u32 val)
{
        writel((__force u32)cpu_to_le32(val), dev->bar[0] + reg);
}

static inline void hifn_write_1(struct hifn_device *dev, u32 reg, u32 val)
{
        writel((__force u32)cpu_to_le32(val), dev->bar[1] + reg);
}

static void hifn_wait_puc(struct hifn_device *dev)
{
        int i;
        u32 ret;

        for (i = 10000; i > 0; --i) {
                ret = hifn_read_0(dev, HIFN_0_PUCTRL);
                if (!(ret & HIFN_PUCTRL_RESET))
                        break;

                udelay(1);
        }

        if (!i)
                dev_err(&dev->pdev->dev, "Failed to reset PUC unit.\n");
}

static void hifn_reset_puc(struct hifn_device *dev)
{
        hifn_write_0(dev, HIFN_0_PUCTRL, HIFN_PUCTRL_DMAENA);
        hifn_wait_puc(dev);
}

static void hifn_stop_device(struct hifn_device *dev)
{
        hifn_write_1(dev, HIFN_1_DMA_CSR,
                HIFN_DMACSR_D_CTRL_DIS | HIFN_DMACSR_R_CTRL_DIS |
                HIFN_DMACSR_S_CTRL_DIS | HIFN_DMACSR_C_CTRL_DIS);
        hifn_write_0(dev, HIFN_0_PUIER, 0);
        hifn_write_1(dev, HIFN_1_DMA_IER, 0);
}

static void hifn_reset_dma(struct hifn_device *dev, int full)
{
        hifn_stop_device(dev);

        /*
         * Setting poll frequency and others to 0.
         */
        hifn_write_1(dev, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
                        HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
        mdelay(1);

        /*
         * Reset DMA.
         */
        if (full) {
                hifn_write_1(dev, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MODE);
                mdelay(1);
        } else {
                hifn_write_1(dev, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MODE |
                                HIFN_DMACNFG_MSTRESET);
                hifn_reset_puc(dev);
        }

        hifn_write_1(dev, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
                        HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);

        hifn_reset_puc(dev);
}

static u32 hifn_next_signature(u32 a, u_int cnt)
{
        int i;
        u32 v;

        for (i = 0; i < cnt; i++) {
                /* get the parity */
                v = a & 0x80080125;
                v ^= v >> 16;
                v ^= v >> 8;
                v ^= v >> 4;
                v ^= v >> 2;
                v ^= v >> 1;

                a = (v & 1) ^ (a << 1);
        }

        return a;
}

static struct pci2id {
        u_short         pci_vendor;
        u_short         pci_prod;
        char            card_id[13];
} pci2id[] = {
        {
                PCI_VENDOR_ID_HIFN,
                PCI_DEVICE_ID_HIFN_7955,
                { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
                  0x00, 0x00, 0x00, 0x00, 0x00 }
        },
        {
                PCI_VENDOR_ID_HIFN,
                PCI_DEVICE_ID_HIFN_7956,
                { 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
                  0x00, 0x00, 0x00, 0x00, 0x00 }
        }
};

#ifdef CONFIG_CRYPTO_DEV_HIFN_795X_RNG
static int hifn_rng_data_present(struct hwrng *rng, int wait)
{
        struct hifn_device *dev = (struct hifn_device *)rng->priv;
        s64 nsec;

        nsec = ktime_to_ns(ktime_sub(ktime_get(), dev->rngtime));
        nsec -= dev->rng_wait_time;
        if (nsec <= 0)
                return 1;
        if (!wait)
                return 0;
        ndelay(nsec);
        return 1;
}

static int hifn_rng_data_read(struct hwrng *rng, u32 *data)
{
        struct hifn_device *dev = (struct hifn_device *)rng->priv;

        *data = hifn_read_1(dev, HIFN_1_RNG_DATA);
        dev->rngtime = ktime_get();
        return 4;
}

static int hifn_register_rng(struct hifn_device *dev)
{
        /*
         * We must wait at least 256 Pk_clk cycles between two reads of the rng.
         */
        dev->rng_wait_time      = DIV_ROUND_UP_ULL(NSEC_PER_SEC,
                                                   dev->pk_clk_freq) * 256;

        dev->rng.name           = dev->name;
        dev->rng.data_present   = hifn_rng_data_present;
        dev->rng.data_read      = hifn_rng_data_read;
        dev->rng.priv           = (unsigned long)dev;

        return hwrng_register(&dev->rng);
}

static void hifn_unregister_rng(struct hifn_device *dev)
{
        hwrng_unregister(&dev->rng);
}
#else
#define hifn_register_rng(dev)          0
#define hifn_unregister_rng(dev)
#endif

static int hifn_init_pubrng(struct hifn_device *dev)
{
        int i;

        hifn_write_1(dev, HIFN_1_PUB_RESET, hifn_read_1(dev, HIFN_1_PUB_RESET) |
                        HIFN_PUBRST_RESET);

        for (i = 100; i > 0; --i) {
                mdelay(1);

                if ((hifn_read_1(dev, HIFN_1_PUB_RESET) & HIFN_PUBRST_RESET) == 0)
                        break;
        }

        if (!i) {
                dev_err(&dev->pdev->dev, "Failed to initialise public key engine.\n");
        } else {
                hifn_write_1(dev, HIFN_1_PUB_IEN, HIFN_PUBIEN_DONE);
                dev->dmareg |= HIFN_DMAIER_PUBDONE;
                hifn_write_1(dev, HIFN_1_DMA_IER, dev->dmareg);

                dev_dbg(&dev->pdev->dev, "Public key engine has been successfully initialised.\n");
        }

        /* Enable RNG engine. */

        hifn_write_1(dev, HIFN_1_RNG_CONFIG,
                        hifn_read_1(dev, HIFN_1_RNG_CONFIG) | HIFN_RNGCFG_ENA);
        dev_dbg(&dev->pdev->dev, "RNG engine has been successfully initialised.\n");

#ifdef CONFIG_CRYPTO_DEV_HIFN_795X_RNG
        /* First value must be discarded */
        hifn_read_1(dev, HIFN_1_RNG_DATA);
        dev->rngtime = ktime_get();
#endif
        return 0;
}

static int hifn_enable_crypto(struct hifn_device *dev)
{
        u32 dmacfg, addr;
        char *offtbl = NULL;
        int i;

        for (i = 0; i < ARRAY_SIZE(pci2id); i++) {
                if (pci2id[i].pci_vendor == dev->pdev->vendor &&
                                pci2id[i].pci_prod == dev->pdev->device) {
                        offtbl = pci2id[i].card_id;
                        break;
                }
        }

        if (!offtbl) {
                dev_err(&dev->pdev->dev, "Unknown card!\n");
                return -ENODEV;
        }

        dmacfg = hifn_read_1(dev, HIFN_1_DMA_CNFG);

        hifn_write_1(dev, HIFN_1_DMA_CNFG,
                        HIFN_DMACNFG_UNLOCK | HIFN_DMACNFG_MSTRESET |
                        HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE);
        mdelay(1);
        addr = hifn_read_1(dev, HIFN_1_UNLOCK_SECRET1);
        mdelay(1);
        hifn_write_1(dev, HIFN_1_UNLOCK_SECRET2, 0);
        mdelay(1);

        for (i = 0; i < 12; ++i) {
                addr = hifn_next_signature(addr, offtbl[i] + 0x101);
                hifn_write_1(dev, HIFN_1_UNLOCK_SECRET2, addr);

                mdelay(1);
        }
        hifn_write_1(dev, HIFN_1_DMA_CNFG, dmacfg);

        dev_dbg(&dev->pdev->dev, "%s %s.\n", dev->name, pci_name(dev->pdev));

        return 0;
}

static void hifn_init_dma(struct hifn_device *dev)
{
        struct hifn_dma *dma = dev->desc_virt;
        u32 dptr = dev->desc_dma;
        int i;

        for (i = 0; i < HIFN_D_CMD_RSIZE; ++i)
                dma->cmdr[i].p = __cpu_to_le32(dptr +
                                offsetof(struct hifn_dma, command_bufs[i][0]));
        for (i = 0; i < HIFN_D_RES_RSIZE; ++i)
                dma->resr[i].p = __cpu_to_le32(dptr +
                                offsetof(struct hifn_dma, result_bufs[i][0]));

        /* Setup LAST descriptors. */
        dma->cmdr[HIFN_D_CMD_RSIZE].p = __cpu_to_le32(dptr +
                        offsetof(struct hifn_dma, cmdr[0]));
        dma->srcr[HIFN_D_SRC_RSIZE].p = __cpu_to_le32(dptr +
                        offsetof(struct hifn_dma, srcr[0]));
        dma->dstr[HIFN_D_DST_RSIZE].p = __cpu_to_le32(dptr +
                        offsetof(struct hifn_dma, dstr[0]));
        dma->resr[HIFN_D_RES_RSIZE].p = __cpu_to_le32(dptr +
                        offsetof(struct hifn_dma, resr[0]));

        dma->cmdu = dma->srcu = dma->dstu = dma->resu = 0;
        dma->cmdi = dma->srci = dma->dsti = dma->resi = 0;
        dma->cmdk = dma->srck = dma->dstk = dma->resk = 0;
}

/*
 * Initialize the PLL. We need to know the frequency of the reference clock
 * to calculate the optimal multiplier. For PCI we assume 66MHz, since that
 * allows us to operate without the risk of overclocking the chip. If it
 * actually uses 33MHz, the chip will operate at half the speed, this can be
 * overridden by specifying the frequency as module parameter (pci33).
 *
 * Unfortunately the PCI clock is not very suitable since the HIFN needs a
 * stable clock and the PCI clock frequency may vary, so the default is the
 * external clock. There is no way to find out its frequency, we default to
 * 66MHz since according to Mike Ham of HiFn, almost every board in existence
 * has an external crystal populated at 66MHz.
 */
static void hifn_init_pll(struct hifn_device *dev)
{
        unsigned int freq, m;
        u32 pllcfg;

        pllcfg = HIFN_1_PLL | HIFN_PLL_RESERVED_1;

        if (strncmp(hifn_pll_ref, "ext", 3) == 0)
                pllcfg |= HIFN_PLL_REF_CLK_PLL;
        else
                pllcfg |= HIFN_PLL_REF_CLK_HBI;

        if (hifn_pll_ref[3] == '\0' ||
            kstrtouint(hifn_pll_ref + 3, 10, &freq)) {
                freq = 66;
                dev_info(&dev->pdev->dev, "assuming %u MHz clock speed, override with hifn_pll_ref=%.3s<frequency>\n",
                         freq, hifn_pll_ref);
        }

        m = HIFN_PLL_FCK_MAX / freq;

        pllcfg |= (m / 2 - 1) << HIFN_PLL_ND_SHIFT;
        if (m <= 8)
                pllcfg |= HIFN_PLL_IS_1_8;
        else
                pllcfg |= HIFN_PLL_IS_9_12;

        /* Select clock source and enable clock bypass */
        hifn_write_1(dev, HIFN_1_PLL, pllcfg |
                     HIFN_PLL_PK_CLK_HBI | HIFN_PLL_PE_CLK_HBI | HIFN_PLL_BP);

        /* Let the chip lock to the input clock */
        mdelay(10);

        /* Disable clock bypass */
        hifn_write_1(dev, HIFN_1_PLL, pllcfg |
                     HIFN_PLL_PK_CLK_HBI | HIFN_PLL_PE_CLK_HBI);

        /* Switch the engines to the PLL */
        hifn_write_1(dev, HIFN_1_PLL, pllcfg |
                     HIFN_PLL_PK_CLK_PLL | HIFN_PLL_PE_CLK_PLL);

        /*
         * The Fpk_clk runs at half the total speed. Its frequency is needed to
         * calculate the minimum time between two reads of the rng. Since 33MHz
         * is actually 33.333... we overestimate the frequency here, resulting
         * in slightly larger intervals.
         */
        dev->pk_clk_freq = 1000000 * (freq + 1) * m / 2;
}

static void hifn_init_registers(struct hifn_device *dev)
{
        u32 dptr = dev->desc_dma;

        /* Initialization magic... */
        hifn_write_0(dev, HIFN_0_PUCTRL, HIFN_PUCTRL_DMAENA);
        hifn_write_0(dev, HIFN_0_FIFOCNFG, HIFN_FIFOCNFG_THRESHOLD);
        hifn_write_0(dev, HIFN_0_PUIER, HIFN_PUIER_DSTOVER);

        /* write all 4 ring address registers */
        hifn_write_1(dev, HIFN_1_DMA_CRAR, dptr +
                                offsetof(struct hifn_dma, cmdr[0]));
        hifn_write_1(dev, HIFN_1_DMA_SRAR, dptr +
                                offsetof(struct hifn_dma, srcr[0]));
        hifn_write_1(dev, HIFN_1_DMA_DRAR, dptr +
                                offsetof(struct hifn_dma, dstr[0]));
        hifn_write_1(dev, HIFN_1_DMA_RRAR, dptr +
                                offsetof(struct hifn_dma, resr[0]));

        mdelay(2);
#if 0
        hifn_write_1(dev, HIFN_1_DMA_CSR,
            HIFN_DMACSR_D_CTRL_DIS | HIFN_DMACSR_R_CTRL_DIS |
            HIFN_DMACSR_S_CTRL_DIS | HIFN_DMACSR_C_CTRL_DIS |
            HIFN_DMACSR_D_ABORT | HIFN_DMACSR_D_DONE | HIFN_DMACSR_D_LAST |
            HIFN_DMACSR_D_WAIT | HIFN_DMACSR_D_OVER |
            HIFN_DMACSR_R_ABORT | HIFN_DMACSR_R_DONE | HIFN_DMACSR_R_LAST |
            HIFN_DMACSR_R_WAIT | HIFN_DMACSR_R_OVER |
            HIFN_DMACSR_S_ABORT | HIFN_DMACSR_S_DONE | HIFN_DMACSR_S_LAST |
            HIFN_DMACSR_S_WAIT |
            HIFN_DMACSR_C_ABORT | HIFN_DMACSR_C_DONE | HIFN_DMACSR_C_LAST |
            HIFN_DMACSR_C_WAIT |
            HIFN_DMACSR_ENGINE |
            HIFN_DMACSR_PUBDONE);
#else
        hifn_write_1(dev, HIFN_1_DMA_CSR,
            HIFN_DMACSR_C_CTRL_ENA | HIFN_DMACSR_S_CTRL_ENA |
            HIFN_DMACSR_D_CTRL_ENA | HIFN_DMACSR_R_CTRL_ENA |
            HIFN_DMACSR_D_ABORT | HIFN_DMACSR_D_DONE | HIFN_DMACSR_D_LAST |
            HIFN_DMACSR_D_WAIT | HIFN_DMACSR_D_OVER |
            HIFN_DMACSR_R_ABORT | HIFN_DMACSR_R_DONE | HIFN_DMACSR_R_LAST |
            HIFN_DMACSR_R_WAIT | HIFN_DMACSR_R_OVER |
            HIFN_DMACSR_S_ABORT | HIFN_DMACSR_S_DONE | HIFN_DMACSR_S_LAST |
            HIFN_DMACSR_S_WAIT |
            HIFN_DMACSR_C_ABORT | HIFN_DMACSR_C_DONE | HIFN_DMACSR_C_LAST |
            HIFN_DMACSR_C_WAIT |
            HIFN_DMACSR_ENGINE |
            HIFN_DMACSR_PUBDONE);
#endif
        hifn_read_1(dev, HIFN_1_DMA_CSR);

        dev->dmareg |= HIFN_DMAIER_R_DONE | HIFN_DMAIER_C_ABORT |
            HIFN_DMAIER_D_OVER | HIFN_DMAIER_R_OVER |
            HIFN_DMAIER_S_ABORT | HIFN_DMAIER_D_ABORT | HIFN_DMAIER_R_ABORT |
            HIFN_DMAIER_ENGINE;
        dev->dmareg &= ~HIFN_DMAIER_C_WAIT;

        hifn_write_1(dev, HIFN_1_DMA_IER, dev->dmareg);
        hifn_read_1(dev, HIFN_1_DMA_IER);
#if 0
        hifn_write_0(dev, HIFN_0_PUCNFG, HIFN_PUCNFG_ENCCNFG |
                    HIFN_PUCNFG_DRFR_128 | HIFN_PUCNFG_TCALLPHASES |
                    HIFN_PUCNFG_TCDRVTOTEM | HIFN_PUCNFG_BUS32 |
                    HIFN_PUCNFG_DRAM);
#else
        hifn_write_0(dev, HIFN_0_PUCNFG, 0x10342);
#endif
        hifn_init_pll(dev);

        hifn_write_0(dev, HIFN_0_PUISR, HIFN_PUISR_DSTOVER);
        hifn_write_1(dev, HIFN_1_DMA_CNFG, HIFN_DMACNFG_MSTRESET |
            HIFN_DMACNFG_DMARESET | HIFN_DMACNFG_MODE | HIFN_DMACNFG_LAST |
            ((HIFN_POLL_FREQUENCY << 16 ) & HIFN_DMACNFG_POLLFREQ) |
            ((HIFN_POLL_SCALAR << 8) & HIFN_DMACNFG_POLLINVAL));
}

static int hifn_setup_base_command(struct hifn_device *dev, u8 *buf,
                unsigned dlen, unsigned slen, u16 mask, u8 snum)
{
        struct hifn_base_command *base_cmd;
        u8 *buf_pos = buf;

        base_cmd = (struct hifn_base_command *)buf_pos;
        base_cmd->masks = __cpu_to_le16(mask);
        base_cmd->total_source_count =
                __cpu_to_le16(slen & HIFN_BASE_CMD_LENMASK_LO);
        base_cmd->total_dest_count =
                __cpu_to_le16(dlen & HIFN_BASE_CMD_LENMASK_LO);

        dlen >>= 16;
        slen >>= 16;
        base_cmd->session_num = __cpu_to_le16(snum |
            ((slen << HIFN_BASE_CMD_SRCLEN_S) & HIFN_BASE_CMD_SRCLEN_M) |
            ((dlen << HIFN_BASE_CMD_DSTLEN_S) & HIFN_BASE_CMD_DSTLEN_M));

        return sizeof(struct hifn_base_command);
}

static int hifn_setup_crypto_command(struct hifn_device *dev,
                u8 *buf, unsigned dlen, unsigned slen,
                u8 *key, int keylen, u8 *iv, int ivsize, u16 mode)
{
        struct hifn_dma *dma = dev->desc_virt;
        struct hifn_crypt_command *cry_cmd;
        u8 *buf_pos = buf;
        u16 cmd_len;

        cry_cmd = (struct hifn_crypt_command *)buf_pos;

        cry_cmd->source_count = __cpu_to_le16(dlen & 0xffff);
        dlen >>= 16;
        cry_cmd->masks = __cpu_to_le16(mode |
                        ((dlen << HIFN_CRYPT_CMD_SRCLEN_S) &
                         HIFN_CRYPT_CMD_SRCLEN_M));
        cry_cmd->header_skip = 0;
        cry_cmd->reserved = 0;

        buf_pos += sizeof(struct hifn_crypt_command);

        dma->cmdu++;
        if (dma->cmdu > 1) {
                dev->dmareg |= HIFN_DMAIER_C_WAIT;
                hifn_write_1(dev, HIFN_1_DMA_IER, dev->dmareg);
        }

        if (keylen) {
                memcpy(buf_pos, key, keylen);
                buf_pos += keylen;
        }
        if (ivsize) {
                memcpy(buf_pos, iv, ivsize);
                buf_pos += ivsize;
        }

        cmd_len = buf_pos - buf;

        return cmd_len;
}

static int hifn_setup_cmd_desc(struct hifn_device *dev,
                struct hifn_context *ctx, struct hifn_request_context *rctx,
                void *priv, unsigned int nbytes)
{
        struct hifn_dma *dma = dev->desc_virt;
        int cmd_len, sa_idx;
        u8 *buf, *buf_pos;
        u16 mask;

        sa_idx = dma->cmdi;
        buf_pos = buf = dma->command_bufs[dma->cmdi];

        mask = 0;
        switch (rctx->op) {
        case ACRYPTO_OP_DECRYPT:
                mask = HIFN_BASE_CMD_CRYPT | HIFN_BASE_CMD_DECODE;
                break;
        case ACRYPTO_OP_ENCRYPT:
                mask = HIFN_BASE_CMD_CRYPT;
                break;
        case ACRYPTO_OP_HMAC:
                mask = HIFN_BASE_CMD_MAC;
                break;
        default:
                goto err_out;
        }

        buf_pos += hifn_setup_base_command(dev, buf_pos, nbytes,
                        nbytes, mask, dev->snum);

        if (rctx->op == ACRYPTO_OP_ENCRYPT || rctx->op == ACRYPTO_OP_DECRYPT) {
                u16 md = 0;

                if (ctx->keysize)
                        md |= HIFN_CRYPT_CMD_NEW_KEY;
                if (rctx->iv && rctx->mode != ACRYPTO_MODE_ECB)
                        md |= HIFN_CRYPT_CMD_NEW_IV;

                switch (rctx->mode) {
                case ACRYPTO_MODE_ECB:
                        md |= HIFN_CRYPT_CMD_MODE_ECB;
                        break;
                case ACRYPTO_MODE_CBC:
                        md |= HIFN_CRYPT_CMD_MODE_CBC;
                        break;
                case ACRYPTO_MODE_CFB:
                        md |= HIFN_CRYPT_CMD_MODE_CFB;
                        break;
                case ACRYPTO_MODE_OFB:
                        md |= HIFN_CRYPT_CMD_MODE_OFB;
                        break;
                default:
                        goto err_out;
                }

                switch (rctx->type) {
                case ACRYPTO_TYPE_AES_128:
                        if (ctx->keysize != 16)
                                goto err_out;
                        md |= HIFN_CRYPT_CMD_KSZ_128 |
                                HIFN_CRYPT_CMD_ALG_AES;
                        break;
                case ACRYPTO_TYPE_AES_192:
                        if (ctx->keysize != 24)
                                goto err_out;
                        md |= HIFN_CRYPT_CMD_KSZ_192 |
                                HIFN_CRYPT_CMD_ALG_AES;
                        break;
                case ACRYPTO_TYPE_AES_256:
                        if (ctx->keysize != 32)
                                goto err_out;
                        md |= HIFN_CRYPT_CMD_KSZ_256 |
                                HIFN_CRYPT_CMD_ALG_AES;
                        break;
                case ACRYPTO_TYPE_3DES:
                        if (ctx->keysize != 24)
                                goto err_out;
                        md |= HIFN_CRYPT_CMD_ALG_3DES;
                        break;
                case ACRYPTO_TYPE_DES:
                        if (ctx->keysize != 8)
                                goto err_out;
                        md |= HIFN_CRYPT_CMD_ALG_DES;
                        break;
                default:
                        goto err_out;
                }

                buf_pos += hifn_setup_crypto_command(dev, buf_pos,
                                nbytes, nbytes, ctx->key, ctx->keysize,
                                rctx->iv, rctx->ivsize, md);
        }

        dev->sa[sa_idx] = priv;
        dev->started++;

        cmd_len = buf_pos - buf;
        dma->cmdr[dma->cmdi].l = __cpu_to_le32(cmd_len | HIFN_D_VALID |
                        HIFN_D_LAST | HIFN_D_MASKDONEIRQ);

        if (++dma->cmdi == HIFN_D_CMD_RSIZE) {
                dma->cmdr[dma->cmdi].l = __cpu_to_le32(
                        HIFN_D_VALID | HIFN_D_LAST |
                        HIFN_D_MASKDONEIRQ | HIFN_D_JUMP);
                dma->cmdi = 0;
        } else {
                dma->cmdr[dma->cmdi - 1].l |= __cpu_to_le32(HIFN_D_VALID);
        }

        if (!(dev->flags & HIFN_FLAG_CMD_BUSY)) {
                hifn_write_1(dev, HIFN_1_DMA_CSR, HIFN_DMACSR_C_CTRL_ENA);
                dev->flags |= HIFN_FLAG_CMD_BUSY;
        }
        return 0;

err_out:
        return -EINVAL;
}

static int hifn_setup_src_desc(struct hifn_device *dev, struct page *page,
                unsigned int offset, unsigned int size, int last)
{
        struct hifn_dma *dma = dev->desc_virt;
        int idx;
        dma_addr_t addr;

        addr = dma_map_page(&dev->pdev->dev, page, offset, size,
                            DMA_TO_DEVICE);

        idx = dma->srci;

        dma->srcr[idx].p = __cpu_to_le32(addr);
        dma->srcr[idx].l = __cpu_to_le32(size | HIFN_D_VALID |
                        HIFN_D_MASKDONEIRQ | (last ? HIFN_D_LAST : 0));

        if (++idx == HIFN_D_SRC_RSIZE) {
                dma->srcr[idx].l = __cpu_to_le32(HIFN_D_VALID |
                                HIFN_D_JUMP | HIFN_D_MASKDONEIRQ |
                                (last ? HIFN_D_LAST : 0));
                idx = 0;
        }

        dma->srci = idx;
        dma->srcu++;

        if (!(dev->flags & HIFN_FLAG_SRC_BUSY)) {
                hifn_write_1(dev, HIFN_1_DMA_CSR, HIFN_DMACSR_S_CTRL_ENA);
                dev->flags |= HIFN_FLAG_SRC_BUSY;
        }

        return size;
}

static void hifn_setup_res_desc(struct hifn_device *dev)
{
        struct hifn_dma *dma = dev->desc_virt;

        dma->resr[dma->resi].l = __cpu_to_le32(HIFN_USED_RESULT |
                        HIFN_D_VALID | HIFN_D_LAST);
        /*
         * dma->resr[dma->resi].l = __cpu_to_le32(HIFN_MAX_RESULT | HIFN_D_VALID |
         *                                      HIFN_D_LAST);
         */

        if (++dma->resi == HIFN_D_RES_RSIZE) {
                dma->resr[HIFN_D_RES_RSIZE].l = __cpu_to_le32(HIFN_D_VALID |
                                HIFN_D_JUMP | HIFN_D_MASKDONEIRQ | HIFN_D_LAST);
                dma->resi = 0;
        }

        dma->resu++;

        if (!(dev->flags & HIFN_FLAG_RES_BUSY)) {
                hifn_write_1(dev, HIFN_1_DMA_CSR, HIFN_DMACSR_R_CTRL_ENA);
                dev->flags |= HIFN_FLAG_RES_BUSY;
        }
}

static void hifn_setup_dst_desc(struct hifn_device *dev, struct page *page,
                unsigned offset, unsigned size, int last)
{
        struct hifn_dma *dma = dev->desc_virt;
        int idx;
        dma_addr_t addr;

        addr = dma_map_page(&dev->pdev->dev, page, offset, size,
                            DMA_FROM_DEVICE);

        idx = dma->dsti;
        dma->dstr[idx].p = __cpu_to_le32(addr);
        dma->dstr[idx].l = __cpu_to_le32(size | HIFN_D_VALID |
                        HIFN_D_MASKDONEIRQ | (last ? HIFN_D_LAST : 0));

        if (++idx == HIFN_D_DST_RSIZE) {
                dma->dstr[idx].l = __cpu_to_le32(HIFN_D_VALID |
                                HIFN_D_JUMP | HIFN_D_MASKDONEIRQ |
                                (last ? HIFN_D_LAST : 0));
                idx = 0;
        }
        dma->dsti = idx;
        dma->dstu++;

        if (!(dev->flags & HIFN_FLAG_DST_BUSY)) {
                hifn_write_1(dev, HIFN_1_DMA_CSR, HIFN_DMACSR_D_CTRL_ENA);
                dev->flags |= HIFN_FLAG_DST_BUSY;
        }
}

static int hifn_setup_dma(struct hifn_device *dev,
                struct hifn_context *ctx, struct hifn_request_context *rctx,
                struct scatterlist *src, struct scatterlist *dst,
                unsigned int nbytes, void *priv)
{
        struct scatterlist *t;
        struct page *spage, *dpage;
        unsigned int soff, doff;
        unsigned int n, len;

        n = nbytes;
        while (n) {
                spage = sg_page(src);
                soff = src->offset;
                len = min(src->length, n);

                hifn_setup_src_desc(dev, spage, soff, len, n - len == 0);

                src++;
                n -= len;
        }

        t = &rctx->walk.cache[0];
        n = nbytes;
        while (n) {
                if (t->length && rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) {
                        BUG_ON(!sg_page(t));
                        dpage = sg_page(t);
                        doff = 0;
                        len = t->length;
                } else {
                        BUG_ON(!sg_page(dst));
                        dpage = sg_page(dst);
                        doff = dst->offset;
                        len = dst->length;
                }
                len = min(len, n);

                hifn_setup_dst_desc(dev, dpage, doff, len, n - len == 0);

                dst++;
                t++;
                n -= len;
        }

        hifn_setup_cmd_desc(dev, ctx, rctx, priv, nbytes);
        hifn_setup_res_desc(dev);
        return 0;
}

static int hifn_cipher_walk_init(struct hifn_cipher_walk *w,
                int num, gfp_t gfp_flags)
{
        int i;

        num = min(ASYNC_SCATTERLIST_CACHE, num);
        sg_init_table(w->cache, num);

        w->num = 0;
        for (i = 0; i < num; ++i) {
                struct page *page = alloc_page(gfp_flags);
                struct scatterlist *s;

                if (!page)
                        break;

                s = &w->cache[i];

                sg_set_page(s, page, PAGE_SIZE, 0);
                w->num++;
        }

        return i;
}

static void hifn_cipher_walk_exit(struct hifn_cipher_walk *w)
{
        int i;

        for (i = 0; i < w->num; ++i) {
                struct scatterlist *s = &w->cache[i];

                __free_page(sg_page(s));

                s->length = 0;
        }

        w->num = 0;
}

static int skcipher_add(unsigned int *drestp, struct scatterlist *dst,
                unsigned int size, unsigned int *nbytesp)
{
        unsigned int copy, drest = *drestp, nbytes = *nbytesp;
        int idx = 0;

        if (drest < size || size > nbytes)
                return -EINVAL;

        while (size) {
                copy = min3(drest, size, dst->length);

                size -= copy;
                drest -= copy;
                nbytes -= copy;

                pr_debug("%s: copy: %u, size: %u, drest: %u, nbytes: %u.\n",
                         __func__, copy, size, drest, nbytes);

                dst++;
                idx++;
        }

        *nbytesp = nbytes;
        *drestp = drest;

        return idx;
}

static int hifn_cipher_walk(struct skcipher_request *req,
                struct hifn_cipher_walk *w)
{
        struct scatterlist *dst, *t;
        unsigned int nbytes = req->cryptlen, offset, copy, diff;
        int idx, tidx, err;

        tidx = idx = 0;
        offset = 0;
        while (nbytes) {
                if (idx >= w->num && (w->flags & ASYNC_FLAGS_MISALIGNED))
                        return -EINVAL;

                dst = &req->dst[idx];

                pr_debug("\n%s: dlen: %u, doff: %u, offset: %u, nbytes: %u.\n",
                         __func__, dst->length, dst->offset, offset, nbytes);

                if (!IS_ALIGNED(dst->offset, HIFN_D_DST_DALIGN) ||
                    !IS_ALIGNED(dst->length, HIFN_D_DST_DALIGN) ||
                    offset) {
                        unsigned slen = min(dst->length - offset, nbytes);
                        unsigned dlen = PAGE_SIZE;

                        t = &w->cache[idx];

                        err = skcipher_add(&dlen, dst, slen, &nbytes);
                        if (err < 0)
                                return err;

                        idx += err;

                        copy = slen & ~(HIFN_D_DST_DALIGN - 1);
                        diff = slen & (HIFN_D_DST_DALIGN - 1);

                        if (dlen < nbytes) {
                                /*
                                 * Destination page does not have enough space
                                 * to put there additional blocksized chunk,
                                 * so we mark that page as containing only
                                 * blocksize aligned chunks:
                                 *      t->length = (slen & ~(HIFN_D_DST_DALIGN - 1));
                                 * and increase number of bytes to be processed
                                 * in next chunk:
                                 *      nbytes += diff;
                                 */
                                nbytes += diff;

                                /*
                                 * Temporary of course...
                                 * Kick author if you will catch this one.
                                 */
                                pr_err("%s: dlen: %u, nbytes: %u, slen: %u, offset: %u.\n",
                                       __func__, dlen, nbytes, slen, offset);
                                pr_err("%s: please contact author to fix this "
                                       "issue, generally you should not catch "
                                       "this path under any condition but who "
                                       "knows how did you use crypto code.\n"
                                       "Thank you.\n",  __func__);
                                BUG();
                        } else {
                                copy += diff + nbytes;

                                dst = &req->dst[idx];

                                err = skcipher_add(&dlen, dst, nbytes, &nbytes);
                                if (err < 0)
                                        return err;

                                idx += err;
                        }

                        t->length = copy;
                        t->offset = offset;
                } else {
                        nbytes -= min(dst->length, nbytes);
                        idx++;
                }

                tidx++;
        }

        return tidx;
}

static int hifn_setup_session(struct skcipher_request *req)
{
        struct hifn_context *ctx = crypto_tfm_ctx(req->base.tfm);
        struct hifn_request_context *rctx = skcipher_request_ctx(req);
        struct hifn_device *dev = ctx->dev;
        unsigned long dlen, flags;
        unsigned int nbytes = req->cryptlen, idx = 0;
        int err = -EINVAL, sg_num;
        struct scatterlist *dst;

        if (rctx->iv && !rctx->ivsize && rctx->mode != ACRYPTO_MODE_ECB)
                goto err_out_exit;

        rctx->walk.flags = 0;

        while (nbytes) {
                dst = &req->dst[idx];
                dlen = min(dst->length, nbytes);

                if (!IS_ALIGNED(dst->offset, HIFN_D_DST_DALIGN) ||
                    !IS_ALIGNED(dlen, HIFN_D_DST_DALIGN))
                        rctx->walk.flags |= ASYNC_FLAGS_MISALIGNED;

                nbytes -= dlen;
                idx++;
        }

        if (rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) {
                err = hifn_cipher_walk_init(&rctx->walk, idx, GFP_ATOMIC);
                if (err < 0)
                        return err;
        }

        sg_num = hifn_cipher_walk(req, &rctx->walk);
        if (sg_num < 0) {
                err = sg_num;
                goto err_out_exit;
        }

        spin_lock_irqsave(&dev->lock, flags);
        if (dev->started + sg_num > HIFN_QUEUE_LENGTH) {
                err = -EAGAIN;
                goto err_out;
        }

        err = hifn_setup_dma(dev, ctx, rctx, req->src, req->dst, req->cryptlen, req);
        if (err)
                goto err_out;

        dev->snum++;

        dev->active = HIFN_DEFAULT_ACTIVE_NUM;
        spin_unlock_irqrestore(&dev->lock, flags);

        return 0;

err_out:
        spin_unlock_irqrestore(&dev->lock, flags);
err_out_exit:
        if (err) {
                dev_info(&dev->pdev->dev, "iv: %p [%d], key: %p [%d], mode: %u, op: %u, "
                         "type: %u, err: %d.\n",
                         rctx->iv, rctx->ivsize,
                         ctx->key, ctx->keysize,
                         rctx->mode, rctx->op, rctx->type, err);
        }

        return err;
}

static int hifn_start_device(struct hifn_device *dev)
{
        int err;

        dev->started = dev->active = 0;
        hifn_reset_dma(dev, 1);

        err = hifn_enable_crypto(dev);
        if (err)
                return err;

        hifn_reset_puc(dev);

        hifn_init_dma(dev);

        hifn_init_registers(dev);

        hifn_init_pubrng(dev);

        return 0;
}

static int skcipher_get(void *saddr, unsigned int *srestp, unsigned int offset,
                struct scatterlist *dst, unsigned int size, unsigned int *nbytesp)
{
        unsigned int srest = *srestp, nbytes = *nbytesp, copy;
        void *daddr;
        int idx = 0;

        if (srest < size || size > nbytes)
                return -EINVAL;

        while (size) {
                copy = min3(srest, dst->length, size);

                daddr = kmap_atomic(sg_page(dst));
                memcpy(daddr + dst->offset + offset, saddr, copy);
                kunmap_atomic(daddr);

                nbytes -= copy;
                size -= copy;
                srest -= copy;
                saddr += copy;
                offset = 0;

                pr_debug("%s: copy: %u, size: %u, srest: %u, nbytes: %u.\n",
                         __func__, copy, size, srest, nbytes);

                dst++;
                idx++;
        }

        *nbytesp = nbytes;
        *srestp = srest;

        return idx;
}

static inline void hifn_complete_sa(struct hifn_device *dev, int i)
{
        unsigned long flags;

        spin_lock_irqsave(&dev->lock, flags);
        dev->sa[i] = NULL;
        dev->started--;
        if (dev->started < 0)
                dev_info(&dev->pdev->dev, "%s: started: %d.\n", __func__,
                         dev->started);
        spin_unlock_irqrestore(&dev->lock, flags);
        BUG_ON(dev->started < 0);
}

static void hifn_process_ready(struct skcipher_request *req, int error)
{
        struct hifn_request_context *rctx = skcipher_request_ctx(req);

        if (rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) {
                unsigned int nbytes = req->cryptlen;
                int idx = 0, err;
                struct scatterlist *dst, *t;
                void *saddr;

                while (nbytes) {
                        t = &rctx->walk.cache[idx];
                        dst = &req->dst[idx];

                        pr_debug("\n%s: sg_page(t): %p, t->length: %u, "
                                "sg_page(dst): %p, dst->length: %u, "
                                "nbytes: %u.\n",
                                __func__, sg_page(t), t->length,
                                sg_page(dst), dst->length, nbytes);

                        if (!t->length) {
                                nbytes -= min(dst->length, nbytes);
                                idx++;
                                continue;
                        }

                        saddr = kmap_atomic(sg_page(t));

                        err = skcipher_get(saddr, &t->length, t->offset,
                                        dst, nbytes, &nbytes);
                        if (err < 0) {
                                kunmap_atomic(saddr);
                                break;
                        }

                        idx += err;
                        kunmap_atomic(saddr);
                }

                hifn_cipher_walk_exit(&rctx->walk);
        }

        skcipher_request_complete(req, error);
}

static void hifn_clear_rings(struct hifn_device *dev, int error)
{
        struct hifn_dma *dma = dev->desc_virt;
        int i, u;

        dev_dbg(&dev->pdev->dev, "ring cleanup 1: i: %d.%d.%d.%d, u: %d.%d.%d.%d, "
                        "k: %d.%d.%d.%d.\n",
                        dma->cmdi, dma->srci, dma->dsti, dma->resi,
                        dma->cmdu, dma->srcu, dma->dstu, dma->resu,
                        dma->cmdk, dma->srck, dma->dstk, dma->resk);

        i = dma->resk; u = dma->resu;
        while (u != 0) {
                if (dma->resr[i].l & __cpu_to_le32(HIFN_D_VALID))
                        break;

                if (dev->sa[i]) {
                        dev->success++;
                        dev->reset = 0;
                        hifn_process_ready(dev->sa[i], error);
                        hifn_complete_sa(dev, i);
                }

                if (++i == HIFN_D_RES_RSIZE)
                        i = 0;
                u--;
        }
        dma->resk = i; dma->resu = u;

        i = dma->srck; u = dma->srcu;
        while (u != 0) {
                if (dma->srcr[i].l & __cpu_to_le32(HIFN_D_VALID))
                        break;
                if (++i == HIFN_D_SRC_RSIZE)
                        i = 0;
                u--;
        }
        dma->srck = i; dma->srcu = u;

        i = dma->cmdk; u = dma->cmdu;
        while (u != 0) {
                if (dma->cmdr[i].l & __cpu_to_le32(HIFN_D_VALID))
                        break;
                if (++i == HIFN_D_CMD_RSIZE)
                        i = 0;
                u--;
        }
        dma->cmdk = i; dma->cmdu = u;

        i = dma->dstk; u = dma->dstu;
        while (u != 0) {
                if (dma->dstr[i].l & __cpu_to_le32(HIFN_D_VALID))
                        break;
                if (++i == HIFN_D_DST_RSIZE)
                        i = 0;
                u--;
        }
        dma->dstk = i; dma->dstu = u;

        dev_dbg(&dev->pdev->dev, "ring cleanup 2: i: %d.%d.%d.%d, u: %d.%d.%d.%d, "
                        "k: %d.%d.%d.%d.\n",
                        dma->cmdi, dma->srci, dma->dsti, dma->resi,
                        dma->cmdu, dma->srcu, dma->dstu, dma->resu,
                        dma->cmdk, dma->srck, dma->dstk, dma->resk);
}

static void hifn_work(struct work_struct *work)
{
        struct delayed_work *dw = to_delayed_work(work);
        struct hifn_device *dev = container_of(dw, struct hifn_device, work);
        unsigned long flags;
        int reset = 0;
        u32 r = 0;

        spin_lock_irqsave(&dev->lock, flags);
        if (dev->active == 0) {
                struct hifn_dma *dma = dev->desc_virt;

                if (dma->cmdu == 0 && (dev->flags & HIFN_FLAG_CMD_BUSY)) {
                        dev->flags &= ~HIFN_FLAG_CMD_BUSY;
                        r |= HIFN_DMACSR_C_CTRL_DIS;
                }
                if (dma->srcu == 0 && (dev->flags & HIFN_FLAG_SRC_BUSY)) {
                        dev->flags &= ~HIFN_FLAG_SRC_BUSY;
                        r |= HIFN_DMACSR_S_CTRL_DIS;
                }
                if (dma->dstu == 0 && (dev->flags & HIFN_FLAG_DST_BUSY)) {
                        dev->flags &= ~HIFN_FLAG_DST_BUSY;
                        r |= HIFN_DMACSR_D_CTRL_DIS;
                }
                if (dma->resu == 0 && (dev->flags & HIFN_FLAG_RES_BUSY)) {
                        dev->flags &= ~HIFN_FLAG_RES_BUSY;
                        r |= HIFN_DMACSR_R_CTRL_DIS;
                }
                if (r)
                        hifn_write_1(dev, HIFN_1_DMA_CSR, r);
        } else
                dev->active--;

        if ((dev->prev_success == dev->success) && dev->started)
                reset = 1;
        dev->prev_success = dev->success;
        spin_unlock_irqrestore(&dev->lock, flags);

        if (reset) {
                if (++dev->reset >= 5) {
                        int i;
                        struct hifn_dma *dma = dev->desc_virt;

                        dev_info(&dev->pdev->dev,
                                 "r: %08x, active: %d, started: %d, "
                                 "success: %lu: qlen: %u/%u, reset: %d.\n",
                                 r, dev->active, dev->started,
                                 dev->success, dev->queue.qlen, dev->queue.max_qlen,
                                 reset);

                        dev_info(&dev->pdev->dev, "%s: res: ", __func__);
                        for (i = 0; i < HIFN_D_RES_RSIZE; ++i) {
                                pr_info("%x.%p ", dma->resr[i].l, dev->sa[i]);
                                if (dev->sa[i]) {
                                        hifn_process_ready(dev->sa[i], -ENODEV);
                                        hifn_complete_sa(dev, i);
                                }
                        }
                        pr_info("\n");

                        hifn_reset_dma(dev, 1);
                        hifn_stop_device(dev);
                        hifn_start_device(dev);
                        dev->reset = 0;
                }

                tasklet_schedule(&dev->tasklet);
        }

        schedule_delayed_work(&dev->work, HZ);
}

static irqreturn_t hifn_interrupt(int irq, void *data)
{
        struct hifn_device *dev = data;
        struct hifn_dma *dma = dev->desc_virt;
        u32 dmacsr, restart;

        dmacsr = hifn_read_1(dev, HIFN_1_DMA_CSR);

        dev_dbg(&dev->pdev->dev, "1 dmacsr: %08x, dmareg: %08x, res: %08x [%d], "
                        "i: %d.%d.%d.%d, u: %d.%d.%d.%d.\n",
                dmacsr, dev->dmareg, dmacsr & dev->dmareg, dma->cmdi,
                dma->cmdi, dma->srci, dma->dsti, dma->resi,
                dma->cmdu, dma->srcu, dma->dstu, dma->resu);

        if ((dmacsr & dev->dmareg) == 0)
                return IRQ_NONE;

        hifn_write_1(dev, HIFN_1_DMA_CSR, dmacsr & dev->dmareg);

        if (dmacsr & HIFN_DMACSR_ENGINE)
                hifn_write_0(dev, HIFN_0_PUISR, hifn_read_0(dev, HIFN_0_PUISR));
        if (dmacsr & HIFN_DMACSR_PUBDONE)
                hifn_write_1(dev, HIFN_1_PUB_STATUS,
                        hifn_read_1(dev, HIFN_1_PUB_STATUS) | HIFN_PUBSTS_DONE);

        restart = dmacsr & (HIFN_DMACSR_R_OVER | HIFN_DMACSR_D_OVER);
        if (restart) {
                u32 puisr = hifn_read_0(dev, HIFN_0_PUISR);

                dev_warn(&dev->pdev->dev, "overflow: r: %d, d: %d, puisr: %08x, d: %u.\n",
                         !!(dmacsr & HIFN_DMACSR_R_OVER),
                         !!(dmacsr & HIFN_DMACSR_D_OVER),
                        puisr, !!(puisr & HIFN_PUISR_DSTOVER));
                if (!!(puisr & HIFN_PUISR_DSTOVER))
                        hifn_write_0(dev, HIFN_0_PUISR, HIFN_PUISR_DSTOVER);
                hifn_write_1(dev, HIFN_1_DMA_CSR, dmacsr & (HIFN_DMACSR_R_OVER |
                                        HIFN_DMACSR_D_OVER));
        }

        restart = dmacsr & (HIFN_DMACSR_C_ABORT | HIFN_DMACSR_S_ABORT |
                        HIFN_DMACSR_D_ABORT | HIFN_DMACSR_R_ABORT);
        if (restart) {
                dev_warn(&dev->pdev->dev, "abort: c: %d, s: %d, d: %d, r: %d.\n",
                         !!(dmacsr & HIFN_DMACSR_C_ABORT),
                         !!(dmacsr & HIFN_DMACSR_S_ABORT),
                         !!(dmacsr & HIFN_DMACSR_D_ABORT),
                         !!(dmacsr & HIFN_DMACSR_R_ABORT));
                hifn_reset_dma(dev, 1);
                hifn_init_dma(dev);
                hifn_init_registers(dev);
        }

        if ((dmacsr & HIFN_DMACSR_C_WAIT) && (dma->cmdu == 0)) {
                dev_dbg(&dev->pdev->dev, "wait on command.\n");
                dev->dmareg &= ~(HIFN_DMAIER_C_WAIT);
                hifn_write_1(dev, HIFN_1_DMA_IER, dev->dmareg);
        }

        tasklet_schedule(&dev->tasklet);

        return IRQ_HANDLED;
}

static void hifn_flush(struct hifn_device *dev)
{
        unsigned long flags;
        struct crypto_async_request *async_req;
        struct skcipher_request *req;
        struct hifn_dma *dma = dev->desc_virt;
        int i;

        for (i = 0; i < HIFN_D_RES_RSIZE; ++i) {
                struct hifn_desc *d = &dma->resr[i];

                if (dev->sa[i]) {
                        hifn_process_ready(dev->sa[i],
                                (d->l & __cpu_to_le32(HIFN_D_VALID)) ? -ENODEV : 0);
                        hifn_complete_sa(dev, i);
                }
        }

        spin_lock_irqsave(&dev->lock, flags);
        while ((async_req = crypto_dequeue_request(&dev->queue))) {
                req = skcipher_request_cast(async_req);
                spin_unlock_irqrestore(&dev->lock, flags);

                hifn_process_ready(req, -ENODEV);

                spin_lock_irqsave(&dev->lock, flags);
        }
        spin_unlock_irqrestore(&dev->lock, flags);
}

static int hifn_setkey(struct crypto_skcipher *cipher, const u8 *key,
                unsigned int len)
{
        struct hifn_context *ctx = crypto_skcipher_ctx(cipher);
        struct hifn_device *dev = ctx->dev;
        int err;

        err = verify_skcipher_des_key(cipher, key);
        if (err)
                return err;

        dev->flags &= ~HIFN_FLAG_OLD_KEY;

        memcpy(ctx->key, key, len);
        ctx->keysize = len;

        return 0;
}

static int hifn_des3_setkey(struct crypto_skcipher *cipher, const u8 *key,
                            unsigned int len)
{
        struct hifn_context *ctx = crypto_skcipher_ctx(cipher);
        struct hifn_device *dev = ctx->dev;
        int err;

        err = verify_skcipher_des3_key(cipher, key);
        if (err)
                return err;

        dev->flags &= ~HIFN_FLAG_OLD_KEY;

        memcpy(ctx->key, key, len);
        ctx->keysize = len;

        return 0;
}

static int hifn_handle_req(struct skcipher_request *req)
{
        struct hifn_context *ctx = crypto_tfm_ctx(req->base.tfm);
        struct hifn_device *dev = ctx->dev;
        int err = -EAGAIN;

        if (dev->started + DIV_ROUND_UP(req->cryptlen, PAGE_SIZE) <= HIFN_QUEUE_LENGTH)
                err = hifn_setup_session(req);

        if (err == -EAGAIN) {
                unsigned long flags;

                spin_lock_irqsave(&dev->lock, flags);
                err = crypto_enqueue_request(&dev->queue, &req->base);
                spin_unlock_irqrestore(&dev->lock, flags);
        }

        return err;
}

static int hifn_setup_crypto_req(struct skcipher_request *req, u8 op,
                u8 type, u8 mode)
{
        struct hifn_context *ctx = crypto_tfm_ctx(req->base.tfm);
        struct hifn_request_context *rctx = skcipher_request_ctx(req);
        unsigned ivsize;

        ivsize = crypto_skcipher_ivsize(crypto_skcipher_reqtfm(req));

        if (req->iv && mode != ACRYPTO_MODE_ECB) {
                if (type == ACRYPTO_TYPE_AES_128)
                        ivsize = HIFN_AES_IV_LENGTH;
                else if (type == ACRYPTO_TYPE_DES)
                        ivsize = HIFN_DES_KEY_LENGTH;
                else if (type == ACRYPTO_TYPE_3DES)
                        ivsize = HIFN_3DES_KEY_LENGTH;
        }

        if (ctx->keysize != 16 && type == ACRYPTO_TYPE_AES_128) {
                if (ctx->keysize == 24)
                        type = ACRYPTO_TYPE_AES_192;
                else if (ctx->keysize == 32)
                        type = ACRYPTO_TYPE_AES_256;
        }

        rctx->op = op;
        rctx->mode = mode;
        rctx->type = type;
        rctx->iv = req->iv;
        rctx->ivsize = ivsize;

        /*
         * HEAVY TODO: needs to kick Herbert XU to write documentation.
         * HEAVY TODO: needs to kick Herbert XU to write documentation.
         * HEAVY TODO: needs to kick Herbert XU to write documentation.
         */

        return hifn_handle_req(req);
}

static int hifn_process_queue(struct hifn_device *dev)
{
        struct crypto_async_request *async_req, *backlog;
        struct skcipher_request *req;
        unsigned long flags;
        int err = 0;

        while (dev->started < HIFN_QUEUE_LENGTH) {
                spin_lock_irqsave(&dev->lock, flags);
                backlog = crypto_get_backlog(&dev->queue);
                async_req = crypto_dequeue_request(&dev->queue);
                spin_unlock_irqrestore(&dev->lock, flags);

                if (!async_req)
                        break;

                if (backlog)
                        crypto_request_complete(backlog, -EINPROGRESS);

                req = skcipher_request_cast(async_req);

                err = hifn_handle_req(req);
                if (err)
                        break;
        }

        return err;
}

static int hifn_setup_crypto(struct skcipher_request *req, u8 op,
                u8 type, u8 mode)
{
        int err;
        struct hifn_context *ctx = crypto_tfm_ctx(req->base.tfm);
        struct hifn_device *dev = ctx->dev;

        err = hifn_setup_crypto_req(req, op, type, mode);
        if (err)
                return err;

        if (dev->started < HIFN_QUEUE_LENGTH && dev->queue.qlen)
                hifn_process_queue(dev);

        return -EINPROGRESS;
}

/*
 * AES ecryption functions.
 */
static inline int hifn_encrypt_aes_ecb(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_ENCRYPT,
                        ACRYPTO_TYPE_AES_128, ACRYPTO_MODE_ECB);
}
static inline int hifn_encrypt_aes_cbc(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_ENCRYPT,
                        ACRYPTO_TYPE_AES_128, ACRYPTO_MODE_CBC);
}

/*
 * AES decryption functions.
 */
static inline int hifn_decrypt_aes_ecb(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_DECRYPT,
                        ACRYPTO_TYPE_AES_128, ACRYPTO_MODE_ECB);
}
static inline int hifn_decrypt_aes_cbc(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_DECRYPT,
                        ACRYPTO_TYPE_AES_128, ACRYPTO_MODE_CBC);
}

/*
 * DES ecryption functions.
 */
static inline int hifn_encrypt_des_ecb(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_ENCRYPT,
                        ACRYPTO_TYPE_DES, ACRYPTO_MODE_ECB);
}
static inline int hifn_encrypt_des_cbc(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_ENCRYPT,
                        ACRYPTO_TYPE_DES, ACRYPTO_MODE_CBC);
}

/*
 * DES decryption functions.
 */
static inline int hifn_decrypt_des_ecb(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_DECRYPT,
                        ACRYPTO_TYPE_DES, ACRYPTO_MODE_ECB);
}
static inline int hifn_decrypt_des_cbc(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_DECRYPT,
                        ACRYPTO_TYPE_DES, ACRYPTO_MODE_CBC);
}

/*
 * 3DES ecryption functions.
 */
static inline int hifn_encrypt_3des_ecb(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_ENCRYPT,
                        ACRYPTO_TYPE_3DES, ACRYPTO_MODE_ECB);
}
static inline int hifn_encrypt_3des_cbc(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_ENCRYPT,
                        ACRYPTO_TYPE_3DES, ACRYPTO_MODE_CBC);
}

/* 3DES decryption functions. */
static inline int hifn_decrypt_3des_ecb(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_DECRYPT,
                        ACRYPTO_TYPE_3DES, ACRYPTO_MODE_ECB);
}
static inline int hifn_decrypt_3des_cbc(struct skcipher_request *req)
{
        return hifn_setup_crypto(req, ACRYPTO_OP_DECRYPT,
                        ACRYPTO_TYPE_3DES, ACRYPTO_MODE_CBC);
}

struct hifn_alg_template {
        char name[CRYPTO_MAX_ALG_NAME];
        char drv_name[CRYPTO_MAX_ALG_NAME];
        unsigned int bsize;
        struct skcipher_alg skcipher;
};

static const struct hifn_alg_template hifn_alg_templates[] = {
        /*
         * 3DES ECB and CBC modes.
         */
        {
                .name = "cbc(des3_ede)", .drv_name = "cbc-3des", .bsize = 8,
                .skcipher = {
                        .ivsize         =       HIFN_IV_LENGTH,
                        .min_keysize    =       HIFN_3DES_KEY_LENGTH,
                        .max_keysize    =       HIFN_3DES_KEY_LENGTH,
                        .setkey         =       hifn_des3_setkey,
                        .encrypt        =       hifn_encrypt_3des_cbc,
                        .decrypt        =       hifn_decrypt_3des_cbc,
                },
        },
        {
                .name = "ecb(des3_ede)", .drv_name = "ecb-3des", .bsize = 8,
                .skcipher = {
                        .min_keysize    =       HIFN_3DES_KEY_LENGTH,
                        .max_keysize    =       HIFN_3DES_KEY_LENGTH,
                        .setkey         =       hifn_des3_setkey,
                        .encrypt        =       hifn_encrypt_3des_ecb,
                        .decrypt        =       hifn_decrypt_3des_ecb,
                },
        },

        /*
         * DES ECB and CBC modes.
         */
        {
                .name = "cbc(des)", .drv_name = "cbc-des", .bsize = 8,
                .skcipher = {
                        .ivsize         =       HIFN_IV_LENGTH,
                        .min_keysize    =       HIFN_DES_KEY_LENGTH,
                        .max_keysize    =       HIFN_DES_KEY_LENGTH,
                        .setkey         =       hifn_setkey,
                        .encrypt        =       hifn_encrypt_des_cbc,
                        .decrypt        =       hifn_decrypt_des_cbc,
                },
        },
        {
                .name = "ecb(des)", .drv_name = "ecb-des", .bsize = 8,
                .skcipher = {
                        .min_keysize    =       HIFN_DES_KEY_LENGTH,
                        .max_keysize    =       HIFN_DES_KEY_LENGTH,
                        .setkey         =       hifn_setkey,
                        .encrypt        =       hifn_encrypt_des_ecb,
                        .decrypt        =       hifn_decrypt_des_ecb,
                },
        },

        /*
         * AES ECB and CBC modes.
         */
        {
                .name = "ecb(aes)", .drv_name = "ecb-aes", .bsize = 16,
                .skcipher = {
                        .min_keysize    =       AES_MIN_KEY_SIZE,
                        .max_keysize    =       AES_MAX_KEY_SIZE,
                        .setkey         =       hifn_setkey,
                        .encrypt        =       hifn_encrypt_aes_ecb,
                        .decrypt        =       hifn_decrypt_aes_ecb,
                },
        },
        {
                .name = "cbc(aes)", .drv_name = "cbc-aes", .bsize = 16,
                .skcipher = {
                        .ivsize         =       HIFN_AES_IV_LENGTH,
                        .min_keysize    =       AES_MIN_KEY_SIZE,
                        .max_keysize    =       AES_MAX_KEY_SIZE,
                        .setkey         =       hifn_setkey,
                        .encrypt        =       hifn_encrypt_aes_cbc,
                        .decrypt        =       hifn_decrypt_aes_cbc,
                },
        },
};

static int hifn_init_tfm(struct crypto_skcipher *tfm)
{
        struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
        struct hifn_crypto_alg *ha = crypto_alg_to_hifn(alg);
        struct hifn_context *ctx = crypto_skcipher_ctx(tfm);

        ctx->dev = ha->dev;
        crypto_skcipher_set_reqsize(tfm, sizeof(struct hifn_request_context));

        return 0;
}

static int hifn_alg_alloc(struct hifn_device *dev, const struct hifn_alg_template *t)
{
        struct hifn_crypto_alg *alg;
        int err;

        alg = kzalloc_obj(*alg);
        if (!alg)
                return -ENOMEM;

        alg->alg = t->skcipher;
        alg->alg.init = hifn_init_tfm;

        err = -EINVAL;
        if (snprintf(alg->alg.base.cra_name, CRYPTO_MAX_ALG_NAME,
                     "%s", t->name) >= CRYPTO_MAX_ALG_NAME)
                goto out_free_alg;
        if (snprintf(alg->alg.base.cra_driver_name, CRYPTO_MAX_ALG_NAME,
                     "%s-%s", t->drv_name, dev->name) >= CRYPTO_MAX_ALG_NAME)
                goto out_free_alg;

        alg->alg.base.cra_priority = 300;
        alg->alg.base.cra_flags = CRYPTO_ALG_KERN_DRIVER_ONLY | CRYPTO_ALG_ASYNC;
        alg->alg.base.cra_blocksize = t->bsize;
        alg->alg.base.cra_ctxsize = sizeof(struct hifn_context);
        alg->alg.base.cra_alignmask = 0;
        alg->alg.base.cra_module = THIS_MODULE;

        alg->dev = dev;

        list_add_tail(&alg->entry, &dev->alg_list);

        err = crypto_register_skcipher(&alg->alg);
        if (err) {
                list_del(&alg->entry);
out_free_alg:
                kfree(alg);
        }

        return err;
}

static void hifn_unregister_alg(struct hifn_device *dev)
{
        struct hifn_crypto_alg *a, *n;

        list_for_each_entry_safe(a, n, &dev->alg_list, entry) {
                list_del(&a->entry);
                crypto_unregister_skcipher(&a->alg);
                kfree(a);
        }
}

static int hifn_register_alg(struct hifn_device *dev)
{
        int i, err;

        for (i = 0; i < ARRAY_SIZE(hifn_alg_templates); ++i) {
                err = hifn_alg_alloc(dev, &hifn_alg_templates[i]);
                if (err)
                        goto err_out_exit;
        }

        return 0;

err_out_exit:
        hifn_unregister_alg(dev);
        return err;
}

static void hifn_tasklet_callback(unsigned long data)
{
        struct hifn_device *dev = (struct hifn_device *)data;

        /*
         * This is ok to call this without lock being held,
         * althogh it modifies some parameters used in parallel,
         * (like dev->success), but they are used in process
         * context or update is atomic (like setting dev->sa[i] to NULL).
         */
        hifn_clear_rings(dev, 0);

        if (dev->started < HIFN_QUEUE_LENGTH && dev->queue.qlen)
                hifn_process_queue(dev);
}

static int hifn_probe(struct pci_dev *pdev, const struct pci_device_id *id)
{
        int err, i;
        struct hifn_device *dev;
        char name[8];

        err = pci_enable_device(pdev);
        if (err)
                return err;
        pci_set_master(pdev);

        err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
        if (err)
                goto err_out_disable_pci_device;

        snprintf(name, sizeof(name), "hifn%d",
                        atomic_inc_return(&hifn_dev_number) - 1);

        err = pci_request_regions(pdev, name);
        if (err)
                goto err_out_disable_pci_device;

        if (pci_resource_len(pdev, 0) < HIFN_BAR0_SIZE ||
            pci_resource_len(pdev, 1) < HIFN_BAR1_SIZE ||
            pci_resource_len(pdev, 2) < HIFN_BAR2_SIZE) {
                dev_err(&pdev->dev, "Broken hardware - I/O regions are too small.\n");
                err = -ENODEV;
                goto err_out_free_regions;
        }

        dev = kzalloc(sizeof(struct hifn_device) + sizeof(struct crypto_alg),
                        GFP_KERNEL);
        if (!dev) {
                err = -ENOMEM;
                goto err_out_free_regions;
        }

        INIT_LIST_HEAD(&dev->alg_list);

        snprintf(dev->name, sizeof(dev->name), "%s", name);
        spin_lock_init(&dev->lock);

        for (i = 0; i < 3; ++i) {
                unsigned long addr, size;

                addr = pci_resource_start(pdev, i);
                size = pci_resource_len(pdev, i);

                dev->bar[i] = ioremap(addr, size);
                if (!dev->bar[i]) {
                        err = -ENOMEM;
                        goto err_out_unmap_bars;
                }
        }

        dev->desc_virt = dma_alloc_coherent(&pdev->dev,
                                            sizeof(struct hifn_dma),
                                            &dev->desc_dma, GFP_KERNEL);
        if (!dev->desc_virt) {
                dev_err(&pdev->dev, "Failed to allocate descriptor rings.\n");
                err = -ENOMEM;
                goto err_out_unmap_bars;
        }

        dev->pdev = pdev;
        dev->irq = pdev->irq;

        for (i = 0; i < HIFN_D_RES_RSIZE; ++i)
                dev->sa[i] = NULL;

        pci_set_drvdata(pdev, dev);

        tasklet_init(&dev->tasklet, hifn_tasklet_callback, (unsigned long)dev);

        crypto_init_queue(&dev->queue, 1);

        err = request_irq(dev->irq, hifn_interrupt, IRQF_SHARED, dev->name, dev);
        if (err) {
                dev_err(&pdev->dev, "Failed to request IRQ%d: err: %d.\n",
                        dev->irq, err);
                dev->irq = 0;
                goto err_out_free_desc;
        }

        err = hifn_start_device(dev);
        if (err)
                goto err_out_free_irq;

        err = hifn_register_rng(dev);
        if (err)
                goto err_out_stop_device;

        err = hifn_register_alg(dev);
        if (err)
                goto err_out_unregister_rng;

        INIT_DELAYED_WORK(&dev->work, hifn_work);
        schedule_delayed_work(&dev->work, HZ);

        dev_dbg(&pdev->dev, "HIFN crypto accelerator card at %s has been "
                "successfully registered as %s.\n",
                pci_name(pdev), dev->name);

        return 0;

err_out_unregister_rng:
        hifn_unregister_rng(dev);
err_out_stop_device:
        hifn_reset_dma(dev, 1);
        hifn_stop_device(dev);
err_out_free_irq:
        free_irq(dev->irq, dev);
        tasklet_kill(&dev->tasklet);
err_out_free_desc:
        dma_free_coherent(&pdev->dev, sizeof(struct hifn_dma), dev->desc_virt,
                          dev->desc_dma);

err_out_unmap_bars:
        for (i = 0; i < 3; ++i)
                if (dev->bar[i])
                        iounmap(dev->bar[i]);
        kfree(dev);

err_out_free_regions:
        pci_release_regions(pdev);

err_out_disable_pci_device:
        pci_disable_device(pdev);

        return err;
}

static void hifn_remove(struct pci_dev *pdev)
{
        int i;
        struct hifn_device *dev;

        dev = pci_get_drvdata(pdev);

        if (dev) {
                cancel_delayed_work_sync(&dev->work);

                hifn_unregister_rng(dev);
                hifn_unregister_alg(dev);
                hifn_reset_dma(dev, 1);
                hifn_stop_device(dev);

                free_irq(dev->irq, dev);
                tasklet_kill(&dev->tasklet);

                hifn_flush(dev);

                dma_free_coherent(&pdev->dev, sizeof(struct hifn_dma),
                                  dev->desc_virt, dev->desc_dma);
                for (i = 0; i < 3; ++i)
                        if (dev->bar[i])
                                iounmap(dev->bar[i]);

                kfree(dev);
        }

        pci_release_regions(pdev);
        pci_disable_device(pdev);
}

static struct pci_device_id hifn_pci_tbl[] = {
        { PCI_DEVICE(PCI_VENDOR_ID_HIFN, PCI_DEVICE_ID_HIFN_7955) },
        { PCI_DEVICE(PCI_VENDOR_ID_HIFN, PCI_DEVICE_ID_HIFN_7956) },
        { 0 }
};
MODULE_DEVICE_TABLE(pci, hifn_pci_tbl);

static struct pci_driver hifn_pci_driver = {
        .name     = "hifn795x",
        .id_table = hifn_pci_tbl,
        .probe    = hifn_probe,
        .remove   = hifn_remove,
};

static int __init hifn_init(void)
{
        unsigned int freq;
        int err;

        if (strncmp(hifn_pll_ref, "ext", 3) &&
            strncmp(hifn_pll_ref, "pci", 3)) {
                pr_err("hifn795x: invalid hifn_pll_ref clock, must be pci or ext");
                return -EINVAL;
        }

        /*
         * For the 7955/7956 the reference clock frequency must be in the
         * range of 20MHz-100MHz. For the 7954 the upper bound is 66.67MHz,
         * but this chip is currently not supported.
         */
        if (hifn_pll_ref[3] != '\0') {
                freq = simple_strtoul(hifn_pll_ref + 3, NULL, 10);
                if (freq < 20 || freq > 100) {
                        pr_err("hifn795x: invalid hifn_pll_ref frequency, must"
                               "be in the range of 20-100");
                        return -EINVAL;
                }
        }

        err = pci_register_driver(&hifn_pci_driver);
        if (err < 0) {
                pr_err("Failed to register PCI driver for %s device.\n",
                       hifn_pci_driver.name);
                return -ENODEV;
        }

        pr_info("Driver for HIFN 795x crypto accelerator chip "
                "has been successfully registered.\n");

        return 0;
}

static void __exit hifn_fini(void)
{
        pci_unregister_driver(&hifn_pci_driver);

        pr_info("Driver for HIFN 795x crypto accelerator chip "
                "has been successfully unregistered.\n");
}

module_init(hifn_init);
module_exit(hifn_fini);

MODULE_LICENSE("GPL");
MODULE_AUTHOR("Evgeniy Polyakov <johnpol@2ka.mipt.ru>");
MODULE_DESCRIPTION("Driver for HIFN 795x crypto accelerator chip.");