crypto: eip93 - Add Inside Secure SafeXcel EIP-93 crypto engine support
authorChristian Marangi <ansuelsmth@gmail.com>
Tue, 14 Jan 2025 12:36:36 +0000 (13:36 +0100)
committerHerbert Xu <herbert@gondor.apana.org.au>
Sun, 9 Feb 2025 10:08:11 +0000 (18:08 +0800)
Add support for the Inside Secure SafeXcel EIP-93 Crypto Engine used on
Mediatek MT7621 SoC and new Airoha SoC.

EIP-93 IP supports AES/DES/3DES ciphers in ECB/CBC and CTR modes as well as
authenc(HMAC(x), cipher(y)) using HMAC MD5, SHA1, SHA224 and SHA256.

EIP-93 provide regs to signal support for specific chipers and the
driver dynamically register only the supported one by the chip.

Signed-off-by: Richard van Schagen <vschagen@icloud.com>
Co-developed-by: Christian Marangi <ansuelsmth@gmail.com>
Signed-off-by: Christian Marangi <ansuelsmth@gmail.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
18 files changed:
MAINTAINERS
drivers/crypto/Kconfig
drivers/crypto/Makefile
drivers/crypto/inside-secure/eip93/Kconfig [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/Makefile [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-aead.c [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-aead.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-aes.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-cipher.c [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-cipher.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-common.c [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-common.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-des.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-hash.c [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-hash.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-main.c [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-main.h [new file with mode: 0644]
drivers/crypto/inside-secure/eip93/eip93-regs.h [new file with mode: 0644]

index 896a307..d286299 100644 (file)
@@ -11466,6 +11466,13 @@ L:     linux-crypto@vger.kernel.org
 S:     Maintained
 F:     drivers/crypto/inside-secure/
 
+INSIDE SECURE EIP93 CRYPTO DRIVER
+M:     Christian Marangi <ansuelsmth@gmail.com>
+L:     linux-crypto@vger.kernel.org
+S:     Maintained
+F:     Documentation/devicetree/bindings/crypto/inside-secure,safexcel-eip93.yaml
+F:     drivers/crypto/inside-secure/eip93/
+
 INTEGRITY MEASUREMENT ARCHITECTURE (IMA)
 M:     Mimi Zohar <zohar@linux.ibm.com>
 M:     Roberto Sassu <roberto.sassu@huawei.com>
index 19ab145..4708278 100644 (file)
@@ -855,5 +855,6 @@ config CRYPTO_DEV_SA2UL
 
 source "drivers/crypto/aspeed/Kconfig"
 source "drivers/crypto/starfive/Kconfig"
+source "drivers/crypto/inside-secure/eip93/Kconfig"
 
 endif # CRYPTO_HW
index fef18ff..62c7ce3 100644 (file)
@@ -50,3 +50,4 @@ obj-y += hisilicon/
 obj-$(CONFIG_CRYPTO_DEV_AMLOGIC_GXL) += amlogic/
 obj-y += intel/
 obj-y += starfive/
+obj-y += inside-secure/eip93/
diff --git a/drivers/crypto/inside-secure/eip93/Kconfig b/drivers/crypto/inside-secure/eip93/Kconfig
new file mode 100644 (file)
index 0000000..8353d3d
--- /dev/null
@@ -0,0 +1,20 @@
+# SPDX-License-Identifier: GPL-2.0
+config CRYPTO_DEV_EIP93
+       tristate "Support for EIP93 crypto HW accelerators"
+       depends on SOC_MT7621 || ARCH_AIROHA ||COMPILE_TEST
+       select CRYPTO_LIB_AES
+       select CRYPTO_LIB_DES
+       select CRYPTO_SKCIPHER
+       select CRYPTO_AEAD
+       select CRYPTO_AUTHENC
+       select CRYPTO_MD5
+       select CRYPTO_SHA1
+       select CRYPTO_SHA256
+       help
+         EIP93 have various crypto HW accelerators. Select this if
+         you want to use the EIP93 modules for any of the crypto algorithms.
+
+         If the IP supports it, this provide offload for AES - ECB, CBC and
+         CTR crypto. Also provide DES and 3DES ECB and CBC.
+
+         Also provide AEAD authenc(hmac(x), cipher(y)) for supported algo.
diff --git a/drivers/crypto/inside-secure/eip93/Makefile b/drivers/crypto/inside-secure/eip93/Makefile
new file mode 100644 (file)
index 0000000..a3d3d36
--- /dev/null
@@ -0,0 +1,5 @@
+obj-$(CONFIG_CRYPTO_DEV_EIP93) += crypto-hw-eip93.o
+
+crypto-hw-eip93-y += eip93-main.o eip93-common.o
+crypto-hw-eip93-y += eip93-cipher.o eip93-aead.o
+crypto-hw-eip93-y += eip93-hash.o
diff --git a/drivers/crypto/inside-secure/eip93/eip93-aead.c b/drivers/crypto/inside-secure/eip93/eip93-aead.c
new file mode 100644 (file)
index 0000000..18dd8a9
--- /dev/null
@@ -0,0 +1,711 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+
+#include <crypto/aead.h>
+#include <crypto/aes.h>
+#include <crypto/authenc.h>
+#include <crypto/ctr.h>
+#include <crypto/hmac.h>
+#include <crypto/internal/aead.h>
+#include <crypto/md5.h>
+#include <crypto/null.h>
+#include <crypto/sha1.h>
+#include <crypto/sha2.h>
+
+#include <crypto/internal/des.h>
+
+#include <linux/crypto.h>
+#include <linux/dma-mapping.h>
+
+#include "eip93-aead.h"
+#include "eip93-cipher.h"
+#include "eip93-common.h"
+#include "eip93-regs.h"
+
+void eip93_aead_handle_result(struct crypto_async_request *async, int err)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(async->tfm);
+       struct eip93_device *eip93 = ctx->eip93;
+       struct aead_request *req = aead_request_cast(async);
+       struct eip93_cipher_reqctx *rctx = aead_request_ctx(req);
+
+       eip93_unmap_dma(eip93, rctx, req->src, req->dst);
+       eip93_handle_result(eip93, rctx, req->iv);
+
+       aead_request_complete(req, err);
+}
+
+static int eip93_aead_send_req(struct crypto_async_request *async)
+{
+       struct aead_request *req = aead_request_cast(async);
+       struct eip93_cipher_reqctx *rctx = aead_request_ctx(req);
+       int err;
+
+       err = check_valid_request(rctx);
+       if (err) {
+               aead_request_complete(req, err);
+               return err;
+       }
+
+       return eip93_send_req(async, req->iv, rctx);
+}
+
+/* Crypto aead API functions */
+static int eip93_aead_cra_init(struct crypto_tfm *tfm)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct eip93_alg_template *tmpl = container_of(tfm->__crt_alg,
+                               struct eip93_alg_template, alg.aead.base);
+
+       crypto_aead_set_reqsize(__crypto_aead_cast(tfm),
+                               sizeof(struct eip93_cipher_reqctx));
+
+       ctx->eip93 = tmpl->eip93;
+       ctx->flags = tmpl->flags;
+       ctx->type = tmpl->type;
+       ctx->set_assoc = true;
+
+       ctx->sa_record = kzalloc(sizeof(*ctx->sa_record), GFP_KERNEL);
+       if (!ctx->sa_record)
+               return -ENOMEM;
+
+       return 0;
+}
+
+static void eip93_aead_cra_exit(struct crypto_tfm *tfm)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       dma_unmap_single(ctx->eip93->dev, ctx->sa_record_base,
+                        sizeof(*ctx->sa_record), DMA_TO_DEVICE);
+       kfree(ctx->sa_record);
+}
+
+static int eip93_aead_setkey(struct crypto_aead *ctfm, const u8 *key,
+                            unsigned int len)
+{
+       struct crypto_tfm *tfm = crypto_aead_tfm(ctfm);
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct crypto_authenc_keys keys;
+       struct crypto_aes_ctx aes;
+       struct sa_record *sa_record = ctx->sa_record;
+       u32 nonce = 0;
+       int ret;
+
+       if (crypto_authenc_extractkeys(&keys, key, len))
+               return -EINVAL;
+
+       if (IS_RFC3686(ctx->flags)) {
+               if (keys.enckeylen < CTR_RFC3686_NONCE_SIZE)
+                       return -EINVAL;
+
+               keys.enckeylen -= CTR_RFC3686_NONCE_SIZE;
+               memcpy(&nonce, keys.enckey + keys.enckeylen,
+                      CTR_RFC3686_NONCE_SIZE);
+       }
+
+       switch ((ctx->flags & EIP93_ALG_MASK)) {
+       case EIP93_ALG_DES:
+               ret = verify_aead_des_key(ctfm, keys.enckey, keys.enckeylen);
+               if (ret)
+                       return ret;
+
+               break;
+       case EIP93_ALG_3DES:
+               if (keys.enckeylen != DES3_EDE_KEY_SIZE)
+                       return -EINVAL;
+
+               ret = verify_aead_des3_key(ctfm, keys.enckey, keys.enckeylen);
+               if (ret)
+                       return ret;
+
+               break;
+       case EIP93_ALG_AES:
+               ret = aes_expandkey(&aes, keys.enckey, keys.enckeylen);
+               if (ret)
+                       return ret;
+
+               break;
+       }
+
+       ctx->blksize = crypto_aead_blocksize(ctfm);
+       /* Encryption key */
+       eip93_set_sa_record(sa_record, keys.enckeylen, ctx->flags);
+       sa_record->sa_cmd0_word &= ~EIP93_SA_CMD_OPCODE;
+       sa_record->sa_cmd0_word |= FIELD_PREP(EIP93_SA_CMD_OPCODE,
+                                             EIP93_SA_CMD_OPCODE_BASIC_OUT_ENC_HASH);
+       sa_record->sa_cmd0_word &= ~EIP93_SA_CMD_DIGEST_LENGTH;
+       sa_record->sa_cmd0_word |= FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH,
+                                             ctx->authsize / sizeof(u32));
+
+       memcpy(sa_record->sa_key, keys.enckey, keys.enckeylen);
+       ctx->sa_nonce = nonce;
+       sa_record->sa_nonce = nonce;
+
+       /* authentication key */
+       ret = eip93_hmac_setkey(ctx->flags, keys.authkey, keys.authkeylen,
+                               ctx->authsize, sa_record->sa_i_digest,
+                               sa_record->sa_o_digest, false);
+
+       ctx->set_assoc = true;
+
+       return ret;
+}
+
+static int eip93_aead_setauthsize(struct crypto_aead *ctfm,
+                                 unsigned int authsize)
+{
+       struct crypto_tfm *tfm = crypto_aead_tfm(ctfm);
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       ctx->authsize = authsize;
+       ctx->sa_record->sa_cmd0_word &= ~EIP93_SA_CMD_DIGEST_LENGTH;
+       ctx->sa_record->sa_cmd0_word |= FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH,
+                                                  ctx->authsize / sizeof(u32));
+
+       return 0;
+}
+
+static void eip93_aead_setassoc(struct eip93_crypto_ctx *ctx,
+                               struct aead_request *req)
+{
+       struct sa_record *sa_record = ctx->sa_record;
+
+       sa_record->sa_cmd1_word &= ~EIP93_SA_CMD_HASH_CRYPT_OFFSET;
+       sa_record->sa_cmd1_word |= FIELD_PREP(EIP93_SA_CMD_HASH_CRYPT_OFFSET,
+                                             req->assoclen / sizeof(u32));
+
+       ctx->assoclen = req->assoclen;
+}
+
+static int eip93_aead_crypt(struct aead_request *req)
+{
+       struct eip93_cipher_reqctx *rctx = aead_request_ctx(req);
+       struct crypto_async_request *async = &req->base;
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(req->base.tfm);
+       struct crypto_aead *aead = crypto_aead_reqtfm(req);
+       int ret;
+
+       ctx->sa_record_base = dma_map_single(ctx->eip93->dev, ctx->sa_record,
+                                            sizeof(*ctx->sa_record), DMA_TO_DEVICE);
+       ret = dma_mapping_error(ctx->eip93->dev, ctx->sa_record_base);
+       if (ret)
+               return ret;
+
+       rctx->textsize = req->cryptlen;
+       rctx->blksize = ctx->blksize;
+       rctx->assoclen = req->assoclen;
+       rctx->authsize = ctx->authsize;
+       rctx->sg_src = req->src;
+       rctx->sg_dst = req->dst;
+       rctx->ivsize = crypto_aead_ivsize(aead);
+       rctx->desc_flags = EIP93_DESC_AEAD;
+       rctx->sa_record_base = ctx->sa_record_base;
+
+       if (IS_DECRYPT(rctx->flags))
+               rctx->textsize -= rctx->authsize;
+
+       return eip93_aead_send_req(async);
+}
+
+static int eip93_aead_encrypt(struct aead_request *req)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(req->base.tfm);
+       struct eip93_cipher_reqctx *rctx = aead_request_ctx(req);
+
+       rctx->flags = ctx->flags;
+       rctx->flags |= EIP93_ENCRYPT;
+       if (ctx->set_assoc) {
+               eip93_aead_setassoc(ctx, req);
+               ctx->set_assoc = false;
+       }
+
+       if (req->assoclen != ctx->assoclen) {
+               dev_err(ctx->eip93->dev, "Request AAD length error\n");
+               return -EINVAL;
+       }
+
+       return eip93_aead_crypt(req);
+}
+
+static int eip93_aead_decrypt(struct aead_request *req)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(req->base.tfm);
+       struct eip93_cipher_reqctx *rctx = aead_request_ctx(req);
+
+       ctx->sa_record->sa_cmd0_word |= EIP93_SA_CMD_DIRECTION_IN;
+       ctx->sa_record->sa_cmd1_word &= ~(EIP93_SA_CMD_COPY_PAD |
+                                         EIP93_SA_CMD_COPY_DIGEST);
+
+       rctx->flags = ctx->flags;
+       rctx->flags |= EIP93_DECRYPT;
+       if (ctx->set_assoc) {
+               eip93_aead_setassoc(ctx, req);
+               ctx->set_assoc = false;
+       }
+
+       if (req->assoclen != ctx->assoclen) {
+               dev_err(ctx->eip93->dev, "Request AAD length error\n");
+               return -EINVAL;
+       }
+
+       return eip93_aead_crypt(req);
+}
+
+/* Available authenc algorithms in this module */
+struct eip93_alg_template eip93_alg_authenc_hmac_md5_cbc_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_MD5 | EIP93_MODE_CBC | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = AES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = MD5_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(md5),cbc(aes))",
+                       .cra_driver_name =
+                               "authenc(hmac(md5-eip93), cbc(aes-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = AES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha1_cbc_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA1 | EIP93_MODE_CBC | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = AES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA1_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha1),cbc(aes))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha1-eip93),cbc(aes-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = AES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha224_cbc_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA224 | EIP93_MODE_CBC | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = AES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA224_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha224),cbc(aes))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha224-eip93),cbc(aes-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = AES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha256_cbc_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA256 | EIP93_MODE_CBC | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = AES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA256_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha256),cbc(aes))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha256-eip93),cbc(aes-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = AES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_md5_rfc3686_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_MD5 |
+                       EIP93_MODE_CTR | EIP93_MODE_RFC3686 | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = CTR_RFC3686_IV_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = MD5_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(md5),rfc3686(ctr(aes)))",
+                       .cra_driver_name =
+                       "authenc(hmac(md5-eip93),rfc3686(ctr(aes-eip93)))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = 1,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha1_rfc3686_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA1 |
+                       EIP93_MODE_CTR | EIP93_MODE_RFC3686 | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = CTR_RFC3686_IV_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA1_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha1),rfc3686(ctr(aes)))",
+                       .cra_driver_name =
+                       "authenc(hmac(sha1-eip93),rfc3686(ctr(aes-eip93)))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = 1,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha224_rfc3686_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA224 |
+                       EIP93_MODE_CTR | EIP93_MODE_RFC3686 | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = CTR_RFC3686_IV_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA224_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha224),rfc3686(ctr(aes)))",
+                       .cra_driver_name =
+                       "authenc(hmac(sha224-eip93),rfc3686(ctr(aes-eip93)))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = 1,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha256_rfc3686_aes = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA256 |
+                       EIP93_MODE_CTR | EIP93_MODE_RFC3686 | EIP93_ALG_AES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = CTR_RFC3686_IV_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA256_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha256),rfc3686(ctr(aes)))",
+                       .cra_driver_name =
+                       "authenc(hmac(sha256-eip93),rfc3686(ctr(aes-eip93)))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = 1,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_md5_cbc_des = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_MD5 | EIP93_MODE_CBC | EIP93_ALG_DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = MD5_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(md5),cbc(des))",
+                       .cra_driver_name =
+                               "authenc(hmac(md5-eip93),cbc(des-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha1_cbc_des = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA1 | EIP93_MODE_CBC | EIP93_ALG_DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA1_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha1),cbc(des))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha1-eip93),cbc(des-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha224_cbc_des = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA224 | EIP93_MODE_CBC | EIP93_ALG_DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA224_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha224),cbc(des))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha224-eip93),cbc(des-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha256_cbc_des = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA256 | EIP93_MODE_CBC | EIP93_ALG_DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA256_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha256),cbc(des))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha256-eip93),cbc(des-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_md5_cbc_des3_ede = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_MD5 | EIP93_MODE_CBC | EIP93_ALG_3DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES3_EDE_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = MD5_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(md5),cbc(des3_ede))",
+                       .cra_driver_name =
+                               "authenc(hmac(md5-eip93),cbc(des3_ede-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES3_EDE_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0x0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha1_cbc_des3_ede = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA1 | EIP93_MODE_CBC | EIP93_ALG_3DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES3_EDE_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA1_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha1),cbc(des3_ede))",
+                       .cra_driver_name =
+                               "authenc(hmac(sha1-eip93),cbc(des3_ede-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES3_EDE_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0x0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha224_cbc_des3_ede = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA224 | EIP93_MODE_CBC | EIP93_ALG_3DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES3_EDE_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA224_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha224),cbc(des3_ede))",
+                       .cra_driver_name =
+                       "authenc(hmac(sha224-eip93),cbc(des3_ede-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES3_EDE_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0x0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_authenc_hmac_sha256_cbc_des3_ede = {
+       .type = EIP93_ALG_TYPE_AEAD,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA256 | EIP93_MODE_CBC | EIP93_ALG_3DES,
+       .alg.aead = {
+               .setkey = eip93_aead_setkey,
+               .encrypt = eip93_aead_encrypt,
+               .decrypt = eip93_aead_decrypt,
+               .ivsize = DES3_EDE_BLOCK_SIZE,
+               .setauthsize = eip93_aead_setauthsize,
+               .maxauthsize = SHA256_DIGEST_SIZE,
+               .base = {
+                       .cra_name = "authenc(hmac(sha256),cbc(des3_ede))",
+                       .cra_driver_name =
+                       "authenc(hmac(sha256-eip93),cbc(des3_ede-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                       CRYPTO_ALG_ALLOCATES_MEMORY,
+                       .cra_blocksize = DES3_EDE_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0x0,
+                       .cra_init = eip93_aead_cra_init,
+                       .cra_exit = eip93_aead_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
diff --git a/drivers/crypto/inside-secure/eip93/eip93-aead.h b/drivers/crypto/inside-secure/eip93/eip93-aead.h
new file mode 100644 (file)
index 0000000..e2fa8fd
--- /dev/null
@@ -0,0 +1,38 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef _EIP93_AEAD_H_
+#define _EIP93_AEAD_H_
+
+extern struct eip93_alg_template eip93_alg_authenc_hmac_md5_cbc_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha1_cbc_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha224_cbc_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha256_cbc_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_md5_ctr_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha1_ctr_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha224_ctr_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha256_ctr_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_md5_rfc3686_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha1_rfc3686_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha224_rfc3686_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha256_rfc3686_aes;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_md5_cbc_des;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha1_cbc_des;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha224_cbc_des;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha256_cbc_des;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_md5_cbc_des3_ede;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha1_cbc_des3_ede;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha224_cbc_des3_ede;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha256_cbc_des3_ede;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_md5_ecb_null;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha1_ecb_null;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha224_ecb_null;
+extern struct eip93_alg_template eip93_alg_authenc_hmac_sha256_ecb_null;
+
+void eip93_aead_handle_result(struct crypto_async_request *async, int err);
+
+#endif /* _EIP93_AEAD_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-aes.h b/drivers/crypto/inside-secure/eip93/eip93-aes.h
new file mode 100644 (file)
index 0000000..1d83d39
--- /dev/null
@@ -0,0 +1,16 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef _EIP93_AES_H_
+#define _EIP93_AES_H_
+
+extern struct eip93_alg_template eip93_alg_ecb_aes;
+extern struct eip93_alg_template eip93_alg_cbc_aes;
+extern struct eip93_alg_template eip93_alg_ctr_aes;
+extern struct eip93_alg_template eip93_alg_rfc3686_aes;
+
+#endif /* _EIP93_AES_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-cipher.c b/drivers/crypto/inside-secure/eip93/eip93-cipher.c
new file mode 100644 (file)
index 0000000..1f2d684
--- /dev/null
@@ -0,0 +1,413 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+
+#include <crypto/aes.h>
+#include <crypto/ctr.h>
+#include <crypto/internal/des.h>
+#include <linux/dma-mapping.h>
+
+#include "eip93-aes.h"
+#include "eip93-cipher.h"
+#include "eip93-common.h"
+#include "eip93-des.h"
+#include "eip93-regs.h"
+
+void eip93_skcipher_handle_result(struct crypto_async_request *async, int err)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(async->tfm);
+       struct eip93_device *eip93 = ctx->eip93;
+       struct skcipher_request *req = skcipher_request_cast(async);
+       struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req);
+
+       eip93_unmap_dma(eip93, rctx, req->src, req->dst);
+       eip93_handle_result(eip93, rctx, req->iv);
+
+       skcipher_request_complete(req, err);
+}
+
+static int eip93_skcipher_send_req(struct crypto_async_request *async)
+{
+       struct skcipher_request *req = skcipher_request_cast(async);
+       struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req);
+       int err;
+
+       err = check_valid_request(rctx);
+
+       if (err) {
+               skcipher_request_complete(req, err);
+               return err;
+       }
+
+       return eip93_send_req(async, req->iv, rctx);
+}
+
+/* Crypto skcipher API functions */
+static int eip93_skcipher_cra_init(struct crypto_tfm *tfm)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct eip93_alg_template *tmpl = container_of(tfm->__crt_alg,
+                               struct eip93_alg_template, alg.skcipher.base);
+
+       crypto_skcipher_set_reqsize(__crypto_skcipher_cast(tfm),
+                                   sizeof(struct eip93_cipher_reqctx));
+
+       memset(ctx, 0, sizeof(*ctx));
+
+       ctx->eip93 = tmpl->eip93;
+       ctx->type = tmpl->type;
+
+       ctx->sa_record = kzalloc(sizeof(*ctx->sa_record), GFP_KERNEL);
+       if (!ctx->sa_record)
+               return -ENOMEM;
+
+       return 0;
+}
+
+static void eip93_skcipher_cra_exit(struct crypto_tfm *tfm)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       dma_unmap_single(ctx->eip93->dev, ctx->sa_record_base,
+                        sizeof(*ctx->sa_record), DMA_TO_DEVICE);
+       kfree(ctx->sa_record);
+}
+
+static int eip93_skcipher_setkey(struct crypto_skcipher *ctfm, const u8 *key,
+                                unsigned int len)
+{
+       struct crypto_tfm *tfm = crypto_skcipher_tfm(ctfm);
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct eip93_alg_template *tmpl = container_of(tfm->__crt_alg,
+                                                    struct eip93_alg_template,
+                                                    alg.skcipher.base);
+       struct sa_record *sa_record = ctx->sa_record;
+       unsigned int keylen = len;
+       u32 flags = tmpl->flags;
+       u32 nonce = 0;
+       int ret;
+
+       if (!key || !keylen)
+               return -EINVAL;
+
+       if (IS_RFC3686(flags)) {
+               if (len < CTR_RFC3686_NONCE_SIZE)
+                       return -EINVAL;
+
+               keylen = len - CTR_RFC3686_NONCE_SIZE;
+               memcpy(&nonce, key + keylen, CTR_RFC3686_NONCE_SIZE);
+       }
+
+       if (flags & EIP93_ALG_DES) {
+               ctx->blksize = DES_BLOCK_SIZE;
+               ret = verify_skcipher_des_key(ctfm, key);
+               if (ret)
+                       return ret;
+       }
+       if (flags & EIP93_ALG_3DES) {
+               ctx->blksize = DES3_EDE_BLOCK_SIZE;
+               ret = verify_skcipher_des3_key(ctfm, key);
+               if (ret)
+                       return ret;
+       }
+
+       if (flags & EIP93_ALG_AES) {
+               struct crypto_aes_ctx aes;
+
+               ctx->blksize = AES_BLOCK_SIZE;
+               ret = aes_expandkey(&aes, key, keylen);
+               if (ret)
+                       return ret;
+       }
+
+       eip93_set_sa_record(sa_record, keylen, flags);
+
+       memcpy(sa_record->sa_key, key, keylen);
+       ctx->sa_nonce = nonce;
+       sa_record->sa_nonce = nonce;
+
+       return 0;
+}
+
+static int eip93_skcipher_crypt(struct skcipher_request *req)
+{
+       struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req);
+       struct crypto_async_request *async = &req->base;
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(req->base.tfm);
+       struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
+       int ret;
+
+       if (!req->cryptlen)
+               return 0;
+
+       /*
+        * ECB and CBC algorithms require message lengths to be
+        * multiples of block size.
+        */
+       if (IS_ECB(rctx->flags) || IS_CBC(rctx->flags))
+               if (!IS_ALIGNED(req->cryptlen,
+                               crypto_skcipher_blocksize(skcipher)))
+                       return -EINVAL;
+
+       ctx->sa_record_base = dma_map_single(ctx->eip93->dev, ctx->sa_record,
+                                            sizeof(*ctx->sa_record), DMA_TO_DEVICE);
+       ret = dma_mapping_error(ctx->eip93->dev, ctx->sa_record_base);
+       if (ret)
+               return ret;
+
+       rctx->assoclen = 0;
+       rctx->textsize = req->cryptlen;
+       rctx->authsize = 0;
+       rctx->sg_src = req->src;
+       rctx->sg_dst = req->dst;
+       rctx->ivsize = crypto_skcipher_ivsize(skcipher);
+       rctx->blksize = ctx->blksize;
+       rctx->desc_flags = EIP93_DESC_SKCIPHER;
+       rctx->sa_record_base = ctx->sa_record_base;
+
+       return eip93_skcipher_send_req(async);
+}
+
+static int eip93_skcipher_encrypt(struct skcipher_request *req)
+{
+       struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req);
+       struct eip93_alg_template *tmpl = container_of(req->base.tfm->__crt_alg,
+                               struct eip93_alg_template, alg.skcipher.base);
+
+       rctx->flags = tmpl->flags;
+       rctx->flags |= EIP93_ENCRYPT;
+
+       return eip93_skcipher_crypt(req);
+}
+
+static int eip93_skcipher_decrypt(struct skcipher_request *req)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(req->base.tfm);
+       struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req);
+       struct eip93_alg_template *tmpl = container_of(req->base.tfm->__crt_alg,
+                               struct eip93_alg_template, alg.skcipher.base);
+
+       ctx->sa_record->sa_cmd0_word |= EIP93_SA_CMD_DIRECTION_IN;
+
+       rctx->flags = tmpl->flags;
+       rctx->flags |= EIP93_DECRYPT;
+
+       return eip93_skcipher_crypt(req);
+}
+
+/* Available algorithms in this module */
+struct eip93_alg_template eip93_alg_ecb_aes = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_ECB | EIP93_ALG_AES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = AES_MIN_KEY_SIZE,
+               .max_keysize = AES_MAX_KEY_SIZE,
+               .ivsize = 0,
+               .base = {
+                       .cra_name = "ecb(aes)",
+                       .cra_driver_name = "ecb(aes-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_NEED_FALLBACK |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = AES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0xf,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_cbc_aes = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_CBC | EIP93_ALG_AES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = AES_MIN_KEY_SIZE,
+               .max_keysize = AES_MAX_KEY_SIZE,
+               .ivsize = AES_BLOCK_SIZE,
+               .base = {
+                       .cra_name = "cbc(aes)",
+                       .cra_driver_name = "cbc(aes-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_NEED_FALLBACK |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = AES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0xf,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_ctr_aes = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_CTR | EIP93_ALG_AES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = AES_MIN_KEY_SIZE,
+               .max_keysize = AES_MAX_KEY_SIZE,
+               .ivsize = AES_BLOCK_SIZE,
+               .base = {
+                       .cra_name = "ctr(aes)",
+                       .cra_driver_name = "ctr(aes-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                    CRYPTO_ALG_NEED_FALLBACK |
+                                    CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = 1,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0xf,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_rfc3686_aes = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_CTR | EIP93_MODE_RFC3686 | EIP93_ALG_AES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = AES_MIN_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
+               .max_keysize = AES_MAX_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
+               .ivsize = CTR_RFC3686_IV_SIZE,
+               .base = {
+                       .cra_name = "rfc3686(ctr(aes))",
+                       .cra_driver_name = "rfc3686(ctr(aes-eip93))",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_NEED_FALLBACK |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = 1,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0xf,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_ecb_des = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_ECB | EIP93_ALG_DES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = DES_KEY_SIZE,
+               .max_keysize = DES_KEY_SIZE,
+               .ivsize = 0,
+               .base = {
+                       .cra_name = "ecb(des)",
+                       .cra_driver_name = "ebc(des-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = DES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_cbc_des = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_CBC | EIP93_ALG_DES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = DES_KEY_SIZE,
+               .max_keysize = DES_KEY_SIZE,
+               .ivsize = DES_BLOCK_SIZE,
+               .base = {
+                       .cra_name = "cbc(des)",
+                       .cra_driver_name = "cbc(des-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = DES_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_ecb_des3_ede = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_ECB | EIP93_ALG_3DES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = DES3_EDE_KEY_SIZE,
+               .max_keysize = DES3_EDE_KEY_SIZE,
+               .ivsize = 0,
+               .base = {
+                       .cra_name = "ecb(des3_ede)",
+                       .cra_driver_name = "ecb(des3_ede-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = DES3_EDE_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_cbc_des3_ede = {
+       .type = EIP93_ALG_TYPE_SKCIPHER,
+       .flags = EIP93_MODE_CBC | EIP93_ALG_3DES,
+       .alg.skcipher = {
+               .setkey = eip93_skcipher_setkey,
+               .encrypt = eip93_skcipher_encrypt,
+               .decrypt = eip93_skcipher_decrypt,
+               .min_keysize = DES3_EDE_KEY_SIZE,
+               .max_keysize = DES3_EDE_KEY_SIZE,
+               .ivsize = DES3_EDE_BLOCK_SIZE,
+               .base = {
+                       .cra_name = "cbc(des3_ede)",
+                       .cra_driver_name = "cbc(des3_ede-eip93)",
+                       .cra_priority = EIP93_CRA_PRIORITY,
+                       .cra_flags = CRYPTO_ALG_ASYNC |
+                                       CRYPTO_ALG_KERN_DRIVER_ONLY,
+                       .cra_blocksize = DES3_EDE_BLOCK_SIZE,
+                       .cra_ctxsize = sizeof(struct eip93_crypto_ctx),
+                       .cra_alignmask = 0,
+                       .cra_init = eip93_skcipher_cra_init,
+                       .cra_exit = eip93_skcipher_cra_exit,
+                       .cra_module = THIS_MODULE,
+               },
+       },
+};
diff --git a/drivers/crypto/inside-secure/eip93/eip93-cipher.h b/drivers/crypto/inside-secure/eip93/eip93-cipher.h
new file mode 100644 (file)
index 0000000..6e2545e
--- /dev/null
@@ -0,0 +1,60 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef _EIP93_CIPHER_H_
+#define _EIP93_CIPHER_H_
+
+#include "eip93-main.h"
+
+struct eip93_crypto_ctx {
+       struct eip93_device             *eip93;
+       u32                             flags;
+       struct sa_record                *sa_record;
+       u32                             sa_nonce;
+       int                             blksize;
+       dma_addr_t                      sa_record_base;
+       /* AEAD specific */
+       unsigned int                    authsize;
+       unsigned int                    assoclen;
+       bool                            set_assoc;
+       enum eip93_alg_type             type;
+};
+
+struct eip93_cipher_reqctx {
+       u16                             desc_flags;
+       u16                             flags;
+       unsigned int                    blksize;
+       unsigned int                    ivsize;
+       unsigned int                    textsize;
+       unsigned int                    assoclen;
+       unsigned int                    authsize;
+       dma_addr_t                      sa_record_base;
+       struct sa_state                 *sa_state;
+       dma_addr_t                      sa_state_base;
+       struct eip93_descriptor         *cdesc;
+       struct scatterlist              *sg_src;
+       struct scatterlist              *sg_dst;
+       int                             src_nents;
+       int                             dst_nents;
+       struct sa_state                 *sa_state_ctr;
+       dma_addr_t                      sa_state_ctr_base;
+};
+
+int check_valid_request(struct eip93_cipher_reqctx *rctx);
+
+void eip93_unmap_dma(struct eip93_device *eip93, struct eip93_cipher_reqctx *rctx,
+                    struct scatterlist *reqsrc, struct scatterlist *reqdst);
+
+void eip93_skcipher_handle_result(struct crypto_async_request *async, int err);
+
+int eip93_send_req(struct crypto_async_request *async,
+                  const u8 *reqiv, struct eip93_cipher_reqctx *rctx);
+
+void eip93_handle_result(struct eip93_device *eip93, struct eip93_cipher_reqctx *rctx,
+                        u8 *reqiv);
+
+#endif /* _EIP93_CIPHER_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-common.c b/drivers/crypto/inside-secure/eip93/eip93-common.c
new file mode 100644 (file)
index 0000000..446a047
--- /dev/null
@@ -0,0 +1,809 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+
+#include <crypto/aes.h>
+#include <crypto/ctr.h>
+#include <crypto/hmac.h>
+#include <crypto/sha1.h>
+#include <crypto/sha2.h>
+#include <linux/kernel.h>
+#include <linux/delay.h>
+#include <linux/dma-mapping.h>
+#include <linux/scatterlist.h>
+
+#include "eip93-cipher.h"
+#include "eip93-hash.h"
+#include "eip93-common.h"
+#include "eip93-main.h"
+#include "eip93-regs.h"
+
+int eip93_parse_ctrl_stat_err(struct eip93_device *eip93, int err)
+{
+       u32 ext_err;
+
+       if (!err)
+               return 0;
+
+       switch (err & ~EIP93_PE_CTRL_PE_EXT_ERR_CODE) {
+       case EIP93_PE_CTRL_PE_AUTH_ERR:
+       case EIP93_PE_CTRL_PE_PAD_ERR:
+               return -EBADMSG;
+       /* let software handle anti-replay errors */
+       case EIP93_PE_CTRL_PE_SEQNUM_ERR:
+               return 0;
+       case EIP93_PE_CTRL_PE_EXT_ERR:
+               break;
+       default:
+               dev_err(eip93->dev, "Unhandled error 0x%08x\n", err);
+               return -EINVAL;
+       }
+
+       /* Parse additional ext errors */
+       ext_err = FIELD_GET(EIP93_PE_CTRL_PE_EXT_ERR_CODE, err);
+       switch (ext_err) {
+       case EIP93_PE_CTRL_PE_EXT_ERR_BUS:
+       case EIP93_PE_CTRL_PE_EXT_ERR_PROCESSING:
+               return -EIO;
+       case EIP93_PE_CTRL_PE_EXT_ERR_DESC_OWNER:
+               return -EACCES;
+       case EIP93_PE_CTRL_PE_EXT_ERR_INVALID_CRYPTO_OP:
+       case EIP93_PE_CTRL_PE_EXT_ERR_INVALID_CRYPTO_ALGO:
+       case EIP93_PE_CTRL_PE_EXT_ERR_SPI:
+               return -EINVAL;
+       case EIP93_PE_CTRL_PE_EXT_ERR_ZERO_LENGTH:
+       case EIP93_PE_CTRL_PE_EXT_ERR_INVALID_PK_LENGTH:
+       case EIP93_PE_CTRL_PE_EXT_ERR_BLOCK_SIZE_ERR:
+               return -EBADMSG;
+       default:
+               dev_err(eip93->dev, "Unhandled ext error 0x%08x\n", ext_err);
+               return -EINVAL;
+       }
+}
+
+static void *eip93_ring_next_wptr(struct eip93_device *eip93,
+                                 struct eip93_desc_ring *ring)
+{
+       void *ptr = ring->write;
+
+       if ((ring->write == ring->read - ring->offset) ||
+           (ring->read == ring->base && ring->write == ring->base_end))
+               return ERR_PTR(-ENOMEM);
+
+       if (ring->write == ring->base_end)
+               ring->write = ring->base;
+       else
+               ring->write += ring->offset;
+
+       return ptr;
+}
+
+static void *eip93_ring_next_rptr(struct eip93_device *eip93,
+                                 struct eip93_desc_ring *ring)
+{
+       void *ptr = ring->read;
+
+       if (ring->write == ring->read)
+               return ERR_PTR(-ENOENT);
+
+       if (ring->read == ring->base_end)
+               ring->read = ring->base;
+       else
+               ring->read += ring->offset;
+
+       return ptr;
+}
+
+int eip93_put_descriptor(struct eip93_device *eip93,
+                        struct eip93_descriptor *desc)
+{
+       struct eip93_descriptor *cdesc;
+       struct eip93_descriptor *rdesc;
+
+       rdesc = eip93_ring_next_wptr(eip93, &eip93->ring->rdr);
+       if (IS_ERR(rdesc))
+               return -ENOENT;
+
+       cdesc = eip93_ring_next_wptr(eip93, &eip93->ring->cdr);
+       if (IS_ERR(cdesc))
+               return -ENOENT;
+
+       memset(rdesc, 0, sizeof(struct eip93_descriptor));
+
+       memcpy(cdesc, desc, sizeof(struct eip93_descriptor));
+
+       return 0;
+}
+
+void *eip93_get_descriptor(struct eip93_device *eip93)
+{
+       struct eip93_descriptor *cdesc;
+       void *ptr;
+
+       cdesc = eip93_ring_next_rptr(eip93, &eip93->ring->cdr);
+       if (IS_ERR(cdesc))
+               return ERR_PTR(-ENOENT);
+
+       memset(cdesc, 0, sizeof(struct eip93_descriptor));
+
+       ptr = eip93_ring_next_rptr(eip93, &eip93->ring->rdr);
+       if (IS_ERR(ptr))
+               return ERR_PTR(-ENOENT);
+
+       return ptr;
+}
+
+static void eip93_free_sg_copy(const int len, struct scatterlist **sg)
+{
+       if (!*sg || !len)
+               return;
+
+       free_pages((unsigned long)sg_virt(*sg), get_order(len));
+       kfree(*sg);
+       *sg = NULL;
+}
+
+static int eip93_make_sg_copy(struct scatterlist *src, struct scatterlist **dst,
+                             const u32 len, const bool copy)
+{
+       void *pages;
+
+       *dst = kmalloc(sizeof(**dst), GFP_KERNEL);
+       if (!*dst)
+               return -ENOMEM;
+
+       pages = (void *)__get_free_pages(GFP_KERNEL | GFP_DMA,
+                                        get_order(len));
+       if (!pages) {
+               kfree(*dst);
+               *dst = NULL;
+               return -ENOMEM;
+       }
+
+       sg_init_table(*dst, 1);
+       sg_set_buf(*dst, pages, len);
+
+       /* copy only as requested */
+       if (copy)
+               sg_copy_to_buffer(src, sg_nents(src), pages, len);
+
+       return 0;
+}
+
+static bool eip93_is_sg_aligned(struct scatterlist *sg, u32 len,
+                               const int blksize)
+{
+       int nents;
+
+       for (nents = 0; sg; sg = sg_next(sg), ++nents) {
+               if (!IS_ALIGNED(sg->offset, 4))
+                       return false;
+
+               if (len <= sg->length) {
+                       if (!IS_ALIGNED(len, blksize))
+                               return false;
+
+                       return true;
+               }
+
+               if (!IS_ALIGNED(sg->length, blksize))
+                       return false;
+
+               len -= sg->length;
+       }
+       return false;
+}
+
+int check_valid_request(struct eip93_cipher_reqctx *rctx)
+{
+       struct scatterlist *src = rctx->sg_src;
+       struct scatterlist *dst = rctx->sg_dst;
+       u32 src_nents, dst_nents;
+       u32 textsize = rctx->textsize;
+       u32 authsize = rctx->authsize;
+       u32 blksize = rctx->blksize;
+       u32 totlen_src = rctx->assoclen + rctx->textsize;
+       u32 totlen_dst = rctx->assoclen + rctx->textsize;
+       u32 copy_len;
+       bool src_align, dst_align;
+       int err = -EINVAL;
+
+       if (!IS_CTR(rctx->flags)) {
+               if (!IS_ALIGNED(textsize, blksize))
+                       return err;
+       }
+
+       if (authsize) {
+               if (IS_ENCRYPT(rctx->flags))
+                       totlen_dst += authsize;
+               else
+                       totlen_src += authsize;
+       }
+
+       src_nents = sg_nents_for_len(src, totlen_src);
+       dst_nents = sg_nents_for_len(dst, totlen_dst);
+
+       if (src == dst) {
+               src_nents = max(src_nents, dst_nents);
+               dst_nents = src_nents;
+               if (unlikely((totlen_src || totlen_dst) && src_nents <= 0))
+                       return err;
+
+       } else {
+               if (unlikely(totlen_src && src_nents <= 0))
+                       return err;
+
+               if (unlikely(totlen_dst && dst_nents <= 0))
+                       return err;
+       }
+
+       if (authsize) {
+               if (dst_nents == 1 && src_nents == 1) {
+                       src_align = eip93_is_sg_aligned(src, totlen_src, blksize);
+                       if (src ==  dst)
+                               dst_align = src_align;
+                       else
+                               dst_align = eip93_is_sg_aligned(dst, totlen_dst, blksize);
+               } else {
+                       src_align = false;
+                       dst_align = false;
+               }
+       } else {
+               src_align = eip93_is_sg_aligned(src, totlen_src, blksize);
+               if (src == dst)
+                       dst_align = src_align;
+               else
+                       dst_align = eip93_is_sg_aligned(dst, totlen_dst, blksize);
+       }
+
+       copy_len = max(totlen_src, totlen_dst);
+       if (!src_align) {
+               err = eip93_make_sg_copy(src, &rctx->sg_src, copy_len, true);
+               if (err)
+                       return err;
+       }
+
+       if (!dst_align) {
+               err = eip93_make_sg_copy(dst, &rctx->sg_dst, copy_len, false);
+               if (err)
+                       return err;
+       }
+
+       rctx->src_nents = sg_nents_for_len(rctx->sg_src, totlen_src);
+       rctx->dst_nents = sg_nents_for_len(rctx->sg_dst, totlen_dst);
+
+       return 0;
+}
+
+/*
+ * Set sa_record function:
+ * Even sa_record is set to "0", keep " = 0" for readability.
+ */
+void eip93_set_sa_record(struct sa_record *sa_record, const unsigned int keylen,
+                        const u32 flags)
+{
+       /* Reset cmd word */
+       sa_record->sa_cmd0_word = 0;
+       sa_record->sa_cmd1_word = 0;
+
+       sa_record->sa_cmd0_word |= EIP93_SA_CMD_IV_FROM_STATE;
+       if (!IS_ECB(flags))
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_SAVE_IV;
+
+       sa_record->sa_cmd0_word |= EIP93_SA_CMD_OP_BASIC;
+
+       switch ((flags & EIP93_ALG_MASK)) {
+       case EIP93_ALG_AES:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_CIPHER_AES;
+               sa_record->sa_cmd1_word |= FIELD_PREP(EIP93_SA_CMD_AES_KEY_LENGTH,
+                                                     keylen >> 3);
+               break;
+       case EIP93_ALG_3DES:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_CIPHER_3DES;
+               break;
+       case EIP93_ALG_DES:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_CIPHER_DES;
+               break;
+       default:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_CIPHER_NULL;
+       }
+
+       switch ((flags & EIP93_HASH_MASK)) {
+       case EIP93_HASH_SHA256:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_HASH_SHA256;
+               break;
+       case EIP93_HASH_SHA224:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_HASH_SHA224;
+               break;
+       case EIP93_HASH_SHA1:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_HASH_SHA1;
+               break;
+       case EIP93_HASH_MD5:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_HASH_MD5;
+               break;
+       default:
+               sa_record->sa_cmd0_word |= EIP93_SA_CMD_HASH_NULL;
+       }
+
+       sa_record->sa_cmd0_word |= EIP93_SA_CMD_PAD_ZERO;
+
+       switch ((flags & EIP93_MODE_MASK)) {
+       case EIP93_MODE_CBC:
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_CHIPER_MODE_CBC;
+               break;
+       case EIP93_MODE_CTR:
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_CHIPER_MODE_CTR;
+               break;
+       case EIP93_MODE_ECB:
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_CHIPER_MODE_ECB;
+               break;
+       }
+
+       sa_record->sa_cmd0_word |= EIP93_SA_CMD_DIGEST_3WORD;
+       if (IS_HASH(flags)) {
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_COPY_PAD;
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_COPY_DIGEST;
+       }
+
+       if (IS_HMAC(flags)) {
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_HMAC;
+               sa_record->sa_cmd1_word |= EIP93_SA_CMD_COPY_HEADER;
+       }
+
+       sa_record->sa_spi = 0x0;
+       sa_record->sa_seqmum_mask[0] = 0xFFFFFFFF;
+       sa_record->sa_seqmum_mask[1] = 0x0;
+}
+
+/*
+ * Poor mans Scatter/gather function:
+ * Create a Descriptor for every segment to avoid copying buffers.
+ * For performance better to wait for hardware to perform multiple DMA
+ */
+static int eip93_scatter_combine(struct eip93_device *eip93,
+                                struct eip93_cipher_reqctx *rctx,
+                                u32 datalen, u32 split, int offsetin)
+{
+       struct eip93_descriptor *cdesc = rctx->cdesc;
+       struct scatterlist *sgsrc = rctx->sg_src;
+       struct scatterlist *sgdst = rctx->sg_dst;
+       unsigned int remainin = sg_dma_len(sgsrc);
+       unsigned int remainout = sg_dma_len(sgdst);
+       dma_addr_t saddr = sg_dma_address(sgsrc);
+       dma_addr_t daddr = sg_dma_address(sgdst);
+       dma_addr_t state_addr;
+       u32 src_addr, dst_addr, len, n;
+       bool nextin = false;
+       bool nextout = false;
+       int offsetout = 0;
+       int err;
+
+       if (IS_ECB(rctx->flags))
+               rctx->sa_state_base = 0;
+
+       if (split < datalen) {
+               state_addr = rctx->sa_state_ctr_base;
+               n = split;
+       } else {
+               state_addr = rctx->sa_state_base;
+               n = datalen;
+       }
+
+       do {
+               if (nextin) {
+                       sgsrc = sg_next(sgsrc);
+                       remainin = sg_dma_len(sgsrc);
+                       if (remainin == 0)
+                               continue;
+
+                       saddr = sg_dma_address(sgsrc);
+                       offsetin = 0;
+                       nextin = false;
+               }
+
+               if (nextout) {
+                       sgdst = sg_next(sgdst);
+                       remainout = sg_dma_len(sgdst);
+                       if (remainout == 0)
+                               continue;
+
+                       daddr = sg_dma_address(sgdst);
+                       offsetout = 0;
+                       nextout = false;
+               }
+               src_addr = saddr + offsetin;
+               dst_addr = daddr + offsetout;
+
+               if (remainin == remainout) {
+                       len = remainin;
+                       if (len > n) {
+                               len = n;
+                               remainin -= n;
+                               remainout -= n;
+                               offsetin += n;
+                               offsetout += n;
+                       } else {
+                               nextin = true;
+                               nextout = true;
+                       }
+               } else if (remainin < remainout) {
+                       len = remainin;
+                       if (len > n) {
+                               len = n;
+                               remainin -= n;
+                               remainout -= n;
+                               offsetin += n;
+                               offsetout += n;
+                       } else {
+                               offsetout += len;
+                               remainout -= len;
+                               nextin = true;
+                       }
+               } else {
+                       len = remainout;
+                       if (len > n) {
+                               len = n;
+                               remainin -= n;
+                               remainout -= n;
+                               offsetin += n;
+                               offsetout += n;
+                       } else {
+                               offsetin += len;
+                               remainin -= len;
+                               nextout = true;
+                       }
+               }
+               n -= len;
+
+               cdesc->src_addr = src_addr;
+               cdesc->dst_addr = dst_addr;
+               cdesc->state_addr = state_addr;
+               cdesc->pe_length_word = FIELD_PREP(EIP93_PE_LENGTH_HOST_PE_READY,
+                                                  EIP93_PE_LENGTH_HOST_READY);
+               cdesc->pe_length_word |= FIELD_PREP(EIP93_PE_LENGTH_LENGTH, len);
+
+               if (n == 0) {
+                       n = datalen - split;
+                       split = datalen;
+                       state_addr = rctx->sa_state_base;
+               }
+
+               if (n == 0)
+                       cdesc->user_id |= FIELD_PREP(EIP93_PE_USER_ID_DESC_FLAGS,
+                                                    EIP93_DESC_LAST);
+
+               /*
+                * Loop - Delay - No need to rollback
+                * Maybe refine by slowing down at EIP93_RING_BUSY
+                */
+again:
+               scoped_guard(spinlock_irqsave, &eip93->ring->write_lock)
+                       err = eip93_put_descriptor(eip93, cdesc);
+               if (err) {
+                       usleep_range(EIP93_RING_BUSY_DELAY,
+                                    EIP93_RING_BUSY_DELAY * 2);
+                       goto again;
+               }
+               /* Writing new descriptor count starts DMA action */
+               writel(1, eip93->base + EIP93_REG_PE_CD_COUNT);
+       } while (n);
+
+       return -EINPROGRESS;
+}
+
+int eip93_send_req(struct crypto_async_request *async,
+                  const u8 *reqiv, struct eip93_cipher_reqctx *rctx)
+{
+       struct eip93_crypto_ctx *ctx = crypto_tfm_ctx(async->tfm);
+       struct eip93_device *eip93 = ctx->eip93;
+       struct scatterlist *src = rctx->sg_src;
+       struct scatterlist *dst = rctx->sg_dst;
+       struct sa_state *sa_state;
+       struct eip93_descriptor cdesc;
+       u32 flags = rctx->flags;
+       int offsetin = 0, err;
+       u32 datalen = rctx->assoclen + rctx->textsize;
+       u32 split = datalen;
+       u32 start, end, ctr, blocks;
+       u32 iv[AES_BLOCK_SIZE / sizeof(u32)];
+       int crypto_async_idr;
+
+       rctx->sa_state_ctr = NULL;
+       rctx->sa_state = NULL;
+
+       if (IS_ECB(flags))
+               goto skip_iv;
+
+       memcpy(iv, reqiv, rctx->ivsize);
+
+       rctx->sa_state = kzalloc(sizeof(*rctx->sa_state), GFP_KERNEL);
+       if (!rctx->sa_state)
+               return -ENOMEM;
+
+       sa_state = rctx->sa_state;
+
+       memcpy(sa_state->state_iv, iv, rctx->ivsize);
+       if (IS_RFC3686(flags)) {
+               sa_state->state_iv[0] = ctx->sa_nonce;
+               sa_state->state_iv[1] = iv[0];
+               sa_state->state_iv[2] = iv[1];
+               sa_state->state_iv[3] = (u32 __force)cpu_to_be32(0x1);
+       } else if (!IS_HMAC(flags) && IS_CTR(flags)) {
+               /* Compute data length. */
+               blocks = DIV_ROUND_UP(rctx->textsize, AES_BLOCK_SIZE);
+               ctr = be32_to_cpu((__be32 __force)iv[3]);
+               /* Check 32bit counter overflow. */
+               start = ctr;
+               end = start + blocks - 1;
+               if (end < start) {
+                       split = AES_BLOCK_SIZE * -start;
+                       /*
+                        * Increment the counter manually to cope with
+                        * the hardware counter overflow.
+                        */
+                       iv[3] = 0xffffffff;
+                       crypto_inc((u8 *)iv, AES_BLOCK_SIZE);
+
+                       rctx->sa_state_ctr = kzalloc(sizeof(*rctx->sa_state_ctr),
+                                                    GFP_KERNEL);
+                       if (!rctx->sa_state_ctr) {
+                               err = -ENOMEM;
+                               goto free_sa_state;
+                       }
+
+                       memcpy(rctx->sa_state_ctr->state_iv, reqiv, rctx->ivsize);
+                       memcpy(sa_state->state_iv, iv, rctx->ivsize);
+
+                       rctx->sa_state_ctr_base = dma_map_single(eip93->dev, rctx->sa_state_ctr,
+                                                                sizeof(*rctx->sa_state_ctr),
+                                                                DMA_TO_DEVICE);
+                       err = dma_mapping_error(eip93->dev, rctx->sa_state_ctr_base);
+                       if (err)
+                               goto free_sa_state_ctr;
+               }
+       }
+
+       rctx->sa_state_base = dma_map_single(eip93->dev, rctx->sa_state,
+                                            sizeof(*rctx->sa_state), DMA_TO_DEVICE);
+       err = dma_mapping_error(eip93->dev, rctx->sa_state_base);
+       if (err)
+               goto free_sa_state_ctr_dma;
+
+skip_iv:
+
+       cdesc.pe_ctrl_stat_word = FIELD_PREP(EIP93_PE_CTRL_PE_READY_DES_TRING_OWN,
+                                            EIP93_PE_CTRL_HOST_READY);
+       cdesc.sa_addr = rctx->sa_record_base;
+       cdesc.arc4_addr = 0;
+
+       scoped_guard(spinlock_bh, &eip93->ring->idr_lock)
+               crypto_async_idr = idr_alloc(&eip93->ring->crypto_async_idr, async, 0,
+                                            EIP93_RING_NUM - 1, GFP_ATOMIC);
+
+       cdesc.user_id = FIELD_PREP(EIP93_PE_USER_ID_CRYPTO_IDR, (u16)crypto_async_idr) |
+                       FIELD_PREP(EIP93_PE_USER_ID_DESC_FLAGS, rctx->desc_flags);
+
+       rctx->cdesc = &cdesc;
+
+       /* map DMA_BIDIRECTIONAL to invalidate cache on destination
+        * implies __dma_cache_wback_inv
+        */
+       if (!dma_map_sg(eip93->dev, dst, rctx->dst_nents, DMA_BIDIRECTIONAL)) {
+               err = -ENOMEM;
+               goto free_sa_state_ctr_dma;
+       }
+
+       if (src != dst &&
+           !dma_map_sg(eip93->dev, src, rctx->src_nents, DMA_TO_DEVICE)) {
+               err = -ENOMEM;
+               goto free_sg_dma;
+       }
+
+       return eip93_scatter_combine(eip93, rctx, datalen, split, offsetin);
+
+free_sg_dma:
+       dma_unmap_sg(eip93->dev, dst, rctx->dst_nents, DMA_BIDIRECTIONAL);
+free_sa_state_ctr_dma:
+       if (rctx->sa_state_ctr)
+               dma_unmap_single(eip93->dev, rctx->sa_state_ctr_base,
+                                sizeof(*rctx->sa_state_ctr),
+                                DMA_TO_DEVICE);
+free_sa_state_ctr:
+       kfree(rctx->sa_state_ctr);
+       if (rctx->sa_state)
+               dma_unmap_single(eip93->dev, rctx->sa_state_base,
+                                sizeof(*rctx->sa_state),
+                                DMA_TO_DEVICE);
+free_sa_state:
+       kfree(rctx->sa_state);
+
+       return err;
+}
+
+void eip93_unmap_dma(struct eip93_device *eip93, struct eip93_cipher_reqctx *rctx,
+                    struct scatterlist *reqsrc, struct scatterlist *reqdst)
+{
+       u32 len = rctx->assoclen + rctx->textsize;
+       u32 authsize = rctx->authsize;
+       u32 flags = rctx->flags;
+       u32 *otag;
+       int i;
+
+       if (rctx->sg_src == rctx->sg_dst) {
+               dma_unmap_sg(eip93->dev, rctx->sg_dst, rctx->dst_nents,
+                            DMA_BIDIRECTIONAL);
+               goto process_tag;
+       }
+
+       dma_unmap_sg(eip93->dev, rctx->sg_src, rctx->src_nents,
+                    DMA_TO_DEVICE);
+
+       if (rctx->sg_src != reqsrc)
+               eip93_free_sg_copy(len +  rctx->authsize, &rctx->sg_src);
+
+       dma_unmap_sg(eip93->dev, rctx->sg_dst, rctx->dst_nents,
+                    DMA_BIDIRECTIONAL);
+
+       /* SHA tags need conversion from net-to-host */
+process_tag:
+       if (IS_DECRYPT(flags))
+               authsize = 0;
+
+       if (authsize) {
+               if (!IS_HASH_MD5(flags)) {
+                       otag = sg_virt(rctx->sg_dst) + len;
+                       for (i = 0; i < (authsize / 4); i++)
+                               otag[i] = be32_to_cpu((__be32 __force)otag[i]);
+               }
+       }
+
+       if (rctx->sg_dst != reqdst) {
+               sg_copy_from_buffer(reqdst, sg_nents(reqdst),
+                                   sg_virt(rctx->sg_dst), len + authsize);
+               eip93_free_sg_copy(len + rctx->authsize, &rctx->sg_dst);
+       }
+}
+
+void eip93_handle_result(struct eip93_device *eip93, struct eip93_cipher_reqctx *rctx,
+                        u8 *reqiv)
+{
+       if (rctx->sa_state_ctr)
+               dma_unmap_single(eip93->dev, rctx->sa_state_ctr_base,
+                                sizeof(*rctx->sa_state_ctr),
+                                DMA_FROM_DEVICE);
+
+       if (rctx->sa_state)
+               dma_unmap_single(eip93->dev, rctx->sa_state_base,
+                                sizeof(*rctx->sa_state),
+                                DMA_FROM_DEVICE);
+
+       if (!IS_ECB(rctx->flags))
+               memcpy(reqiv, rctx->sa_state->state_iv, rctx->ivsize);
+
+       kfree(rctx->sa_state_ctr);
+       kfree(rctx->sa_state);
+}
+
+int eip93_hmac_setkey(u32 ctx_flags, const u8 *key, unsigned int keylen,
+                     unsigned int hashlen, u8 *dest_ipad, u8 *dest_opad,
+                     bool skip_ipad)
+{
+       u8 ipad[SHA256_BLOCK_SIZE], opad[SHA256_BLOCK_SIZE];
+       struct crypto_ahash *ahash_tfm;
+       struct eip93_hash_reqctx *rctx;
+       struct ahash_request *req;
+       DECLARE_CRYPTO_WAIT(wait);
+       struct scatterlist sg[1];
+       const char *alg_name;
+       int i, ret;
+
+       switch (ctx_flags & EIP93_HASH_MASK) {
+       case EIP93_HASH_SHA256:
+               alg_name = "sha256-eip93";
+               break;
+       case EIP93_HASH_SHA224:
+               alg_name = "sha224-eip93";
+               break;
+       case EIP93_HASH_SHA1:
+               alg_name = "sha1-eip93";
+               break;
+       case EIP93_HASH_MD5:
+               alg_name = "md5-eip93";
+               break;
+       default: /* Impossible */
+               return -EINVAL;
+       }
+
+       ahash_tfm = crypto_alloc_ahash(alg_name, 0, CRYPTO_ALG_ASYNC);
+       if (IS_ERR(ahash_tfm))
+               return PTR_ERR(ahash_tfm);
+
+       req = ahash_request_alloc(ahash_tfm, GFP_ATOMIC);
+       if (!req) {
+               ret = -ENOMEM;
+               goto err_ahash;
+       }
+
+       rctx = ahash_request_ctx_dma(req);
+       crypto_init_wait(&wait);
+       ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
+                                  crypto_req_done, &wait);
+
+       /* Hash the key if > SHA256_BLOCK_SIZE */
+       if (keylen > SHA256_BLOCK_SIZE) {
+               sg_init_one(&sg[0], key, keylen);
+
+               ahash_request_set_crypt(req, sg, ipad, keylen);
+               ret = crypto_wait_req(crypto_ahash_digest(req), &wait);
+               if (ret)
+                       goto err_req;
+
+               keylen = hashlen;
+       } else {
+               memcpy(ipad, key, keylen);
+       }
+
+       /* Copy to opad */
+       memset(ipad + keylen, 0, SHA256_BLOCK_SIZE - keylen);
+       memcpy(opad, ipad, SHA256_BLOCK_SIZE);
+
+       /* Pad with HMAC constants */
+       for (i = 0; i < SHA256_BLOCK_SIZE; i++) {
+               ipad[i] ^= HMAC_IPAD_VALUE;
+               opad[i] ^= HMAC_OPAD_VALUE;
+       }
+
+       if (skip_ipad) {
+               memcpy(dest_ipad, ipad, SHA256_BLOCK_SIZE);
+       } else {
+               /* Hash ipad */
+               sg_init_one(&sg[0], ipad, SHA256_BLOCK_SIZE);
+               ahash_request_set_crypt(req, sg, dest_ipad, SHA256_BLOCK_SIZE);
+               ret = crypto_ahash_init(req);
+               if (ret)
+                       goto err_req;
+
+               /* Disable HASH_FINALIZE for ipad hash */
+               rctx->partial_hash = true;
+
+               ret = crypto_wait_req(crypto_ahash_finup(req), &wait);
+               if (ret)
+                       goto err_req;
+       }
+
+       /* Hash opad */
+       sg_init_one(&sg[0], opad, SHA256_BLOCK_SIZE);
+       ahash_request_set_crypt(req, sg, dest_opad, SHA256_BLOCK_SIZE);
+       ret = crypto_ahash_init(req);
+       if (ret)
+               goto err_req;
+
+       /* Disable HASH_FINALIZE for opad hash */
+       rctx->partial_hash = true;
+
+       ret = crypto_wait_req(crypto_ahash_finup(req), &wait);
+       if (ret)
+               goto err_req;
+
+       if (!IS_HASH_MD5(ctx_flags)) {
+               for (i = 0; i < SHA256_DIGEST_SIZE / sizeof(u32); i++) {
+                       u32 *ipad_hash = (u32 *)dest_ipad;
+                       u32 *opad_hash = (u32 *)dest_opad;
+
+                       if (!skip_ipad)
+                               ipad_hash[i] = (u32 __force)cpu_to_be32(ipad_hash[i]);
+                       opad_hash[i] = (u32 __force)cpu_to_be32(opad_hash[i]);
+               }
+       }
+
+err_req:
+       ahash_request_free(req);
+err_ahash:
+       crypto_free_ahash(ahash_tfm);
+
+       return ret;
+}
diff --git a/drivers/crypto/inside-secure/eip93/eip93-common.h b/drivers/crypto/inside-secure/eip93/eip93-common.h
new file mode 100644 (file)
index 0000000..80964cf
--- /dev/null
@@ -0,0 +1,24 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+
+#ifndef _EIP93_COMMON_H_
+#define _EIP93_COMMON_H_
+
+void *eip93_get_descriptor(struct eip93_device *eip93);
+int eip93_put_descriptor(struct eip93_device *eip93, struct eip93_descriptor *desc);
+
+void eip93_set_sa_record(struct sa_record *sa_record, const unsigned int keylen,
+                        const u32 flags);
+
+int eip93_parse_ctrl_stat_err(struct eip93_device *eip93, int err);
+
+int eip93_hmac_setkey(u32 ctx_flags, const u8 *key, unsigned int keylen,
+                     unsigned int hashlen, u8 *ipad, u8 *opad,
+                     bool skip_ipad);
+
+#endif /* _EIP93_COMMON_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-des.h b/drivers/crypto/inside-secure/eip93/eip93-des.h
new file mode 100644 (file)
index 0000000..74748df
--- /dev/null
@@ -0,0 +1,16 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef _EIP93_DES_H_
+#define _EIP93_DES_H_
+
+extern struct eip93_alg_template eip93_alg_ecb_des;
+extern struct eip93_alg_template eip93_alg_cbc_des;
+extern struct eip93_alg_template eip93_alg_ecb_des3_ede;
+extern struct eip93_alg_template eip93_alg_cbc_des3_ede;
+
+#endif /* _EIP93_DES_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-hash.c b/drivers/crypto/inside-secure/eip93/eip93-hash.c
new file mode 100644 (file)
index 0000000..5e96274
--- /dev/null
@@ -0,0 +1,866 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2024
+ *
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+
+#include <crypto/sha1.h>
+#include <crypto/sha2.h>
+#include <crypto/md5.h>
+#include <crypto/hmac.h>
+#include <linux/dma-mapping.h>
+#include <linux/delay.h>
+
+#include "eip93-cipher.h"
+#include "eip93-hash.h"
+#include "eip93-main.h"
+#include "eip93-common.h"
+#include "eip93-regs.h"
+
+static void eip93_hash_free_data_blocks(struct ahash_request *req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct eip93_device *eip93 = ctx->eip93;
+       struct mkt_hash_block *block, *tmp;
+
+       list_for_each_entry_safe(block, tmp, &rctx->blocks, list) {
+               dma_unmap_single(eip93->dev, block->data_dma,
+                                SHA256_BLOCK_SIZE, DMA_TO_DEVICE);
+               kfree(block);
+       }
+       if (!list_empty(&rctx->blocks))
+               INIT_LIST_HEAD(&rctx->blocks);
+
+       if (rctx->finalize)
+               dma_unmap_single(eip93->dev, rctx->data_dma,
+                                rctx->data_used,
+                                DMA_TO_DEVICE);
+}
+
+static void eip93_hash_free_sa_record(struct ahash_request *req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct eip93_device *eip93 = ctx->eip93;
+
+       if (IS_HMAC(ctx->flags))
+               dma_unmap_single(eip93->dev, rctx->sa_record_hmac_base,
+                                sizeof(rctx->sa_record_hmac), DMA_TO_DEVICE);
+
+       dma_unmap_single(eip93->dev, rctx->sa_record_base,
+                        sizeof(rctx->sa_record), DMA_TO_DEVICE);
+}
+
+void eip93_hash_handle_result(struct crypto_async_request *async, int err)
+{
+       struct ahash_request *req = ahash_request_cast(async);
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       int digestsize = crypto_ahash_digestsize(ahash);
+       struct sa_state *sa_state = &rctx->sa_state;
+       struct eip93_device *eip93 = ctx->eip93;
+       int i;
+
+       dma_unmap_single(eip93->dev, rctx->sa_state_base,
+                        sizeof(*sa_state), DMA_FROM_DEVICE);
+
+       /*
+        * With partial_hash assume SHA256_DIGEST_SIZE buffer is passed.
+        * This is to handle SHA224 that have a 32 byte intermediate digest.
+        */
+       if (rctx->partial_hash)
+               digestsize = SHA256_DIGEST_SIZE;
+
+       if (rctx->finalize || rctx->partial_hash) {
+               /* bytes needs to be swapped for req->result */
+               if (!IS_HASH_MD5(ctx->flags)) {
+                       for (i = 0; i < digestsize / sizeof(u32); i++) {
+                               u32 *digest = (u32 *)sa_state->state_i_digest;
+
+                               digest[i] = be32_to_cpu((__be32 __force)digest[i]);
+                       }
+               }
+
+               memcpy(req->result, sa_state->state_i_digest, digestsize);
+       }
+
+       eip93_hash_free_sa_record(req);
+       eip93_hash_free_data_blocks(req);
+
+       ahash_request_complete(req, err);
+}
+
+static void eip93_hash_init_sa_state_digest(u32 hash, u8 *digest)
+{
+       u32 sha256_init[] = { SHA256_H0, SHA256_H1, SHA256_H2, SHA256_H3,
+                             SHA256_H4, SHA256_H5, SHA256_H6, SHA256_H7 };
+       u32 sha224_init[] = { SHA224_H0, SHA224_H1, SHA224_H2, SHA224_H3,
+                             SHA224_H4, SHA224_H5, SHA224_H6, SHA224_H7 };
+       u32 sha1_init[] = { SHA1_H0, SHA1_H1, SHA1_H2, SHA1_H3, SHA1_H4 };
+       u32 md5_init[] = { MD5_H0, MD5_H1, MD5_H2, MD5_H3 };
+
+       /* Init HASH constant */
+       switch (hash) {
+       case EIP93_HASH_SHA256:
+               memcpy(digest, sha256_init, sizeof(sha256_init));
+               return;
+       case EIP93_HASH_SHA224:
+               memcpy(digest, sha224_init, sizeof(sha224_init));
+               return;
+       case EIP93_HASH_SHA1:
+               memcpy(digest, sha1_init, sizeof(sha1_init));
+               return;
+       case EIP93_HASH_MD5:
+               memcpy(digest, md5_init, sizeof(md5_init));
+               return;
+       default: /* Impossible */
+               return;
+       }
+}
+
+static void eip93_hash_export_sa_state(struct ahash_request *req,
+                                      struct eip93_hash_export_state *state)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct sa_state *sa_state = &rctx->sa_state;
+
+       /*
+        * EIP93 have special handling for state_byte_cnt in sa_state.
+        * Even if a zero packet is passed (and a BADMSG is returned),
+        * state_byte_cnt is incremented to the digest handled (with the hash
+        * primitive). This is problematic with export/import as EIP93
+        * expect 0 state_byte_cnt for the very first iteration.
+        */
+       if (!rctx->len)
+               memset(state->state_len, 0, sizeof(u32) * 2);
+       else
+               memcpy(state->state_len, sa_state->state_byte_cnt,
+                      sizeof(u32) * 2);
+       memcpy(state->state_hash, sa_state->state_i_digest,
+              SHA256_DIGEST_SIZE);
+       state->len = rctx->len;
+       state->data_used = rctx->data_used;
+}
+
+static void __eip93_hash_init(struct ahash_request *req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct sa_record *sa_record = &rctx->sa_record;
+       int digestsize;
+
+       digestsize = crypto_ahash_digestsize(ahash);
+
+       eip93_set_sa_record(sa_record, 0, ctx->flags);
+       sa_record->sa_cmd0_word |= EIP93_SA_CMD_HASH_FROM_STATE;
+       sa_record->sa_cmd0_word |= EIP93_SA_CMD_SAVE_HASH;
+       sa_record->sa_cmd0_word &= ~EIP93_SA_CMD_OPCODE;
+       sa_record->sa_cmd0_word |= FIELD_PREP(EIP93_SA_CMD_OPCODE,
+                                             EIP93_SA_CMD_OPCODE_BASIC_OUT_HASH);
+       sa_record->sa_cmd0_word &= ~EIP93_SA_CMD_DIGEST_LENGTH;
+       sa_record->sa_cmd0_word |= FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH,
+                                             digestsize / sizeof(u32));
+
+       /*
+        * HMAC special handling
+        * Enabling CMD_HMAC force the inner hash to be always finalized.
+        * This cause problems on handling message > 64 byte as we
+        * need to produce intermediate inner hash on sending intermediate
+        * 64 bytes blocks.
+        *
+        * To handle this, enable CMD_HMAC only on the last block.
+        * We make a duplicate of sa_record and on the last descriptor,
+        * we pass a dedicated sa_record with CMD_HMAC enabled to make
+        * EIP93 apply the outer hash.
+        */
+       if (IS_HMAC(ctx->flags)) {
+               struct sa_record *sa_record_hmac = &rctx->sa_record_hmac;
+
+               memcpy(sa_record_hmac, sa_record, sizeof(*sa_record));
+               /* Copy pre-hashed opad for HMAC */
+               memcpy(sa_record_hmac->sa_o_digest, ctx->opad, SHA256_DIGEST_SIZE);
+
+               /* Disable HMAC for hash normal sa_record */
+               sa_record->sa_cmd1_word &= ~EIP93_SA_CMD_HMAC;
+       }
+
+       rctx->len = 0;
+       rctx->data_used = 0;
+       rctx->partial_hash = false;
+       rctx->finalize = false;
+       INIT_LIST_HEAD(&rctx->blocks);
+}
+
+static int eip93_send_hash_req(struct crypto_async_request *async, u8 *data,
+                              dma_addr_t *data_dma, u32 len, bool last)
+{
+       struct ahash_request *req = ahash_request_cast(async);
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct eip93_device *eip93 = ctx->eip93;
+       struct eip93_descriptor cdesc = { };
+       dma_addr_t src_addr;
+       int ret;
+
+       /* Map block data to DMA */
+       src_addr = dma_map_single(eip93->dev, data, len, DMA_TO_DEVICE);
+       ret = dma_mapping_error(eip93->dev, src_addr);
+       if (ret)
+               return ret;
+
+       cdesc.pe_ctrl_stat_word = FIELD_PREP(EIP93_PE_CTRL_PE_READY_DES_TRING_OWN,
+                                            EIP93_PE_CTRL_HOST_READY);
+       cdesc.sa_addr = rctx->sa_record_base;
+       cdesc.arc4_addr = 0;
+
+       cdesc.state_addr = rctx->sa_state_base;
+       cdesc.src_addr = src_addr;
+       cdesc.pe_length_word = FIELD_PREP(EIP93_PE_LENGTH_HOST_PE_READY,
+                                         EIP93_PE_LENGTH_HOST_READY);
+       cdesc.pe_length_word |= FIELD_PREP(EIP93_PE_LENGTH_LENGTH,
+                                          len);
+
+       cdesc.user_id |= FIELD_PREP(EIP93_PE_USER_ID_DESC_FLAGS, EIP93_DESC_HASH);
+
+       if (last) {
+               int crypto_async_idr;
+
+               if (rctx->finalize && !rctx->partial_hash) {
+                       /* For last block, pass sa_record with CMD_HMAC enabled */
+                       if (IS_HMAC(ctx->flags)) {
+                               struct sa_record *sa_record_hmac = &rctx->sa_record_hmac;
+
+                               rctx->sa_record_hmac_base = dma_map_single(eip93->dev,
+                                                                          sa_record_hmac,
+                                                                          sizeof(*sa_record_hmac),
+                                                                          DMA_TO_DEVICE);
+                               ret = dma_mapping_error(eip93->dev, rctx->sa_record_hmac_base);
+                               if (ret)
+                                       return ret;
+
+                               cdesc.sa_addr = rctx->sa_record_hmac_base;
+                       }
+
+                       cdesc.pe_ctrl_stat_word |= EIP93_PE_CTRL_PE_HASH_FINAL;
+               }
+
+               scoped_guard(spinlock_bh, &eip93->ring->idr_lock)
+                       crypto_async_idr = idr_alloc(&eip93->ring->crypto_async_idr, async, 0,
+                                                    EIP93_RING_NUM - 1, GFP_ATOMIC);
+
+               cdesc.user_id |= FIELD_PREP(EIP93_PE_USER_ID_CRYPTO_IDR, (u16)crypto_async_idr) |
+                                FIELD_PREP(EIP93_PE_USER_ID_DESC_FLAGS, EIP93_DESC_LAST);
+       }
+
+again:
+       ret = eip93_put_descriptor(eip93, &cdesc);
+       if (ret) {
+               usleep_range(EIP93_RING_BUSY_DELAY,
+                            EIP93_RING_BUSY_DELAY * 2);
+               goto again;
+       }
+
+       /* Writing new descriptor count starts DMA action */
+       writel(1, eip93->base + EIP93_REG_PE_CD_COUNT);
+
+       *data_dma = src_addr;
+       return 0;
+}
+
+static int eip93_hash_init(struct ahash_request *req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct sa_state *sa_state = &rctx->sa_state;
+
+       memset(sa_state->state_byte_cnt, 0, sizeof(u32) * 2);
+       eip93_hash_init_sa_state_digest(ctx->flags & EIP93_HASH_MASK,
+                                       sa_state->state_i_digest);
+
+       __eip93_hash_init(req);
+
+       /* For HMAC setup the initial block for ipad */
+       if (IS_HMAC(ctx->flags)) {
+               memcpy(rctx->data, ctx->ipad, SHA256_BLOCK_SIZE);
+
+               rctx->data_used = SHA256_BLOCK_SIZE;
+               rctx->len += SHA256_BLOCK_SIZE;
+       }
+
+       return 0;
+}
+
+/*
+ * With complete_req true, we wait for the engine to consume all the block in list,
+ * else we just queue the block to the engine as final() will wait. This is useful
+ * for finup().
+ */
+static int __eip93_hash_update(struct ahash_request *req, bool complete_req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_async_request *async = &req->base;
+       unsigned int read, to_consume = req->nbytes;
+       unsigned int max_read, consumed = 0;
+       struct mkt_hash_block *block;
+       bool wait_req = false;
+       int offset;
+       int ret;
+
+       /* Get the offset and available space to fill req data */
+       offset = rctx->data_used;
+       max_read = SHA256_BLOCK_SIZE - offset;
+
+       /* Consume req in block of SHA256_BLOCK_SIZE.
+        * to_read is initially set to space available in the req data
+        * and then reset to SHA256_BLOCK_SIZE.
+        */
+       while (to_consume > max_read) {
+               block = kzalloc(sizeof(*block), GFP_ATOMIC);
+               if (!block) {
+                       ret = -ENOMEM;
+                       goto free_blocks;
+               }
+
+               read = sg_pcopy_to_buffer(req->src, sg_nents(req->src),
+                                         block->data + offset,
+                                         max_read, consumed);
+
+               /*
+                * For first iteration only, copy req data to block
+                * and reset offset and max_read for next iteration.
+                */
+               if (offset > 0) {
+                       memcpy(block->data, rctx->data, offset);
+                       offset = 0;
+                       max_read = SHA256_BLOCK_SIZE;
+               }
+
+               list_add(&block->list, &rctx->blocks);
+               to_consume -= read;
+               consumed += read;
+       }
+
+       /* Write the remaining data to req data */
+       read = sg_pcopy_to_buffer(req->src, sg_nents(req->src),
+                                 rctx->data + offset, to_consume,
+                                 consumed);
+       rctx->data_used = offset + read;
+
+       /* Update counter with processed bytes */
+       rctx->len += read + consumed;
+
+       /* Consume all the block added to list */
+       list_for_each_entry_reverse(block, &rctx->blocks, list) {
+               wait_req = complete_req &&
+                           list_is_first(&block->list, &rctx->blocks);
+
+               ret = eip93_send_hash_req(async, block->data,
+                                         &block->data_dma,
+                                         SHA256_BLOCK_SIZE, wait_req);
+               if (ret)
+                       goto free_blocks;
+       }
+
+       return wait_req ? -EINPROGRESS : 0;
+
+free_blocks:
+       eip93_hash_free_data_blocks(req);
+
+       return ret;
+}
+
+static int eip93_hash_update(struct ahash_request *req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct sa_record *sa_record = &rctx->sa_record;
+       struct sa_state *sa_state = &rctx->sa_state;
+       struct eip93_device *eip93 = ctx->eip93;
+       int ret;
+
+       if (!req->nbytes)
+               return 0;
+
+       rctx->sa_state_base = dma_map_single(eip93->dev, sa_state,
+                                            sizeof(*sa_state),
+                                            DMA_TO_DEVICE);
+       ret = dma_mapping_error(eip93->dev, rctx->sa_state_base);
+       if (ret)
+               return ret;
+
+       rctx->sa_record_base = dma_map_single(eip93->dev, sa_record,
+                                             sizeof(*sa_record),
+                                             DMA_TO_DEVICE);
+       ret = dma_mapping_error(eip93->dev, rctx->sa_record_base);
+       if (ret)
+               goto free_sa_state;
+
+       ret = __eip93_hash_update(req, true);
+       if (ret && ret != -EINPROGRESS)
+               goto free_sa_record;
+
+       return ret;
+
+free_sa_record:
+       dma_unmap_single(eip93->dev, rctx->sa_record_base,
+                        sizeof(*sa_record), DMA_TO_DEVICE);
+
+free_sa_state:
+       dma_unmap_single(eip93->dev, rctx->sa_state_base,
+                        sizeof(*sa_state), DMA_TO_DEVICE);
+
+       return ret;
+}
+
+/*
+ * With map_data true, we map the sa_record and sa_state. This is needed
+ * for finup() as the they are mapped before calling update()
+ */
+static int __eip93_hash_final(struct ahash_request *req, bool map_dma)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct crypto_async_request *async = &req->base;
+       struct sa_record *sa_record = &rctx->sa_record;
+       struct sa_state *sa_state = &rctx->sa_state;
+       struct eip93_device *eip93 = ctx->eip93;
+       int ret;
+
+       /* EIP93 can't handle zero bytes hash */
+       if (!rctx->len && !IS_HMAC(ctx->flags)) {
+               switch ((ctx->flags & EIP93_HASH_MASK)) {
+               case EIP93_HASH_SHA256:
+                       memcpy(req->result, sha256_zero_message_hash,
+                              SHA256_DIGEST_SIZE);
+                       break;
+               case EIP93_HASH_SHA224:
+                       memcpy(req->result, sha224_zero_message_hash,
+                              SHA224_DIGEST_SIZE);
+                       break;
+               case EIP93_HASH_SHA1:
+                       memcpy(req->result, sha1_zero_message_hash,
+                              SHA1_DIGEST_SIZE);
+                       break;
+               case EIP93_HASH_MD5:
+                       memcpy(req->result, md5_zero_message_hash,
+                              MD5_DIGEST_SIZE);
+                       break;
+               default: /* Impossible */
+                       return -EINVAL;
+               }
+
+               return 0;
+       }
+
+       /* Signal interrupt from engine is for last block */
+       rctx->finalize = true;
+
+       if (map_dma) {
+               rctx->sa_state_base = dma_map_single(eip93->dev, sa_state,
+                                                    sizeof(*sa_state),
+                                                    DMA_TO_DEVICE);
+               ret = dma_mapping_error(eip93->dev, rctx->sa_state_base);
+               if (ret)
+                       return ret;
+
+               rctx->sa_record_base = dma_map_single(eip93->dev, sa_record,
+                                                     sizeof(*sa_record),
+                                                     DMA_TO_DEVICE);
+               ret = dma_mapping_error(eip93->dev, rctx->sa_record_base);
+               if (ret)
+                       goto free_sa_state;
+       }
+
+       /* Send last block */
+       ret = eip93_send_hash_req(async, rctx->data, &rctx->data_dma,
+                                 rctx->data_used, true);
+       if (ret)
+               goto free_blocks;
+
+       return -EINPROGRESS;
+
+free_blocks:
+       eip93_hash_free_data_blocks(req);
+
+       dma_unmap_single(eip93->dev, rctx->sa_record_base,
+                        sizeof(*sa_record), DMA_TO_DEVICE);
+
+free_sa_state:
+       dma_unmap_single(eip93->dev, rctx->sa_state_base,
+                        sizeof(*sa_state), DMA_TO_DEVICE);
+
+       return ret;
+}
+
+static int eip93_hash_final(struct ahash_request *req)
+{
+       return __eip93_hash_final(req, true);
+}
+
+static int eip93_hash_finup(struct ahash_request *req)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct crypto_ahash *ahash = crypto_ahash_reqtfm(req);
+       struct eip93_hash_ctx *ctx = crypto_ahash_ctx(ahash);
+       struct sa_record *sa_record = &rctx->sa_record;
+       struct sa_state *sa_state = &rctx->sa_state;
+       struct eip93_device *eip93 = ctx->eip93;
+       int ret;
+
+       if (rctx->len + req->nbytes || IS_HMAC(ctx->flags)) {
+               rctx->sa_state_base = dma_map_single(eip93->dev, sa_state,
+                                                    sizeof(*sa_state),
+                                                    DMA_TO_DEVICE);
+               ret = dma_mapping_error(eip93->dev, rctx->sa_state_base);
+               if (ret)
+                       return ret;
+
+               rctx->sa_record_base = dma_map_single(eip93->dev, sa_record,
+                                                     sizeof(*sa_record),
+                                                     DMA_TO_DEVICE);
+               ret = dma_mapping_error(eip93->dev, rctx->sa_record_base);
+               if (ret)
+                       goto free_sa_state;
+
+               ret = __eip93_hash_update(req, false);
+               if (ret)
+                       goto free_sa_record;
+       }
+
+       return __eip93_hash_final(req, false);
+
+free_sa_record:
+       dma_unmap_single(eip93->dev, rctx->sa_record_base,
+                        sizeof(*sa_record), DMA_TO_DEVICE);
+free_sa_state:
+       dma_unmap_single(eip93->dev, rctx->sa_state_base,
+                        sizeof(*sa_state), DMA_TO_DEVICE);
+
+       return ret;
+}
+
+static int eip93_hash_hmac_setkey(struct crypto_ahash *ahash, const u8 *key,
+                                 u32 keylen)
+{
+       unsigned int digestsize = crypto_ahash_digestsize(ahash);
+       struct crypto_tfm *tfm = crypto_ahash_tfm(ahash);
+       struct eip93_hash_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       return eip93_hmac_setkey(ctx->flags, key, keylen, digestsize,
+                                ctx->ipad, ctx->opad, true);
+}
+
+static int eip93_hash_cra_init(struct crypto_tfm *tfm)
+{
+       struct eip93_hash_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct eip93_alg_template *tmpl = container_of(tfm->__crt_alg,
+                               struct eip93_alg_template, alg.ahash.halg.base);
+
+       crypto_ahash_set_reqsize_dma(__crypto_ahash_cast(tfm),
+                                    sizeof(struct eip93_hash_reqctx));
+
+       ctx->eip93 = tmpl->eip93;
+       ctx->flags = tmpl->flags;
+
+       return 0;
+}
+
+static int eip93_hash_digest(struct ahash_request *req)
+{
+       int ret;
+
+       ret = eip93_hash_init(req);
+       if (ret)
+               return ret;
+
+       return eip93_hash_finup(req);
+}
+
+static int eip93_hash_import(struct ahash_request *req, const void *in)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       const struct eip93_hash_export_state *state = in;
+       struct sa_state *sa_state = &rctx->sa_state;
+
+       memcpy(sa_state->state_byte_cnt, state->state_len, sizeof(u32) * 2);
+       memcpy(sa_state->state_i_digest, state->state_hash, SHA256_DIGEST_SIZE);
+
+       __eip93_hash_init(req);
+
+       rctx->len = state->len;
+       rctx->data_used = state->data_used;
+
+       /* Skip copying data if we have nothing to copy */
+       if (rctx->len)
+               memcpy(rctx->data, state->data, rctx->data_used);
+
+       return 0;
+}
+
+static int eip93_hash_export(struct ahash_request *req, void *out)
+{
+       struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req);
+       struct eip93_hash_export_state *state = out;
+
+       /* Save the first block in state data */
+       if (rctx->len)
+               memcpy(state->data, rctx->data, rctx->data_used);
+
+       eip93_hash_export_sa_state(req, state);
+
+       return 0;
+}
+
+struct eip93_alg_template eip93_alg_md5 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_MD5,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = MD5_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "md5",
+                               .cra_driver_name = "md5-eip93",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = MD5_HMAC_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_sha1 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_SHA1,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = SHA1_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "sha1",
+                               .cra_driver_name = "sha1-eip93",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = SHA1_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_sha224 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_SHA224,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = SHA224_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "sha224",
+                               .cra_driver_name = "sha224-eip93",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = SHA224_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_sha256 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_SHA256,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = SHA256_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "sha256",
+                               .cra_driver_name = "sha256-eip93",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = SHA256_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_hmac_md5 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_MD5,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .setkey = eip93_hash_hmac_setkey,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = MD5_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "hmac(md5)",
+                               .cra_driver_name = "hmac(md5-eip93)",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = MD5_HMAC_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_hmac_sha1 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA1,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .setkey = eip93_hash_hmac_setkey,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = SHA1_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "hmac(sha1)",
+                               .cra_driver_name = "hmac(sha1-eip93)",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = SHA1_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_hmac_sha224 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA224,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .setkey = eip93_hash_hmac_setkey,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = SHA224_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "hmac(sha224)",
+                               .cra_driver_name = "hmac(sha224-eip93)",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = SHA224_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
+
+struct eip93_alg_template eip93_alg_hmac_sha256 = {
+       .type = EIP93_ALG_TYPE_HASH,
+       .flags = EIP93_HASH_HMAC | EIP93_HASH_SHA256,
+       .alg.ahash = {
+               .init = eip93_hash_init,
+               .update = eip93_hash_update,
+               .final = eip93_hash_final,
+               .finup = eip93_hash_finup,
+               .digest = eip93_hash_digest,
+               .setkey = eip93_hash_hmac_setkey,
+               .export = eip93_hash_export,
+               .import = eip93_hash_import,
+               .halg = {
+                       .digestsize = SHA256_DIGEST_SIZE,
+                       .statesize = sizeof(struct eip93_hash_export_state),
+                       .base = {
+                               .cra_name = "hmac(sha256)",
+                               .cra_driver_name = "hmac(sha256-eip93)",
+                               .cra_priority = 300,
+                               .cra_flags = CRYPTO_ALG_ASYNC |
+                                               CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                               CRYPTO_ALG_ALLOCATES_MEMORY,
+                               .cra_blocksize = SHA256_BLOCK_SIZE,
+                               .cra_ctxsize = sizeof(struct eip93_hash_ctx),
+                               .cra_init = eip93_hash_cra_init,
+                               .cra_module = THIS_MODULE,
+                       },
+               },
+       },
+};
diff --git a/drivers/crypto/inside-secure/eip93/eip93-hash.h b/drivers/crypto/inside-secure/eip93/eip93-hash.h
new file mode 100644 (file)
index 0000000..556f22f
--- /dev/null
@@ -0,0 +1,82 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef _EIP93_HASH_H_
+#define _EIP93_HASH_H_
+
+#include <crypto/sha2.h>
+
+#include "eip93-main.h"
+#include "eip93-regs.h"
+
+struct eip93_hash_ctx {
+       struct eip93_device     *eip93;
+       u32                     flags;
+
+       u8                      ipad[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
+       u8                      opad[SHA256_DIGEST_SIZE] __aligned(sizeof(u32));
+};
+
+struct eip93_hash_reqctx {
+       /* Placement is important for DMA align */
+       struct {
+               struct sa_record        sa_record;
+               struct sa_record        sa_record_hmac;
+               struct sa_state         sa_state;
+       } __aligned(CRYPTO_DMA_ALIGN);
+
+       dma_addr_t              sa_record_base;
+       dma_addr_t              sa_record_hmac_base;
+       dma_addr_t              sa_state_base;
+
+       /* Don't enable HASH_FINALIZE when last block is sent */
+       bool                    partial_hash;
+
+       /* Set to signal interrupt is for final packet */
+       bool                    finalize;
+
+       /*
+        * EIP93 requires data to be accumulated in block of 64 bytes
+        * for intermediate hash calculation.
+        */
+       u64                     len;
+       u32                     data_used;
+
+       u8                      data[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
+       dma_addr_t              data_dma;
+
+       struct list_head        blocks;
+};
+
+struct mkt_hash_block {
+       struct list_head        list;
+       u8                      data[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
+       dma_addr_t              data_dma;
+};
+
+struct eip93_hash_export_state {
+       u64                     len;
+       u32                     data_used;
+
+       u32                     state_len[2];
+       u8                      state_hash[SHA256_DIGEST_SIZE] __aligned(sizeof(u32));
+
+       u8                      data[SHA256_BLOCK_SIZE] __aligned(sizeof(u32));
+};
+
+void eip93_hash_handle_result(struct crypto_async_request *async, int err);
+
+extern struct eip93_alg_template eip93_alg_md5;
+extern struct eip93_alg_template eip93_alg_sha1;
+extern struct eip93_alg_template eip93_alg_sha224;
+extern struct eip93_alg_template eip93_alg_sha256;
+extern struct eip93_alg_template eip93_alg_hmac_md5;
+extern struct eip93_alg_template eip93_alg_hmac_sha1;
+extern struct eip93_alg_template eip93_alg_hmac_sha224;
+extern struct eip93_alg_template eip93_alg_hmac_sha256;
+
+#endif /* _EIP93_HASH_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-main.c b/drivers/crypto/inside-secure/eip93/eip93-main.c
new file mode 100644 (file)
index 0000000..0b38a56
--- /dev/null
@@ -0,0 +1,501 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+
+#include <linux/atomic.h>
+#include <linux/clk.h>
+#include <linux/delay.h>
+#include <linux/dma-mapping.h>
+#include <linux/interrupt.h>
+#include <linux/module.h>
+#include <linux/of.h>
+#include <linux/platform_device.h>
+#include <linux/spinlock.h>
+#include <crypto/aes.h>
+#include <crypto/ctr.h>
+
+#include "eip93-main.h"
+#include "eip93-regs.h"
+#include "eip93-common.h"
+#include "eip93-cipher.h"
+#include "eip93-aes.h"
+#include "eip93-des.h"
+#include "eip93-aead.h"
+#include "eip93-hash.h"
+
+static struct eip93_alg_template *eip93_algs[] = {
+       &eip93_alg_ecb_des,
+       &eip93_alg_cbc_des,
+       &eip93_alg_ecb_des3_ede,
+       &eip93_alg_cbc_des3_ede,
+       &eip93_alg_ecb_aes,
+       &eip93_alg_cbc_aes,
+       &eip93_alg_ctr_aes,
+       &eip93_alg_rfc3686_aes,
+       &eip93_alg_authenc_hmac_md5_cbc_des,
+       &eip93_alg_authenc_hmac_sha1_cbc_des,
+       &eip93_alg_authenc_hmac_sha224_cbc_des,
+       &eip93_alg_authenc_hmac_sha256_cbc_des,
+       &eip93_alg_authenc_hmac_md5_cbc_des3_ede,
+       &eip93_alg_authenc_hmac_sha1_cbc_des3_ede,
+       &eip93_alg_authenc_hmac_sha224_cbc_des3_ede,
+       &eip93_alg_authenc_hmac_sha256_cbc_des3_ede,
+       &eip93_alg_authenc_hmac_md5_cbc_aes,
+       &eip93_alg_authenc_hmac_sha1_cbc_aes,
+       &eip93_alg_authenc_hmac_sha224_cbc_aes,
+       &eip93_alg_authenc_hmac_sha256_cbc_aes,
+       &eip93_alg_authenc_hmac_md5_rfc3686_aes,
+       &eip93_alg_authenc_hmac_sha1_rfc3686_aes,
+       &eip93_alg_authenc_hmac_sha224_rfc3686_aes,
+       &eip93_alg_authenc_hmac_sha256_rfc3686_aes,
+       &eip93_alg_md5,
+       &eip93_alg_sha1,
+       &eip93_alg_sha224,
+       &eip93_alg_sha256,
+       &eip93_alg_hmac_md5,
+       &eip93_alg_hmac_sha1,
+       &eip93_alg_hmac_sha224,
+       &eip93_alg_hmac_sha256,
+};
+
+inline void eip93_irq_disable(struct eip93_device *eip93, u32 mask)
+{
+       __raw_writel(mask, eip93->base + EIP93_REG_MASK_DISABLE);
+}
+
+inline void eip93_irq_enable(struct eip93_device *eip93, u32 mask)
+{
+       __raw_writel(mask, eip93->base + EIP93_REG_MASK_ENABLE);
+}
+
+inline void eip93_irq_clear(struct eip93_device *eip93, u32 mask)
+{
+       __raw_writel(mask, eip93->base + EIP93_REG_INT_CLR);
+}
+
+static void eip93_unregister_algs(unsigned int i)
+{
+       unsigned int j;
+
+       for (j = 0; j < i; j++) {
+               switch (eip93_algs[j]->type) {
+               case EIP93_ALG_TYPE_SKCIPHER:
+                       crypto_unregister_skcipher(&eip93_algs[j]->alg.skcipher);
+                       break;
+               case EIP93_ALG_TYPE_AEAD:
+                       crypto_unregister_aead(&eip93_algs[j]->alg.aead);
+                       break;
+               case EIP93_ALG_TYPE_HASH:
+                       crypto_unregister_ahash(&eip93_algs[i]->alg.ahash);
+                       break;
+               }
+       }
+}
+
+static int eip93_register_algs(struct eip93_device *eip93, u32 supported_algo_flags)
+{
+       unsigned int i;
+       int ret = 0;
+
+       for (i = 0; i < ARRAY_SIZE(eip93_algs); i++) {
+               u32 alg_flags = eip93_algs[i]->flags;
+
+               eip93_algs[i]->eip93 = eip93;
+
+               if ((IS_DES(alg_flags) || IS_3DES(alg_flags)) &&
+                   !(supported_algo_flags & EIP93_PE_OPTION_TDES))
+                       continue;
+
+               if (IS_AES(alg_flags)) {
+                       if (!(supported_algo_flags & EIP93_PE_OPTION_AES))
+                               continue;
+
+                       if (!IS_HMAC(alg_flags)) {
+                               if (supported_algo_flags & EIP93_PE_OPTION_AES_KEY128)
+                                       eip93_algs[i]->alg.skcipher.max_keysize =
+                                               AES_KEYSIZE_128;
+
+                               if (supported_algo_flags & EIP93_PE_OPTION_AES_KEY192)
+                                       eip93_algs[i]->alg.skcipher.max_keysize =
+                                               AES_KEYSIZE_192;
+
+                               if (supported_algo_flags & EIP93_PE_OPTION_AES_KEY256)
+                                       eip93_algs[i]->alg.skcipher.max_keysize =
+                                               AES_KEYSIZE_256;
+
+                               if (IS_RFC3686(alg_flags))
+                                       eip93_algs[i]->alg.skcipher.max_keysize +=
+                                               CTR_RFC3686_NONCE_SIZE;
+                       }
+               }
+
+               if (IS_HASH_MD5(alg_flags) &&
+                   !(supported_algo_flags & EIP93_PE_OPTION_MD5))
+                       continue;
+
+               if (IS_HASH_SHA1(alg_flags) &&
+                   !(supported_algo_flags & EIP93_PE_OPTION_SHA_1))
+                       continue;
+
+               if (IS_HASH_SHA224(alg_flags) &&
+                   !(supported_algo_flags & EIP93_PE_OPTION_SHA_224))
+                       continue;
+
+               if (IS_HASH_SHA256(alg_flags) &&
+                   !(supported_algo_flags & EIP93_PE_OPTION_SHA_256))
+                       continue;
+
+               switch (eip93_algs[i]->type) {
+               case EIP93_ALG_TYPE_SKCIPHER:
+                       ret = crypto_register_skcipher(&eip93_algs[i]->alg.skcipher);
+                       break;
+               case EIP93_ALG_TYPE_AEAD:
+                       ret = crypto_register_aead(&eip93_algs[i]->alg.aead);
+                       break;
+               case EIP93_ALG_TYPE_HASH:
+                       ret = crypto_register_ahash(&eip93_algs[i]->alg.ahash);
+                       break;
+               }
+               if (ret)
+                       goto fail;
+       }
+
+       return 0;
+
+fail:
+       eip93_unregister_algs(i);
+
+       return ret;
+}
+
+static void eip93_handle_result_descriptor(struct eip93_device *eip93)
+{
+       struct crypto_async_request *async;
+       struct eip93_descriptor *rdesc;
+       u16 desc_flags, crypto_idr;
+       bool last_entry;
+       int handled, left, err;
+       u32 pe_ctrl_stat;
+       u32 pe_length;
+
+get_more:
+       handled = 0;
+
+       left = readl(eip93->base + EIP93_REG_PE_RD_COUNT) & EIP93_PE_RD_COUNT;
+
+       if (!left) {
+               eip93_irq_clear(eip93, EIP93_INT_RDR_THRESH);
+               eip93_irq_enable(eip93, EIP93_INT_RDR_THRESH);
+               return;
+       }
+
+       last_entry = false;
+
+       while (left) {
+               scoped_guard(spinlock_irqsave, &eip93->ring->read_lock)
+                       rdesc = eip93_get_descriptor(eip93);
+               if (IS_ERR(rdesc)) {
+                       dev_err(eip93->dev, "Ndesc: %d nreq: %d\n",
+                               handled, left);
+                       err = -EIO;
+                       break;
+               }
+               /* make sure DMA is finished writing */
+               do {
+                       pe_ctrl_stat = READ_ONCE(rdesc->pe_ctrl_stat_word);
+                       pe_length = READ_ONCE(rdesc->pe_length_word);
+               } while (FIELD_GET(EIP93_PE_CTRL_PE_READY_DES_TRING_OWN, pe_ctrl_stat) !=
+                        EIP93_PE_CTRL_PE_READY ||
+                        FIELD_GET(EIP93_PE_LENGTH_HOST_PE_READY, pe_length) !=
+                        EIP93_PE_LENGTH_PE_READY);
+
+               err = rdesc->pe_ctrl_stat_word & (EIP93_PE_CTRL_PE_EXT_ERR_CODE |
+                                                 EIP93_PE_CTRL_PE_EXT_ERR |
+                                                 EIP93_PE_CTRL_PE_SEQNUM_ERR |
+                                                 EIP93_PE_CTRL_PE_PAD_ERR |
+                                                 EIP93_PE_CTRL_PE_AUTH_ERR);
+
+               desc_flags = FIELD_GET(EIP93_PE_USER_ID_DESC_FLAGS, rdesc->user_id);
+               crypto_idr = FIELD_GET(EIP93_PE_USER_ID_CRYPTO_IDR, rdesc->user_id);
+
+               writel(1, eip93->base + EIP93_REG_PE_RD_COUNT);
+               eip93_irq_clear(eip93, EIP93_INT_RDR_THRESH);
+
+               handled++;
+               left--;
+
+               if (desc_flags & EIP93_DESC_LAST) {
+                       last_entry = true;
+                       break;
+               }
+       }
+
+       if (!last_entry)
+               goto get_more;
+
+       /* Get crypto async ref only for last descriptor */
+       scoped_guard(spinlock_bh, &eip93->ring->idr_lock) {
+               async = idr_find(&eip93->ring->crypto_async_idr, crypto_idr);
+               idr_remove(&eip93->ring->crypto_async_idr, crypto_idr);
+       }
+
+       /* Parse error in ctrl stat word */
+       err = eip93_parse_ctrl_stat_err(eip93, err);
+
+       if (desc_flags & EIP93_DESC_SKCIPHER)
+               eip93_skcipher_handle_result(async, err);
+
+       if (desc_flags & EIP93_DESC_AEAD)
+               eip93_aead_handle_result(async, err);
+
+       if (desc_flags & EIP93_DESC_HASH)
+               eip93_hash_handle_result(async, err);
+
+       goto get_more;
+}
+
+static void eip93_done_task(unsigned long data)
+{
+       struct eip93_device *eip93 = (struct eip93_device *)data;
+
+       eip93_handle_result_descriptor(eip93);
+}
+
+static irqreturn_t eip93_irq_handler(int irq, void *data)
+{
+       struct eip93_device *eip93 = data;
+       u32 irq_status;
+
+       irq_status = readl(eip93->base + EIP93_REG_INT_MASK_STAT);
+       if (FIELD_GET(EIP93_INT_RDR_THRESH, irq_status)) {
+               eip93_irq_disable(eip93, EIP93_INT_RDR_THRESH);
+               tasklet_schedule(&eip93->ring->done_task);
+               return IRQ_HANDLED;
+       }
+
+       /* Ignore errors in AUTO mode, handled by the RDR */
+       eip93_irq_clear(eip93, irq_status);
+       if (irq_status)
+               eip93_irq_disable(eip93, irq_status);
+
+       return IRQ_NONE;
+}
+
+static void eip93_initialize(struct eip93_device *eip93, u32 supported_algo_flags)
+{
+       u32 val;
+
+       /* Reset PE and rings */
+       val = EIP93_PE_CONFIG_RST_PE | EIP93_PE_CONFIG_RST_RING;
+       val |= EIP93_PE_TARGET_AUTO_RING_MODE;
+       /* For Auto more, update the CDR ring owner after processing */
+       val |= EIP93_PE_CONFIG_EN_CDR_UPDATE;
+       writel(val, eip93->base + EIP93_REG_PE_CONFIG);
+
+       /* Wait for PE and ring to reset */
+       usleep_range(10, 20);
+
+       /* Release PE and ring reset */
+       val = readl(eip93->base + EIP93_REG_PE_CONFIG);
+       val &= ~(EIP93_PE_CONFIG_RST_PE | EIP93_PE_CONFIG_RST_RING);
+       writel(val, eip93->base + EIP93_REG_PE_CONFIG);
+
+       /* Config Clocks */
+       val = EIP93_PE_CLOCK_EN_PE_CLK;
+       if (supported_algo_flags & EIP93_PE_OPTION_TDES)
+               val |= EIP93_PE_CLOCK_EN_DES_CLK;
+       if (supported_algo_flags & EIP93_PE_OPTION_AES)
+               val |= EIP93_PE_CLOCK_EN_AES_CLK;
+       if (supported_algo_flags &
+           (EIP93_PE_OPTION_MD5 | EIP93_PE_OPTION_SHA_1 | EIP93_PE_OPTION_SHA_224 |
+            EIP93_PE_OPTION_SHA_256))
+               val |= EIP93_PE_CLOCK_EN_HASH_CLK;
+       writel(val, eip93->base + EIP93_REG_PE_CLOCK_CTRL);
+
+       /* Config DMA thresholds */
+       val = FIELD_PREP(EIP93_PE_OUTBUF_THRESH, 128) |
+             FIELD_PREP(EIP93_PE_INBUF_THRESH, 128);
+       writel(val, eip93->base + EIP93_REG_PE_BUF_THRESH);
+
+       /* Clear/ack all interrupts before disable all */
+       eip93_irq_clear(eip93, EIP93_INT_ALL);
+       eip93_irq_disable(eip93, EIP93_INT_ALL);
+
+       /* Setup CRD threshold to trigger interrupt */
+       val = FIELD_PREP(EIPR93_PE_CDR_THRESH, EIP93_RING_NUM - EIP93_RING_BUSY);
+       /*
+        * Configure RDR interrupt to be triggered if RD counter is not 0
+        * for more than 2^(N+10) system clocks.
+        */
+       val |= FIELD_PREP(EIPR93_PE_RD_TIMEOUT, 5) | EIPR93_PE_TIMEROUT_EN;
+       writel(val, eip93->base + EIP93_REG_PE_RING_THRESH);
+}
+
+static void eip93_desc_free(struct eip93_device *eip93)
+{
+       writel(0, eip93->base + EIP93_REG_PE_RING_CONFIG);
+       writel(0, eip93->base + EIP93_REG_PE_CDR_BASE);
+       writel(0, eip93->base + EIP93_REG_PE_RDR_BASE);
+}
+
+static int eip93_set_ring(struct eip93_device *eip93, struct eip93_desc_ring *ring)
+{
+       ring->offset = sizeof(struct eip93_descriptor);
+       ring->base = dmam_alloc_coherent(eip93->dev,
+                                        sizeof(struct eip93_descriptor) * EIP93_RING_NUM,
+                                        &ring->base_dma, GFP_KERNEL);
+       if (!ring->base)
+               return -ENOMEM;
+
+       ring->write = ring->base;
+       ring->base_end = ring->base + sizeof(struct eip93_descriptor) * (EIP93_RING_NUM - 1);
+       ring->read  = ring->base;
+
+       return 0;
+}
+
+static int eip93_desc_init(struct eip93_device *eip93)
+{
+       struct eip93_desc_ring *cdr = &eip93->ring->cdr;
+       struct eip93_desc_ring *rdr = &eip93->ring->rdr;
+       int ret;
+       u32 val;
+
+       ret = eip93_set_ring(eip93, cdr);
+       if (ret)
+               return ret;
+
+       ret = eip93_set_ring(eip93, rdr);
+       if (ret)
+               return ret;
+
+       writel((u32 __force)cdr->base_dma, eip93->base + EIP93_REG_PE_CDR_BASE);
+       writel((u32 __force)rdr->base_dma, eip93->base + EIP93_REG_PE_RDR_BASE);
+
+       val = FIELD_PREP(EIP93_PE_RING_SIZE, EIP93_RING_NUM - 1);
+       writel(val, eip93->base + EIP93_REG_PE_RING_CONFIG);
+
+       return 0;
+}
+
+static void eip93_cleanup(struct eip93_device *eip93)
+{
+       tasklet_kill(&eip93->ring->done_task);
+
+       /* Clear/ack all interrupts before disable all */
+       eip93_irq_clear(eip93, EIP93_INT_ALL);
+       eip93_irq_disable(eip93, EIP93_INT_ALL);
+
+       writel(0, eip93->base + EIP93_REG_PE_CLOCK_CTRL);
+
+       eip93_desc_free(eip93);
+
+       idr_destroy(&eip93->ring->crypto_async_idr);
+}
+
+static int eip93_crypto_probe(struct platform_device *pdev)
+{
+       struct device *dev = &pdev->dev;
+       struct eip93_device *eip93;
+       u32 ver, algo_flags;
+       int ret;
+
+       eip93 = devm_kzalloc(dev, sizeof(*eip93), GFP_KERNEL);
+       if (!eip93)
+               return -ENOMEM;
+
+       eip93->dev = dev;
+       platform_set_drvdata(pdev, eip93);
+
+       eip93->base = devm_platform_ioremap_resource(pdev, 0);
+       if (IS_ERR(eip93->base))
+               return PTR_ERR(eip93->base);
+
+       eip93->irq = platform_get_irq(pdev, 0);
+       if (eip93->irq < 0)
+               return eip93->irq;
+
+       ret = devm_request_threaded_irq(eip93->dev, eip93->irq, eip93_irq_handler,
+                                       NULL, IRQF_ONESHOT,
+                                       dev_name(eip93->dev), eip93);
+
+       eip93->ring = devm_kcalloc(eip93->dev, 1, sizeof(*eip93->ring), GFP_KERNEL);
+       if (!eip93->ring)
+               return -ENOMEM;
+
+       ret = eip93_desc_init(eip93);
+
+       if (ret)
+               return ret;
+
+       tasklet_init(&eip93->ring->done_task, eip93_done_task, (unsigned long)eip93);
+
+       spin_lock_init(&eip93->ring->read_lock);
+       spin_lock_init(&eip93->ring->write_lock);
+
+       spin_lock_init(&eip93->ring->idr_lock);
+       idr_init(&eip93->ring->crypto_async_idr);
+
+       algo_flags = readl(eip93->base + EIP93_REG_PE_OPTION_1);
+
+       eip93_initialize(eip93, algo_flags);
+
+       /* Init finished, enable RDR interrupt */
+       eip93_irq_enable(eip93, EIP93_INT_RDR_THRESH);
+
+       ret = eip93_register_algs(eip93, algo_flags);
+       if (ret) {
+               eip93_cleanup(eip93);
+               return ret;
+       }
+
+       ver = readl(eip93->base + EIP93_REG_PE_REVISION);
+       /* EIP_EIP_NO:MAJOR_HW_REV:MINOR_HW_REV:HW_PATCH,PE(ALGO_FLAGS) */
+       dev_info(eip93->dev, "EIP%lu:%lx:%lx:%lx,PE(0x%x:0x%x)\n",
+                FIELD_GET(EIP93_PE_REVISION_EIP_NO, ver),
+                FIELD_GET(EIP93_PE_REVISION_MAJ_HW_REV, ver),
+                FIELD_GET(EIP93_PE_REVISION_MIN_HW_REV, ver),
+                FIELD_GET(EIP93_PE_REVISION_HW_PATCH, ver),
+                algo_flags,
+                readl(eip93->base + EIP93_REG_PE_OPTION_0));
+
+       return 0;
+}
+
+static void eip93_crypto_remove(struct platform_device *pdev)
+{
+       struct eip93_device *eip93 = platform_get_drvdata(pdev);
+
+       eip93_unregister_algs(ARRAY_SIZE(eip93_algs));
+       eip93_cleanup(eip93);
+}
+
+static const struct of_device_id eip93_crypto_of_match[] = {
+       { .compatible = "inside-secure,safexcel-eip93i", },
+       { .compatible = "inside-secure,safexcel-eip93ie", },
+       { .compatible = "inside-secure,safexcel-eip93is", },
+       { .compatible = "inside-secure,safexcel-eip93ies", },
+       /* IW not supported currently, missing AES-XCB-MAC/AES-CCM */
+       /* { .compatible = "inside-secure,safexcel-eip93iw", }, */
+       {}
+};
+MODULE_DEVICE_TABLE(of, eip93_crypto_of_match);
+
+static struct platform_driver eip93_crypto_driver = {
+       .probe = eip93_crypto_probe,
+       .remove = eip93_crypto_remove,
+       .driver = {
+               .name = "inside-secure-eip93",
+               .of_match_table = eip93_crypto_of_match,
+       },
+};
+module_platform_driver(eip93_crypto_driver);
+
+MODULE_AUTHOR("Richard van Schagen <vschagen@cs.com>");
+MODULE_AUTHOR("Christian Marangi <ansuelsmth@gmail.com>");
+MODULE_DESCRIPTION("Mediatek EIP-93 crypto engine driver");
+MODULE_LICENSE("GPL");
diff --git a/drivers/crypto/inside-secure/eip93/eip93-main.h b/drivers/crypto/inside-secure/eip93/eip93-main.h
new file mode 100644 (file)
index 0000000..79b078f
--- /dev/null
@@ -0,0 +1,151 @@
+/* SPDX-License-Identifier: GPL-2.0
+ *
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef _EIP93_MAIN_H_
+#define _EIP93_MAIN_H_
+
+#include <crypto/internal/aead.h>
+#include <crypto/internal/hash.h>
+#include <crypto/internal/skcipher.h>
+#include <linux/bitfield.h>
+#include <linux/interrupt.h>
+
+#define EIP93_RING_BUSY_DELAY          500
+
+#define EIP93_RING_NUM                 512
+#define EIP93_RING_BUSY                        32
+#define EIP93_CRA_PRIORITY             1500
+
+#define EIP93_RING_SA_STATE_ADDR(base, idx)    ((base) + (idx))
+#define EIP93_RING_SA_STATE_DMA(dma_base, idx) ((u32 __force)(dma_base) + \
+                                                ((idx) * sizeof(struct sa_state)))
+
+/* cipher algorithms */
+#define EIP93_ALG_DES                  BIT(0)
+#define EIP93_ALG_3DES                 BIT(1)
+#define EIP93_ALG_AES                  BIT(2)
+#define EIP93_ALG_MASK                 GENMASK(2, 0)
+/* hash and hmac algorithms */
+#define EIP93_HASH_MD5                 BIT(3)
+#define EIP93_HASH_SHA1                        BIT(4)
+#define EIP93_HASH_SHA224                      BIT(5)
+#define EIP93_HASH_SHA256                      BIT(6)
+#define EIP93_HASH_HMAC                        BIT(7)
+#define EIP93_HASH_MASK                        GENMASK(6, 3)
+/* cipher modes */
+#define EIP93_MODE_CBC                 BIT(8)
+#define EIP93_MODE_ECB                 BIT(9)
+#define EIP93_MODE_CTR                 BIT(10)
+#define EIP93_MODE_RFC3686             BIT(11)
+#define EIP93_MODE_MASK                        GENMASK(10, 8)
+
+/* cipher encryption/decryption operations */
+#define EIP93_ENCRYPT                  BIT(12)
+#define EIP93_DECRYPT                  BIT(13)
+
+#define EIP93_BUSY                     BIT(14)
+
+/* descriptor flags */
+#define EIP93_DESC_DMA_IV                      BIT(0)
+#define EIP93_DESC_IPSEC                       BIT(1)
+#define EIP93_DESC_FINISH                      BIT(2)
+#define EIP93_DESC_LAST                        BIT(3)
+#define EIP93_DESC_FAKE_HMAC           BIT(4)
+#define EIP93_DESC_PRNG                        BIT(5)
+#define EIP93_DESC_HASH                        BIT(6)
+#define EIP93_DESC_AEAD                        BIT(7)
+#define EIP93_DESC_SKCIPHER            BIT(8)
+#define EIP93_DESC_ASYNC                       BIT(9)
+
+#define IS_DMA_IV(desc_flags)          ((desc_flags) & EIP93_DESC_DMA_IV)
+
+#define IS_DES(flags)                  ((flags) & EIP93_ALG_DES)
+#define IS_3DES(flags)                 ((flags) & EIP93_ALG_3DES)
+#define IS_AES(flags)                  ((flags) & EIP93_ALG_AES)
+
+#define IS_HASH_MD5(flags)             ((flags) & EIP93_HASH_MD5)
+#define IS_HASH_SHA1(flags)            ((flags) & EIP93_HASH_SHA1)
+#define IS_HASH_SHA224(flags)          ((flags) & EIP93_HASH_SHA224)
+#define IS_HASH_SHA256(flags)          ((flags) & EIP93_HASH_SHA256)
+#define IS_HMAC(flags)                 ((flags) & EIP93_HASH_HMAC)
+
+#define IS_CBC(mode)                   ((mode) & EIP93_MODE_CBC)
+#define IS_ECB(mode)                   ((mode) & EIP93_MODE_ECB)
+#define IS_CTR(mode)                   ((mode) & EIP93_MODE_CTR)
+#define IS_RFC3686(mode)               ((mode) & EIP93_MODE_RFC3686)
+
+#define IS_BUSY(flags)                 ((flags) & EIP93_BUSY)
+
+#define IS_ENCRYPT(dir)                        ((dir) & EIP93_ENCRYPT)
+#define IS_DECRYPT(dir)                        ((dir) & EIP93_DECRYPT)
+
+#define IS_CIPHER(flags)               ((flags) & (EIP93_ALG_DES | \
+                                                   EIP93_ALG_3DES |  \
+                                                   EIP93_ALG_AES))
+
+#define IS_HASH(flags)                 ((flags) & (EIP93_HASH_MD5 |  \
+                                                   EIP93_HASH_SHA1 |   \
+                                                   EIP93_HASH_SHA224 | \
+                                                   EIP93_HASH_SHA256))
+
+/**
+ * struct eip93_device - crypto engine device structure
+ */
+struct eip93_device {
+       void __iomem            *base;
+       struct device           *dev;
+       struct clk              *clk;
+       int                     irq;
+       struct eip93_ring               *ring;
+};
+
+struct eip93_desc_ring {
+       void                    *base;
+       void                    *base_end;
+       dma_addr_t              base_dma;
+       /* write and read pointers */
+       void                    *read;
+       void                    *write;
+       /* descriptor element offset */
+       u32                     offset;
+};
+
+struct eip93_state_pool {
+       void                    *base;
+       dma_addr_t              base_dma;
+};
+
+struct eip93_ring {
+       struct tasklet_struct           done_task;
+       /* command/result rings */
+       struct eip93_desc_ring          cdr;
+       struct eip93_desc_ring          rdr;
+       spinlock_t                      write_lock;
+       spinlock_t                      read_lock;
+       /* aync idr */
+       spinlock_t                      idr_lock;
+       struct idr                      crypto_async_idr;
+};
+
+enum eip93_alg_type {
+       EIP93_ALG_TYPE_AEAD,
+       EIP93_ALG_TYPE_SKCIPHER,
+       EIP93_ALG_TYPE_HASH,
+};
+
+struct eip93_alg_template {
+       struct eip93_device     *eip93;
+       enum eip93_alg_type     type;
+       u32                     flags;
+       union {
+               struct aead_alg         aead;
+               struct skcipher_alg     skcipher;
+               struct ahash_alg        ahash;
+       } alg;
+};
+
+#endif /* _EIP93_MAIN_H_ */
diff --git a/drivers/crypto/inside-secure/eip93/eip93-regs.h b/drivers/crypto/inside-secure/eip93/eip93-regs.h
new file mode 100644 (file)
index 0000000..0490b8d
--- /dev/null
@@ -0,0 +1,335 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/*
+ * Copyright (C) 2019 - 2021
+ *
+ * Richard van Schagen <vschagen@icloud.com>
+ * Christian Marangi <ansuelsmth@gmail.com
+ */
+#ifndef REG_EIP93_H
+#define REG_EIP93_H
+
+#define EIP93_REG_PE_CTRL_STAT                 0x0
+#define   EIP93_PE_CTRL_PE_PAD_CTRL_STAT       GENMASK(31, 24)
+#define   EIP93_PE_CTRL_PE_EXT_ERR_CODE                GENMASK(23, 20)
+#define   EIP93_PE_CTRL_PE_EXT_ERR_PROCESSING  0x8
+#define   EIP93_PE_CTRL_PE_EXT_ERR_BLOCK_SIZE_ERR 0x7
+#define   EIP93_PE_CTRL_PE_EXT_ERR_INVALID_PK_LENGTH 0x6
+#define   EIP93_PE_CTRL_PE_EXT_ERR_ZERO_LENGTH 0x5
+#define   EIP93_PE_CTRL_PE_EXT_ERR_SPI         0x4
+#define   EIP93_PE_CTRL_PE_EXT_ERR_INVALID_CRYPTO_ALGO 0x3
+#define   EIP93_PE_CTRL_PE_EXT_ERR_INVALID_CRYPTO_OP 0x2
+#define   EIP93_PE_CTRL_PE_EXT_ERR_DESC_OWNER  0x1
+#define   EIP93_PE_CTRL_PE_EXT_ERR_BUS         0x0
+#define   EIP93_PE_CTRL_PE_EXT_ERR             BIT(19)
+#define   EIP93_PE_CTRL_PE_SEQNUM_ERR          BIT(18)
+#define   EIP93_PE_CTRL_PE_PAD_ERR             BIT(17)
+#define   EIP93_PE_CTRL_PE_AUTH_ERR            BIT(16)
+#define   EIP93_PE_CTRL_PE_PAD_VALUE           GENMASK(15, 8)
+#define   EIP93_PE_CTRL_PE_PRNG_MODE           GENMASK(7, 6)
+#define   EIP93_PE_CTRL_PE_HASH_FINAL          BIT(4)
+#define   EIP93_PE_CTRL_PE_INIT_ARC4           BIT(3)
+#define   EIP93_PE_CTRL_PE_READY_DES_TRING_OWN GENMASK(1, 0)
+#define   EIP93_PE_CTRL_PE_READY               0x2
+#define   EIP93_PE_CTRL_HOST_READY             0x1
+#define EIP93_REG_PE_SOURCE_ADDR               0x4
+#define EIP93_REG_PE_DEST_ADDR                 0x8
+#define EIP93_REG_PE_SA_ADDR                   0xc
+#define EIP93_REG_PE_ADDR                      0x10 /* STATE_ADDR */
+/*
+ * Special implementation for user ID
+ * user_id in eip93_descriptor is used to identify the
+ * descriptor and is opaque and can be used by the driver
+ * in custom way.
+ *
+ * The usage of this should be to put an address to the crypto
+ * request struct from the kernel but this can't work in 64bit
+ * world.
+ *
+ * Also it's required to put some flags to identify the last
+ * descriptor.
+ *
+ * To handle this, split the u32 in 2 part:
+ * - 31:16 descriptor flags
+ * - 15:0 IDR to connect the crypto request address
+ */
+#define EIP93_REG_PE_USER_ID                   0x18
+#define   EIP93_PE_USER_ID_DESC_FLAGS          GENMASK(31, 16)
+#define   EIP93_PE_USER_ID_CRYPTO_IDR          GENMASK(15, 0)
+#define EIP93_REG_PE_LENGTH                    0x1c
+#define   EIP93_PE_LENGTH_BYPASS               GENMASK(31, 24)
+#define   EIP93_PE_LENGTH_HOST_PE_READY                GENMASK(23, 22)
+#define   EIP93_PE_LENGTH_PE_READY             0x2
+#define   EIP93_PE_LENGTH_HOST_READY           0x1
+#define   EIP93_PE_LENGTH_LENGTH               GENMASK(19, 0)
+
+/* PACKET ENGINE RING configuration registers */
+#define EIP93_REG_PE_CDR_BASE                  0x80
+#define EIP93_REG_PE_RDR_BASE                  0x84
+#define EIP93_REG_PE_RING_CONFIG               0x88
+#define   EIP93_PE_EN_EXT_TRIG                 BIT(31)
+/* Absent in later revision of eip93 */
+/* #define   EIP93_PE_RING_OFFSET              GENMASK(23, 15) */
+#define   EIP93_PE_RING_SIZE                   GENMASK(9, 0)
+#define EIP93_REG_PE_RING_THRESH               0x8c
+#define   EIPR93_PE_TIMEROUT_EN                        BIT(31)
+#define   EIPR93_PE_RD_TIMEOUT                 GENMASK(29, 26)
+#define   EIPR93_PE_RDR_THRESH                 GENMASK(25, 16)
+#define   EIPR93_PE_CDR_THRESH                 GENMASK(9, 0)
+#define EIP93_REG_PE_CD_COUNT                  0x90
+#define   EIP93_PE_CD_COUNT                    GENMASK(10, 0)
+/*
+ * In the same register, writing a value in GENMASK(7, 0) will
+ * increment the descriptor count and start DMA action.
+ */
+#define   EIP93_PE_CD_COUNT_INCR               GENMASK(7, 0)
+#define EIP93_REG_PE_RD_COUNT                  0x94
+#define   EIP93_PE_RD_COUNT                    GENMASK(10, 0)
+/*
+ * In the same register, writing a value in GENMASK(7, 0) will
+ * increment the descriptor count and start DMA action.
+ */
+#define   EIP93_PE_RD_COUNT_INCR               GENMASK(7, 0)
+#define EIP93_REG_PE_RING_RW_PNTR              0x98 /* RING_PNTR */
+
+/* PACKET ENGINE  configuration registers */
+#define EIP93_REG_PE_CONFIG                    0x100
+#define   EIP93_PE_CONFIG_SWAP_TARGET          BIT(20)
+#define   EIP93_PE_CONFIG_SWAP_DATA            BIT(18)
+#define   EIP93_PE_CONFIG_SWAP_SA              BIT(17)
+#define   EIP93_PE_CONFIG_SWAP_CDRD            BIT(16)
+#define   EIP93_PE_CONFIG_EN_CDR_UPDATE                BIT(10)
+#define   EIP93_PE_CONFIG_PE_MODE              GENMASK(9, 8)
+#define   EIP93_PE_TARGET_AUTO_RING_MODE       FIELD_PREP(EIP93_PE_CONFIG_PE_MODE, 0x3)
+#define   EIP93_PE_TARGET_COMMAND_NO_RDR_MODE  FIELD_PREP(EIP93_PE_CONFIG_PE_MODE, 0x2)
+#define   EIP93_PE_TARGET_COMMAND_WITH_RDR_MODE        FIELD_PREP(EIP93_PE_CONFIG_PE_MODE, 0x1)
+#define   EIP93_PE_DIRECT_HOST_MODE            FIELD_PREP(EIP93_PE_CONFIG_PE_MODE, 0x0)
+#define   EIP93_PE_CONFIG_RST_RING             BIT(2)
+#define   EIP93_PE_CONFIG_RST_PE               BIT(0)
+#define EIP93_REG_PE_STATUS                    0x104
+#define EIP93_REG_PE_BUF_THRESH                        0x10c
+#define   EIP93_PE_OUTBUF_THRESH               GENMASK(23, 16)
+#define   EIP93_PE_INBUF_THRESH                        GENMASK(7, 0)
+#define EIP93_REG_PE_INBUF_COUNT               0x100
+#define EIP93_REG_PE_OUTBUF_COUNT              0x114
+#define EIP93_REG_PE_BUF_RW_PNTR               0x118 /* BUF_PNTR */
+
+/* PACKET ENGINE endian config */
+#define EIP93_REG_PE_ENDIAN_CONFIG             0x1cc
+#define EIP93_AIROHA_REG_PE_ENDIAN_CONFIG      0x1d0
+#define   EIP93_PE_ENDIAN_TARGET_BYTE_SWAP     GENMASK(23, 16)
+#define   EIP93_PE_ENDIAN_MASTER_BYTE_SWAP     GENMASK(7, 0)
+/*
+ * Byte goes 2 and 2 and are referenced by ID
+ * Split GENMASK(7, 0) in 4 part, one for each byte.
+ * Example LITTLE ENDIAN: Example BIG ENDIAN
+ * GENMASK(7, 6) 0x3     GENMASK(7, 6) 0x0
+ * GENMASK(5, 4) 0x2     GENMASK(7, 6) 0x1
+ * GENMASK(3, 2) 0x1     GENMASK(3, 2) 0x2
+ * GENMASK(1, 0) 0x0     GENMASK(1, 0) 0x3
+ */
+#define   EIP93_PE_ENDIAN_BYTE0                        0x0
+#define   EIP93_PE_ENDIAN_BYTE1                        0x1
+#define   EIP93_PE_ENDIAN_BYTE2                        0x2
+#define   EIP93_PE_ENDIAN_BYTE3                        0x3
+
+/* EIP93 CLOCK control registers */
+#define EIP93_REG_PE_CLOCK_CTRL                        0x1e8
+#define   EIP93_PE_CLOCK_EN_HASH_CLK           BIT(4)
+#define   EIP93_PE_CLOCK_EN_ARC4_CLK           BIT(3)
+#define   EIP93_PE_CLOCK_EN_AES_CLK            BIT(2)
+#define   EIP93_PE_CLOCK_EN_DES_CLK            BIT(1)
+#define   EIP93_PE_CLOCK_EN_PE_CLK             BIT(0)
+
+/* EIP93 Device Option and Revision Register */
+#define EIP93_REG_PE_OPTION_1                  0x1f4
+#define   EIP93_PE_OPTION_MAC_KEY256           BIT(31)
+#define   EIP93_PE_OPTION_MAC_KEY192           BIT(30)
+#define   EIP93_PE_OPTION_MAC_KEY128           BIT(29)
+#define   EIP93_PE_OPTION_AES_CBC_MAC          BIT(28)
+#define   EIP93_PE_OPTION_AES_XCBX             BIT(23)
+#define   EIP93_PE_OPTION_SHA_256              BIT(19)
+#define   EIP93_PE_OPTION_SHA_224              BIT(18)
+#define   EIP93_PE_OPTION_SHA_1                        BIT(17)
+#define   EIP93_PE_OPTION_MD5                  BIT(16)
+#define   EIP93_PE_OPTION_AES_KEY256           BIT(15)
+#define   EIP93_PE_OPTION_AES_KEY192           BIT(14)
+#define   EIP93_PE_OPTION_AES_KEY128           BIT(13)
+#define   EIP93_PE_OPTION_AES                  BIT(2)
+#define   EIP93_PE_OPTION_ARC4                 BIT(1)
+#define   EIP93_PE_OPTION_TDES                 BIT(0) /* DES and TDES */
+#define EIP93_REG_PE_OPTION_0                  0x1f8
+#define EIP93_REG_PE_REVISION                  0x1fc
+#define   EIP93_PE_REVISION_MAJ_HW_REV         GENMASK(27, 24)
+#define   EIP93_PE_REVISION_MIN_HW_REV         GENMASK(23, 20)
+#define   EIP93_PE_REVISION_HW_PATCH           GENMASK(19, 16)
+#define   EIP93_PE_REVISION_EIP_NO             GENMASK(7, 0)
+
+/* EIP93 Interrupt Control Register */
+#define EIP93_REG_INT_UNMASK_STAT              0x200
+#define EIP93_REG_INT_MASK_STAT                        0x204
+#define EIP93_REG_INT_CLR                      0x204
+#define EIP93_REG_INT_MASK                     0x208 /* INT_EN */
+/* Each int reg have the same bitmap */
+#define   EIP93_INT_INTERFACE_ERR              BIT(18)
+#define   EIP93_INT_RPOC_ERR                   BIT(17)
+#define   EIP93_INT_PE_RING_ERR                        BIT(16)
+#define   EIP93_INT_HALT                       BIT(15)
+#define   EIP93_INT_OUTBUF_THRESH              BIT(11)
+#define   EIP93_INT_INBUF_THRESH               BIT(10)
+#define   EIP93_INT_OPERATION_DONE             BIT(9)
+#define   EIP93_INT_RDR_THRESH                 BIT(1)
+#define   EIP93_INT_CDR_THRESH                 BIT(0)
+#define   EIP93_INT_ALL                                (EIP93_INT_INTERFACE_ERR | \
+                                                EIP93_INT_RPOC_ERR | \
+                                                EIP93_INT_PE_RING_ERR | \
+                                                EIP93_INT_HALT | \
+                                                EIP93_INT_OUTBUF_THRESH | \
+                                                EIP93_INT_INBUF_THRESH | \
+                                                EIP93_INT_OPERATION_DONE | \
+                                                EIP93_INT_RDR_THRESH | \
+                                                EIP93_INT_CDR_THRESH)
+
+#define EIP93_REG_INT_CFG                      0x20c
+#define   EIP93_INT_TYPE_PULSE                 BIT(0)
+#define EIP93_REG_MASK_ENABLE                  0x210
+#define EIP93_REG_MASK_DISABLE                 0x214
+
+/* EIP93 SA Record register */
+#define EIP93_REG_SA_CMD_0                     0x400
+#define   EIP93_SA_CMD_SAVE_HASH               BIT(29)
+#define   EIP93_SA_CMD_SAVE_IV                 BIT(28)
+#define   EIP93_SA_CMD_HASH_SOURCE             GENMASK(27, 26)
+#define   EIP93_SA_CMD_HASH_NO_LOAD            FIELD_PREP(EIP93_SA_CMD_HASH_SOURCE, 0x3)
+#define   EIP93_SA_CMD_HASH_FROM_STATE         FIELD_PREP(EIP93_SA_CMD_HASH_SOURCE, 0x2)
+#define   EIP93_SA_CMD_HASH_FROM_SA            FIELD_PREP(EIP93_SA_CMD_HASH_SOURCE, 0x0)
+#define   EIP93_SA_CMD_IV_SOURCE               GENMASK(25, 24)
+#define   EIP93_SA_CMD_IV_FROM_PRNG            FIELD_PREP(EIP93_SA_CMD_IV_SOURCE, 0x3)
+#define   EIP93_SA_CMD_IV_FROM_STATE           FIELD_PREP(EIP93_SA_CMD_IV_SOURCE, 0x2)
+#define   EIP93_SA_CMD_IV_FROM_INPUT           FIELD_PREP(EIP93_SA_CMD_IV_SOURCE, 0x1)
+#define   EIP93_SA_CMD_IV_NO_LOAD              FIELD_PREP(EIP93_SA_CMD_IV_SOURCE, 0x0)
+#define   EIP93_SA_CMD_DIGEST_LENGTH           GENMASK(23, 20)
+#define   EIP93_SA_CMD_DIGEST_10WORD           FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0xa) /* SRTP and TLS */
+#define   EIP93_SA_CMD_DIGEST_8WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x8) /* SHA-256 */
+#define   EIP93_SA_CMD_DIGEST_7WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x7) /* SHA-224 */
+#define   EIP93_SA_CMD_DIGEST_6WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x6)
+#define   EIP93_SA_CMD_DIGEST_5WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x5) /* SHA1 */
+#define   EIP93_SA_CMD_DIGEST_4WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x4) /* MD5 and AES-based */
+#define   EIP93_SA_CMD_DIGEST_3WORD_IPSEC      FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x3) /* IPSEC */
+#define   EIP93_SA_CMD_DIGEST_2WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x2)
+#define   EIP93_SA_CMD_DIGEST_1WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x1)
+#define   EIP93_SA_CMD_DIGEST_3WORD            FIELD_PREP(EIP93_SA_CMD_DIGEST_LENGTH, 0x0) /* 96bit output */
+#define   EIP93_SA_CMD_HDR_PROC                        BIT(19)
+#define   EIP93_SA_CMD_EXT_PAD                 BIT(18)
+#define   EIP93_SA_CMD_SCPAD                   BIT(17)
+#define   EIP93_SA_CMD_HASH                    GENMASK(15, 12)
+#define   EIP93_SA_CMD_HASH_NULL               FIELD_PREP(EIP93_SA_CMD_HASH, 0xf)
+#define   EIP93_SA_CMD_HASH_SHA256             FIELD_PREP(EIP93_SA_CMD_HASH, 0x3)
+#define   EIP93_SA_CMD_HASH_SHA224             FIELD_PREP(EIP93_SA_CMD_HASH, 0x2)
+#define   EIP93_SA_CMD_HASH_SHA1               FIELD_PREP(EIP93_SA_CMD_HASH, 0x1)
+#define   EIP93_SA_CMD_HASH_MD5                        FIELD_PREP(EIP93_SA_CMD_HASH, 0x0)
+#define   EIP93_SA_CMD_CIPHER                  GENMASK(11, 8)
+#define   EIP93_SA_CMD_CIPHER_NULL             FIELD_PREP(EIP93_SA_CMD_CIPHER, 0xf)
+#define   EIP93_SA_CMD_CIPHER_AES              FIELD_PREP(EIP93_SA_CMD_CIPHER, 0x3)
+#define   EIP93_SA_CMD_CIPHER_ARC4             FIELD_PREP(EIP93_SA_CMD_CIPHER, 0x2)
+#define   EIP93_SA_CMD_CIPHER_3DES             FIELD_PREP(EIP93_SA_CMD_CIPHER, 0x1)
+#define   EIP93_SA_CMD_CIPHER_DES              FIELD_PREP(EIP93_SA_CMD_CIPHER, 0x0)
+#define   EIP93_SA_CMD_PAD_TYPE                        GENMASK(7, 6)
+#define   EIP93_SA_CMD_PAD_CONST_SSL           FIELD_PREP(EIP93_SA_CMD_PAD_TYPE, 0x6)
+#define   EIP93_SA_CMD_PAD_TLS_DTLS            FIELD_PREP(EIP93_SA_CMD_PAD_TYPE, 0x5)
+#define   EIP93_SA_CMD_PAD_ZERO                        FIELD_PREP(EIP93_SA_CMD_PAD_TYPE, 0x3)
+#define   EIP93_SA_CMD_PAD_CONST               FIELD_PREP(EIP93_SA_CMD_PAD_TYPE, 0x2)
+#define   EIP93_SA_CMD_PAD_PKCS7               FIELD_PREP(EIP93_SA_CMD_PAD_TYPE, 0x1)
+#define   EIP93_SA_CMD_PAD_IPSEC               FIELD_PREP(EIP93_SA_CMD_PAD_TYPE, 0x0)
+#define   EIP93_SA_CMD_OPGROUP                 GENMASK(5, 4)
+#define   EIP93_SA_CMD_OP_EXT                  FIELD_PREP(EIP93_SA_CMD_OPGROUP, 0x2)
+#define   EIP93_SA_CMD_OP_PROTOCOL             FIELD_PREP(EIP93_SA_CMD_OPGROUP, 0x1)
+#define   EIP93_SA_CMD_OP_BASIC                        FIELD_PREP(EIP93_SA_CMD_OPGROUP, 0x0)
+#define   EIP93_SA_CMD_DIRECTION_IN            BIT(3) /* 0: outbount 1: inbound */
+#define   EIP93_SA_CMD_OPCODE                  GENMASK(2, 0)
+#define   EIP93_SA_CMD_OPCODE_BASIC_OUT_PRNG   0x7
+#define   EIP93_SA_CMD_OPCODE_BASIC_OUT_HASH   0x3
+#define   EIP93_SA_CMD_OPCODE_BASIC_OUT_ENC_HASH 0x1
+#define   EIP93_SA_CMD_OPCODE_BASIC_OUT_ENC    0x0
+#define   EIP93_SA_CMD_OPCODE_BASIC_IN_HASH    0x3
+#define   EIP93_SA_CMD_OPCODE_BASIC_IN_HASH_DEC        0x1
+#define   EIP93_SA_CMD_OPCODE_BASIC_IN_DEC     0x0
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_OUT_ESP 0x0
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_OUT_SSL 0x4
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_OUT_TLS 0x5
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_OUT_SRTP        0x7
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_IN_ESP  0x0
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_IN_SSL  0x2
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_IN_TLS  0x3
+#define   EIP93_SA_CMD_OPCODE_PROTOCOL_IN_SRTP 0x7
+#define   EIP93_SA_CMD_OPCODE_EXT_OUT_DTSL     0x1
+#define   EIP93_SA_CMD_OPCODE_EXT_OUT_SSL      0x4
+#define   EIP93_SA_CMD_OPCODE_EXT_OUT_TLSV10   0x5
+#define   EIP93_SA_CMD_OPCODE_EXT_OUT_TLSV11   0x6
+#define   EIP93_SA_CMD_OPCODE_EXT_IN_DTSL      0x1
+#define   EIP93_SA_CMD_OPCODE_EXT_IN_SSL       0x4
+#define   EIP93_SA_CMD_OPCODE_EXT_IN_TLSV10    0x5
+#define   EIP93_SA_CMD_OPCODE_EXT_IN_TLSV11    0x6
+#define EIP93_REG_SA_CMD_1                     0x404
+#define   EIP93_SA_CMD_EN_SEQNUM_CHK           BIT(29)
+/* This mask can be either used for ARC4 or AES */
+#define   EIP93_SA_CMD_ARC4_KEY_LENGHT         GENMASK(28, 24)
+#define   EIP93_SA_CMD_AES_DEC_KEY             BIT(28) /* 0: encrypt key 1: decrypt key */
+#define   EIP93_SA_CMD_AES_KEY_LENGTH          GENMASK(26, 24)
+#define   EIP93_SA_CMD_AES_KEY_256BIT          FIELD_PREP(EIP93_SA_CMD_AES_KEY_LENGTH, 0x4)
+#define   EIP93_SA_CMD_AES_KEY_192BIT          FIELD_PREP(EIP93_SA_CMD_AES_KEY_LENGTH, 0x3)
+#define   EIP93_SA_CMD_AES_KEY_128BIT          FIELD_PREP(EIP93_SA_CMD_AES_KEY_LENGTH, 0x2)
+#define   EIP93_SA_CMD_HASH_CRYPT_OFFSET       GENMASK(23, 16)
+#define   EIP93_SA_CMD_BYTE_OFFSET             BIT(13) /* 0: CRYPT_OFFSET in 32bit word 1: CRYPT_OFFSET in 8bit bytes */
+#define   EIP93_SA_CMD_HMAC                    BIT(12)
+#define   EIP93_SA_CMD_SSL_MAC                 BIT(12)
+/* This mask can be either used for ARC4 or AES */
+#define   EIP93_SA_CMD_CHIPER_MODE             GENMASK(9, 8)
+/* AES or DES operations */
+#define   EIP93_SA_CMD_CHIPER_MODE_ICM         FIELD_PREP(EIP93_SA_CMD_CHIPER_MODE, 0x3)
+#define   EIP93_SA_CMD_CHIPER_MODE_CTR         FIELD_PREP(EIP93_SA_CMD_CHIPER_MODE, 0x2)
+#define   EIP93_SA_CMD_CHIPER_MODE_CBC         FIELD_PREP(EIP93_SA_CMD_CHIPER_MODE, 0x1)
+#define   EIP93_SA_CMD_CHIPER_MODE_ECB         FIELD_PREP(EIP93_SA_CMD_CHIPER_MODE, 0x0)
+/* ARC4 operations */
+#define   EIP93_SA_CMD_CHIPER_MODE_STATEFULL   FIELD_PREP(EIP93_SA_CMD_CHIPER_MODE, 0x1)
+#define   EIP93_SA_CMD_CHIPER_MODE_STATELESS   FIELD_PREP(EIP93_SA_CMD_CHIPER_MODE, 0x0)
+#define   EIP93_SA_CMD_COPY_PAD                        BIT(3)
+#define   EIP93_SA_CMD_COPY_PAYLOAD            BIT(2)
+#define   EIP93_SA_CMD_COPY_HEADER             BIT(1)
+#define   EIP93_SA_CMD_COPY_DIGEST             BIT(0) /* With this enabled, COPY_PAD is required */
+
+/* State save register */
+#define EIP93_REG_STATE_IV_0                   0x500
+#define EIP93_REG_STATE_IV_1                   0x504
+
+#define EIP93_REG_PE_ARC4STATE                 0x700
+
+struct sa_record {
+       u32 sa_cmd0_word;
+       u32 sa_cmd1_word;
+       u32 sa_key[8];
+       u8 sa_i_digest[32];
+       u8 sa_o_digest[32];
+       u32 sa_spi;
+       u32 sa_seqnum[2];
+       u32 sa_seqmum_mask[2];
+       u32 sa_nonce;
+} __packed;
+
+struct sa_state {
+       u32 state_iv[4];
+       u32 state_byte_cnt[2];
+       u8 state_i_digest[32];
+} __packed;
+
+struct eip93_descriptor {
+       u32 pe_ctrl_stat_word;
+       u32 src_addr;
+       u32 dst_addr;
+       u32 sa_addr;
+       u32 state_addr;
+       u32 arc4_addr;
+       u32 user_id;
+       u32 pe_length_word;
+} __packed;
+
+#endif