--- /dev/null
+/*
+ * Copyright (C) 2017 Xilinx, Inc.
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2 of the License, or
+ * (at your option) any later version.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ * GNU General Public License for more details.
+ */
+
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/crypto.h>
+#include <linux/spinlock.h>
+#include <crypto/algapi.h>
+#include <crypto/aes.h>
+#include <linux/io.h>
+#include <linux/device.h>
+#include <linux/of_device.h>
+#include <linux/platform_device.h>
+#include <linux/scatterlist.h>
+#include <crypto/scatterwalk.h>
+#include <linux/soc/xilinx/zynqmp/firmware.h>
+
+#define ZYNQMP_RSA_QUEUE_LENGTH 1
+#define ZYNQMP_RSA_MAX_KEY_SIZE 1024
+
+struct zynqmp_rsa_dev;
+
+struct zynqmp_rsa_op {
+ struct zynqmp_rsa_dev *dd;
+ void *src;
+ void *dst;
+ int len;
+ u8 key[ZYNQMP_RSA_MAX_KEY_SIZE];
+ u8 *iv;
+ u32 keylen;
+};
+
+struct zynqmp_rsa_dev {
+ struct list_head list;
+ struct device *dev;
+ /* the lock protects queue and dev list*/
+ spinlock_t lock;
+ struct crypto_queue queue;
+};
+
+struct zynqmp_rsa_drv {
+ struct list_head dev_list;
+ /* the lock protects queue and dev list*/
+ spinlock_t lock;
+};
+
+static struct zynqmp_rsa_drv zynqmp_rsa = {
+ .dev_list = LIST_HEAD_INIT(zynqmp_rsa.dev_list),
+ .lock = __SPIN_LOCK_UNLOCKED(zynqmp_rsa.lock),
+};
+
+static struct zynqmp_rsa_dev *zynqmp_rsa_find_dev(struct zynqmp_rsa_op *ctx)
+{
+ struct zynqmp_rsa_dev *rsa_dd = NULL;
+ struct zynqmp_rsa_dev *tmp;
+
+ spin_lock_bh(&zynqmp_rsa.lock);
+ if (!ctx->dd) {
+ list_for_each_entry(tmp, &zynqmp_rsa.dev_list, list) {
+ rsa_dd = tmp;
+ break;
+ }
+ ctx->dd = rsa_dd;
+ } else {
+ rsa_dd = ctx->dd;
+ }
+ spin_unlock_bh(&zynqmp_rsa.lock);
+
+ return rsa_dd;
+}
+
+static int zynqmp_setkey_blk(struct crypto_tfm *tfm, const u8 *key,
+ unsigned int len)
+{
+ struct zynqmp_rsa_op *op = crypto_tfm_ctx(tfm);
+
+ op->keylen = len;
+ memcpy(op->key, key, len);
+ return 0;
+}
+
+static int
+zynqmp_rsa_decrypt(struct blkcipher_desc *desc,
+ struct scatterlist *dst, struct scatterlist *src,
+ unsigned int nbytes)
+{
+ struct zynqmp_rsa_op *op = crypto_blkcipher_ctx(desc->tfm);
+ struct zynqmp_rsa_dev *dd = zynqmp_rsa_find_dev(op);
+ struct blkcipher_walk walk;
+ char *kbuf;
+ size_t dma_size;
+ dma_addr_t dma_addr;
+ int err;
+
+ blkcipher_walk_init(&walk, dst, src, nbytes);
+ err = blkcipher_walk_virt(desc, &walk);
+ op->iv = walk.iv;
+ op->src = walk.src.virt.addr;
+ op->dst = walk.dst.virt.addr;
+
+ dma_size = nbytes + op->keylen;
+ kbuf = dma_alloc_coherent(dd->dev, dma_size, &dma_addr, GFP_KERNEL);
+ if (!kbuf)
+ return -ENOMEM;
+
+ memcpy(kbuf, op->src, nbytes);
+ memcpy(kbuf + nbytes, op->key, op->keylen);
+ zynqmp_pm_rsa(dma_addr, nbytes, 0);
+ memcpy(walk.dst.virt.addr, kbuf, nbytes);
+ err = blkcipher_walk_done(desc, &walk, nbytes - 1);
+ dma_free_coherent(dd->dev, dma_size, kbuf, dma_addr);
+
+ return err;
+}
+
+static int
+zynqmp_rsa_encrypt(struct blkcipher_desc *desc,
+ struct scatterlist *dst, struct scatterlist *src,
+ unsigned int nbytes)
+{
+ struct zynqmp_rsa_op *op = crypto_blkcipher_ctx(desc->tfm);
+ struct zynqmp_rsa_dev *dd = zynqmp_rsa_find_dev(op);
+ struct blkcipher_walk walk;
+ char *kbuf;
+ size_t dma_size;
+ dma_addr_t dma_addr;
+ int err;
+
+ blkcipher_walk_init(&walk, dst, src, nbytes);
+ err = blkcipher_walk_virt(desc, &walk);
+ op->iv = walk.iv;
+ op->src = walk.src.virt.addr;
+ op->dst = walk.dst.virt.addr;
+
+ dma_size = nbytes + op->keylen;
+ kbuf = dma_alloc_coherent(dd->dev, dma_size, &dma_addr, GFP_KERNEL);
+ if (!kbuf)
+ return -ENOMEM;
+
+ memcpy(kbuf, op->src, nbytes);
+ memcpy(kbuf + nbytes, op->key, op->keylen);
+ zynqmp_pm_rsa(dma_addr, nbytes, 1);
+ memcpy(walk.dst.virt.addr, kbuf, nbytes);
+ err = blkcipher_walk_done(desc, &walk, nbytes - 1);
+ dma_free_coherent(dd->dev, dma_size, kbuf, dma_addr);
+
+ return err;
+}
+
+static struct crypto_alg zynqmp_alg = {
+ .cra_name = "xilinx-zynqmp-rsa",
+ .cra_driver_name = "zynqmp-rsa",
+ .cra_priority = 400,
+ .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER |
+ CRYPTO_ALG_KERN_DRIVER_ONLY,
+ .cra_blocksize = 1,
+ .cra_ctxsize = sizeof(struct zynqmp_rsa_op),
+ .cra_alignmask = 15,
+ .cra_type = &crypto_blkcipher_type,
+ .cra_module = THIS_MODULE,
+ .cra_u = {
+ .blkcipher = {
+ .min_keysize = 0,
+ .max_keysize = ZYNQMP_RSA_MAX_KEY_SIZE,
+ .setkey = zynqmp_setkey_blk,
+ .encrypt = zynqmp_rsa_encrypt,
+ .decrypt = zynqmp_rsa_decrypt,
+ .ivsize = 1,
+ }
+ }
+};
+
+static const struct of_device_id zynqmp_rsa_dt_ids[] = {
+ { .compatible = "xlnx,zynqmp-rsa" },
+ { /* sentinel */ }
+};
+
+MODULE_DEVICE_TABLE(of, zynqmp_rsa_dt_ids);
+
+static int zynqmp_rsa_probe(struct platform_device *pdev)
+{
+ struct zynqmp_rsa_dev *rsa_dd;
+ struct device *dev = &pdev->dev;
+ int ret;
+
+ rsa_dd = devm_kzalloc(&pdev->dev, sizeof(*rsa_dd), GFP_KERNEL);
+ if (!rsa_dd)
+ return -ENOMEM;
+
+ rsa_dd->dev = dev;
+ platform_set_drvdata(pdev, rsa_dd);
+
+ ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(44));
+ if (ret < 0)
+ dev_err(dev, "no usable DMA configuration");
+
+ INIT_LIST_HEAD(&rsa_dd->list);
+ spin_lock_init(&rsa_dd->lock);
+ crypto_init_queue(&rsa_dd->queue, ZYNQMP_RSA_QUEUE_LENGTH);
+ spin_lock(&zynqmp_rsa.lock);
+ list_add_tail(&rsa_dd->list, &zynqmp_rsa.dev_list);
+ spin_unlock(&zynqmp_rsa.lock);
+
+ ret = crypto_register_alg(&zynqmp_alg);
+ if (ret)
+ goto err_algs;
+
+ return 0;
+
+err_algs:
+ spin_lock(&zynqmp_rsa.lock);
+ list_del(&rsa_dd->list);
+ spin_unlock(&zynqmp_rsa.lock);
+ dev_err(dev, "initialization failed.\n");
+ return ret;
+}
+
+static int zynqmp_rsa_remove(struct platform_device *pdev)
+{
+ crypto_unregister_alg(&zynqmp_alg);
+ return 0;
+}
+
+static struct platform_driver xilinx_rsa_driver = {
+ .probe = zynqmp_rsa_probe,
+ .remove = zynqmp_rsa_remove,
+ .driver = {
+ .name = "zynqmp_rsa",
+ .of_match_table = of_match_ptr(zynqmp_rsa_dt_ids),
+ },
+};
+
+module_platform_driver(xilinx_rsa_driver);
+
+MODULE_DESCRIPTION("ZynqMP RSA hw acceleration support.");
+MODULE_LICENSE("GPL");
+MODULE_AUTHOR("Nava kishore Manne <navam@xilinx.com>");