1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (c) 2010-2014, The Linux Foundation. All rights reserved. 4 */ 5 6 #include <linux/clk.h> 7 #include <linux/interrupt.h> 8 #include <linux/module.h> 9 #include <linux/mod_devicetable.h> 10 #include <linux/platform_device.h> 11 #include <linux/spinlock.h> 12 #include <linux/types.h> 13 #include <crypto/algapi.h> 14 #include <crypto/internal/hash.h> 15 #include <crypto/sha.h> 16 17 #include "core.h" 18 #include "cipher.h" 19 #include "sha.h" 20 21 #define QCE_MAJOR_VERSION5 0x05 22 #define QCE_QUEUE_LENGTH 1 23 24 static const struct qce_algo_ops *qce_ops[] = { 25 #ifdef CONFIG_CRYPTO_DEV_QCE_SKCIPHER 26 &skcipher_ops, 27 #endif 28 #ifdef CONFIG_CRYPTO_DEV_QCE_SHA 29 &ahash_ops, 30 #endif 31 }; 32 33 static void qce_unregister_algs(struct qce_device *qce) 34 { 35 const struct qce_algo_ops *ops; 36 int i; 37 38 for (i = 0; i < ARRAY_SIZE(qce_ops); i++) { 39 ops = qce_ops[i]; 40 ops->unregister_algs(qce); 41 } 42 } 43 44 static int qce_register_algs(struct qce_device *qce) 45 { 46 const struct qce_algo_ops *ops; 47 int i, ret = -ENODEV; 48 49 for (i = 0; i < ARRAY_SIZE(qce_ops); i++) { 50 ops = qce_ops[i]; 51 ret = ops->register_algs(qce); 52 if (ret) 53 break; 54 } 55 56 return ret; 57 } 58 59 static int qce_handle_request(struct crypto_async_request *async_req) 60 { 61 int ret = -EINVAL, i; 62 const struct qce_algo_ops *ops; 63 u32 type = crypto_tfm_alg_type(async_req->tfm); 64 65 for (i = 0; i < ARRAY_SIZE(qce_ops); i++) { 66 ops = qce_ops[i]; 67 if (type != ops->type) 68 continue; 69 ret = ops->async_req_handle(async_req); 70 break; 71 } 72 73 return ret; 74 } 75 76 static int qce_handle_queue(struct qce_device *qce, 77 struct crypto_async_request *req) 78 { 79 struct crypto_async_request *async_req, *backlog; 80 unsigned long flags; 81 int ret = 0, err; 82 83 spin_lock_irqsave(&qce->lock, flags); 84 85 if (req) 86 ret = crypto_enqueue_request(&qce->queue, req); 87 88 /* busy, do not dequeue request */ 89 if (qce->req) { 90 spin_unlock_irqrestore(&qce->lock, flags); 91 return ret; 92 } 93 94 backlog = crypto_get_backlog(&qce->queue); 95 async_req = crypto_dequeue_request(&qce->queue); 96 if (async_req) 97 qce->req = async_req; 98 99 spin_unlock_irqrestore(&qce->lock, flags); 100 101 if (!async_req) 102 return ret; 103 104 if (backlog) { 105 spin_lock_bh(&qce->lock); 106 backlog->complete(backlog, -EINPROGRESS); 107 spin_unlock_bh(&qce->lock); 108 } 109 110 err = qce_handle_request(async_req); 111 if (err) { 112 qce->result = err; 113 tasklet_schedule(&qce->done_tasklet); 114 } 115 116 return ret; 117 } 118 119 static void qce_tasklet_req_done(unsigned long data) 120 { 121 struct qce_device *qce = (struct qce_device *)data; 122 struct crypto_async_request *req; 123 unsigned long flags; 124 125 spin_lock_irqsave(&qce->lock, flags); 126 req = qce->req; 127 qce->req = NULL; 128 spin_unlock_irqrestore(&qce->lock, flags); 129 130 if (req) 131 req->complete(req, qce->result); 132 133 qce_handle_queue(qce, NULL); 134 } 135 136 static int qce_async_request_enqueue(struct qce_device *qce, 137 struct crypto_async_request *req) 138 { 139 return qce_handle_queue(qce, req); 140 } 141 142 static void qce_async_request_done(struct qce_device *qce, int ret) 143 { 144 qce->result = ret; 145 tasklet_schedule(&qce->done_tasklet); 146 } 147 148 static int qce_check_version(struct qce_device *qce) 149 { 150 u32 major, minor, step; 151 152 qce_get_version(qce, &major, &minor, &step); 153 154 /* 155 * the driver does not support v5 with minor 0 because it has special 156 * alignment requirements. 157 */ 158 if (major != QCE_MAJOR_VERSION5 || minor == 0) 159 return -ENODEV; 160 161 qce->burst_size = QCE_BAM_BURST_SIZE; 162 qce->pipe_pair_id = 1; 163 164 dev_dbg(qce->dev, "Crypto device found, version %d.%d.%d\n", 165 major, minor, step); 166 167 return 0; 168 } 169 170 static int qce_crypto_probe(struct platform_device *pdev) 171 { 172 struct device *dev = &pdev->dev; 173 struct qce_device *qce; 174 int ret; 175 176 qce = devm_kzalloc(dev, sizeof(*qce), GFP_KERNEL); 177 if (!qce) 178 return -ENOMEM; 179 180 qce->dev = dev; 181 platform_set_drvdata(pdev, qce); 182 183 qce->base = devm_platform_ioremap_resource(pdev, 0); 184 if (IS_ERR(qce->base)) 185 return PTR_ERR(qce->base); 186 187 ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); 188 if (ret < 0) 189 return ret; 190 191 qce->core = devm_clk_get(qce->dev, "core"); 192 if (IS_ERR(qce->core)) 193 return PTR_ERR(qce->core); 194 195 qce->iface = devm_clk_get(qce->dev, "iface"); 196 if (IS_ERR(qce->iface)) 197 return PTR_ERR(qce->iface); 198 199 qce->bus = devm_clk_get(qce->dev, "bus"); 200 if (IS_ERR(qce->bus)) 201 return PTR_ERR(qce->bus); 202 203 ret = clk_prepare_enable(qce->core); 204 if (ret) 205 return ret; 206 207 ret = clk_prepare_enable(qce->iface); 208 if (ret) 209 goto err_clks_core; 210 211 ret = clk_prepare_enable(qce->bus); 212 if (ret) 213 goto err_clks_iface; 214 215 ret = qce_dma_request(qce->dev, &qce->dma); 216 if (ret) 217 goto err_clks; 218 219 ret = qce_check_version(qce); 220 if (ret) 221 goto err_clks; 222 223 spin_lock_init(&qce->lock); 224 tasklet_init(&qce->done_tasklet, qce_tasklet_req_done, 225 (unsigned long)qce); 226 crypto_init_queue(&qce->queue, QCE_QUEUE_LENGTH); 227 228 qce->async_req_enqueue = qce_async_request_enqueue; 229 qce->async_req_done = qce_async_request_done; 230 231 ret = qce_register_algs(qce); 232 if (ret) 233 goto err_dma; 234 235 return 0; 236 237 err_dma: 238 qce_dma_release(&qce->dma); 239 err_clks: 240 clk_disable_unprepare(qce->bus); 241 err_clks_iface: 242 clk_disable_unprepare(qce->iface); 243 err_clks_core: 244 clk_disable_unprepare(qce->core); 245 return ret; 246 } 247 248 static int qce_crypto_remove(struct platform_device *pdev) 249 { 250 struct qce_device *qce = platform_get_drvdata(pdev); 251 252 tasklet_kill(&qce->done_tasklet); 253 qce_unregister_algs(qce); 254 qce_dma_release(&qce->dma); 255 clk_disable_unprepare(qce->bus); 256 clk_disable_unprepare(qce->iface); 257 clk_disable_unprepare(qce->core); 258 return 0; 259 } 260 261 static const struct of_device_id qce_crypto_of_match[] = { 262 { .compatible = "qcom,crypto-v5.1", }, 263 {} 264 }; 265 MODULE_DEVICE_TABLE(of, qce_crypto_of_match); 266 267 static struct platform_driver qce_crypto_driver = { 268 .probe = qce_crypto_probe, 269 .remove = qce_crypto_remove, 270 .driver = { 271 .name = KBUILD_MODNAME, 272 .of_match_table = qce_crypto_of_match, 273 }, 274 }; 275 module_platform_driver(qce_crypto_driver); 276 277 MODULE_LICENSE("GPL v2"); 278 MODULE_DESCRIPTION("Qualcomm crypto engine driver"); 279 MODULE_ALIAS("platform:" KBUILD_MODNAME); 280 MODULE_AUTHOR("The Linux Foundation"); 281