Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Cryptographic API.
4 *
5 * Support for ATMEL DES/TDES HW acceleration.
6 *
7 * Copyright (c) 2012 Eukréa Electromatique - ATMEL
8 * Author: Nicolas Royer <nicolas@eukrea.com>
9 *
10 * Some ideas are from omap-aes.c drivers.
11 */
12
13
14#include <linux/kernel.h>
15#include <linux/module.h>
16#include <linux/slab.h>
17#include <linux/err.h>
18#include <linux/clk.h>
19#include <linux/io.h>
20#include <linux/hw_random.h>
21#include <linux/platform_device.h>
22
23#include <linux/device.h>
24#include <linux/dmaengine.h>
25#include <linux/init.h>
26#include <linux/errno.h>
27#include <linux/interrupt.h>
28#include <linux/irq.h>
29#include <linux/scatterlist.h>
30#include <linux/dma-mapping.h>
31#include <linux/mod_devicetable.h>
32#include <linux/delay.h>
33#include <linux/crypto.h>
34#include <crypto/scatterwalk.h>
35#include <crypto/algapi.h>
36#include <crypto/internal/des.h>
37#include <crypto/internal/skcipher.h>
38#include "atmel-tdes-regs.h"
39
40#define ATMEL_TDES_PRIORITY 300
41
42/* TDES flags */
43/* Reserve bits [17:16], [13:12], [2:0] for AES Mode Register */
44#define TDES_FLAGS_ENCRYPT TDES_MR_CYPHER_ENC
45#define TDES_FLAGS_OPMODE_MASK (TDES_MR_OPMOD_MASK | TDES_MR_CFBS_MASK)
46#define TDES_FLAGS_ECB TDES_MR_OPMOD_ECB
47#define TDES_FLAGS_CBC TDES_MR_OPMOD_CBC
48
49#define TDES_FLAGS_MODE_MASK (TDES_FLAGS_OPMODE_MASK | TDES_FLAGS_ENCRYPT)
50
51#define TDES_FLAGS_INIT BIT(3)
52#define TDES_FLAGS_FAST BIT(4)
53#define TDES_FLAGS_BUSY BIT(5)
54#define TDES_FLAGS_DMA BIT(6)
55
56#define ATMEL_TDES_QUEUE_LENGTH 50
57
58struct atmel_tdes_caps {
59 bool has_dma;
60};
61
62struct atmel_tdes_dev;
63
64struct atmel_tdes_ctx {
65 struct atmel_tdes_dev *dd;
66
67 int keylen;
68 u32 key[DES3_EDE_KEY_SIZE / sizeof(u32)];
69 unsigned long flags;
70
71 u16 block_size;
72};
73
74struct atmel_tdes_reqctx {
75 unsigned long mode;
76 u8 lastc[DES_BLOCK_SIZE];
77};
78
79struct atmel_tdes_dma {
80 struct dma_chan *chan;
81 struct dma_slave_config dma_conf;
82};
83
84struct atmel_tdes_dev {
85 struct list_head list;
86 unsigned long phys_base;
87 void __iomem *io_base;
88
89 struct atmel_tdes_ctx *ctx;
90 struct device *dev;
91 struct clk *iclk;
92 int irq;
93
94 unsigned long flags;
95
96 spinlock_t lock;
97 struct crypto_queue queue;
98
99 struct tasklet_struct done_task;
100 struct tasklet_struct queue_task;
101
102 struct skcipher_request *req;
103 size_t total;
104
105 struct scatterlist *in_sg;
106 unsigned int nb_in_sg;
107 size_t in_offset;
108 struct scatterlist *out_sg;
109 unsigned int nb_out_sg;
110 size_t out_offset;
111
112 size_t buflen;
113 size_t dma_size;
114
115 void *buf_in;
116 int dma_in;
117 dma_addr_t dma_addr_in;
118 struct atmel_tdes_dma dma_lch_in;
119
120 void *buf_out;
121 int dma_out;
122 dma_addr_t dma_addr_out;
123 struct atmel_tdes_dma dma_lch_out;
124
125 struct atmel_tdes_caps caps;
126
127 u32 hw_version;
128};
129
130struct atmel_tdes_drv {
131 struct list_head dev_list;
132 spinlock_t lock;
133};
134
135static struct atmel_tdes_drv atmel_tdes = {
136 .dev_list = LIST_HEAD_INIT(atmel_tdes.dev_list),
137 .lock = __SPIN_LOCK_UNLOCKED(atmel_tdes.lock),
138};
139
140static int atmel_tdes_sg_copy(struct scatterlist **sg, size_t *offset,
141 void *buf, size_t buflen, size_t total, int out)
142{
143 size_t count, off = 0;
144
145 while (buflen && total) {
146 count = min((*sg)->length - *offset, total);
147 count = min(count, buflen);
148
149 if (!count)
150 return off;
151
152 scatterwalk_map_and_copy(buf + off, *sg, *offset, count, out);
153
154 off += count;
155 buflen -= count;
156 *offset += count;
157 total -= count;
158
159 if (*offset == (*sg)->length) {
160 *sg = sg_next(*sg);
161 if (*sg)
162 *offset = 0;
163 else
164 total = 0;
165 }
166 }
167
168 return off;
169}
170
171static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset)
172{
173 return readl_relaxed(dd->io_base + offset);
174}
175
176static inline void atmel_tdes_write(struct atmel_tdes_dev *dd,
177 u32 offset, u32 value)
178{
179 writel_relaxed(value, dd->io_base + offset);
180}
181
182static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset,
183 const u32 *value, int count)
184{
185 for (; count--; value++, offset += 4)
186 atmel_tdes_write(dd, offset, *value);
187}
188
189static struct atmel_tdes_dev *atmel_tdes_dev_alloc(void)
190{
191 struct atmel_tdes_dev *tdes_dd;
192
193 spin_lock_bh(&atmel_tdes.lock);
194 /* One TDES IP per SoC. */
195 tdes_dd = list_first_entry_or_null(&atmel_tdes.dev_list,
196 struct atmel_tdes_dev, list);
197 spin_unlock_bh(&atmel_tdes.lock);
198 return tdes_dd;
199}
200
201static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd)
202{
203 int err;
204
205 err = clk_prepare_enable(dd->iclk);
206 if (err)
207 return err;
208
209 if (!(dd->flags & TDES_FLAGS_INIT)) {
210 atmel_tdes_write(dd, TDES_CR, TDES_CR_SWRST);
211 dd->flags |= TDES_FLAGS_INIT;
212 }
213
214 return 0;
215}
216
217static inline unsigned int atmel_tdes_get_version(struct atmel_tdes_dev *dd)
218{
219 return atmel_tdes_read(dd, TDES_HW_VERSION) & 0x00000fff;
220}
221
222static int atmel_tdes_hw_version_init(struct atmel_tdes_dev *dd)
223{
224 int err;
225
226 err = atmel_tdes_hw_init(dd);
227 if (err)
228 return err;
229
230 dd->hw_version = atmel_tdes_get_version(dd);
231
232 dev_info(dd->dev,
233 "version: 0x%x\n", dd->hw_version);
234
235 clk_disable_unprepare(dd->iclk);
236
237 return 0;
238}
239
240static void atmel_tdes_dma_callback(void *data)
241{
242 struct atmel_tdes_dev *dd = data;
243
244 /* dma_lch_out - completed */
245 tasklet_schedule(&dd->done_task);
246}
247
248static int atmel_tdes_write_ctrl(struct atmel_tdes_dev *dd)
249{
250 int err;
251 u32 valmr = TDES_MR_SMOD_PDC;
252
253 err = atmel_tdes_hw_init(dd);
254
255 if (err)
256 return err;
257
258 if (!dd->caps.has_dma)
259 atmel_tdes_write(dd, TDES_PTCR,
260 TDES_PTCR_TXTDIS | TDES_PTCR_RXTDIS);
261
262 /* MR register must be set before IV registers */
263 if (dd->ctx->keylen > (DES_KEY_SIZE << 1)) {
264 valmr |= TDES_MR_KEYMOD_3KEY;
265 valmr |= TDES_MR_TDESMOD_TDES;
266 } else if (dd->ctx->keylen > DES_KEY_SIZE) {
267 valmr |= TDES_MR_KEYMOD_2KEY;
268 valmr |= TDES_MR_TDESMOD_TDES;
269 } else {
270 valmr |= TDES_MR_TDESMOD_DES;
271 }
272
273 valmr |= dd->flags & TDES_FLAGS_MODE_MASK;
274
275 atmel_tdes_write(dd, TDES_MR, valmr);
276
277 atmel_tdes_write_n(dd, TDES_KEY1W1R, dd->ctx->key,
278 dd->ctx->keylen >> 2);
279
280 if (dd->req->iv && (valmr & TDES_MR_OPMOD_MASK) != TDES_MR_OPMOD_ECB)
281 atmel_tdes_write_n(dd, TDES_IV1R, (void *)dd->req->iv, 2);
282
283 return 0;
284}
285
286static int atmel_tdes_crypt_pdc_stop(struct atmel_tdes_dev *dd)
287{
288 int err = 0;
289 size_t count;
290
291 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
292
293 if (dd->flags & TDES_FLAGS_FAST) {
294 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
295 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
296 } else {
297 dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
298 dd->dma_size, DMA_FROM_DEVICE);
299
300 /* copy data */
301 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
302 dd->buf_out, dd->buflen, dd->dma_size, 1);
303 if (count != dd->dma_size) {
304 err = -EINVAL;
305 dev_dbg(dd->dev, "not all data converted: %zu\n", count);
306 }
307 }
308
309 return err;
310}
311
312static int atmel_tdes_buff_init(struct atmel_tdes_dev *dd)
313{
314 int err = -ENOMEM;
315
316 dd->buf_in = (void *)__get_free_pages(GFP_KERNEL, 0);
317 dd->buf_out = (void *)__get_free_pages(GFP_KERNEL, 0);
318 dd->buflen = PAGE_SIZE;
319 dd->buflen &= ~(DES_BLOCK_SIZE - 1);
320
321 if (!dd->buf_in || !dd->buf_out) {
322 dev_dbg(dd->dev, "unable to alloc pages.\n");
323 goto err_alloc;
324 }
325
326 /* MAP here */
327 dd->dma_addr_in = dma_map_single(dd->dev, dd->buf_in,
328 dd->buflen, DMA_TO_DEVICE);
329 err = dma_mapping_error(dd->dev, dd->dma_addr_in);
330 if (err) {
331 dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen);
332 goto err_map_in;
333 }
334
335 dd->dma_addr_out = dma_map_single(dd->dev, dd->buf_out,
336 dd->buflen, DMA_FROM_DEVICE);
337 err = dma_mapping_error(dd->dev, dd->dma_addr_out);
338 if (err) {
339 dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen);
340 goto err_map_out;
341 }
342
343 return 0;
344
345err_map_out:
346 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
347 DMA_TO_DEVICE);
348err_map_in:
349err_alloc:
350 free_page((unsigned long)dd->buf_out);
351 free_page((unsigned long)dd->buf_in);
352 return err;
353}
354
355static void atmel_tdes_buff_cleanup(struct atmel_tdes_dev *dd)
356{
357 dma_unmap_single(dd->dev, dd->dma_addr_out, dd->buflen,
358 DMA_FROM_DEVICE);
359 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
360 DMA_TO_DEVICE);
361 free_page((unsigned long)dd->buf_out);
362 free_page((unsigned long)dd->buf_in);
363}
364
365static int atmel_tdes_crypt_pdc(struct atmel_tdes_dev *dd,
366 dma_addr_t dma_addr_in,
367 dma_addr_t dma_addr_out, int length)
368{
369 int len32;
370
371 dd->dma_size = length;
372
373 if (!(dd->flags & TDES_FLAGS_FAST)) {
374 dma_sync_single_for_device(dd->dev, dma_addr_in, length,
375 DMA_TO_DEVICE);
376 }
377
378 len32 = DIV_ROUND_UP(length, sizeof(u32));
379
380 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
381 atmel_tdes_write(dd, TDES_TPR, dma_addr_in);
382 atmel_tdes_write(dd, TDES_TCR, len32);
383 atmel_tdes_write(dd, TDES_RPR, dma_addr_out);
384 atmel_tdes_write(dd, TDES_RCR, len32);
385
386 /* Enable Interrupt */
387 atmel_tdes_write(dd, TDES_IER, TDES_INT_ENDRX);
388
389 /* Start DMA transfer */
390 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTEN | TDES_PTCR_RXTEN);
391
392 return 0;
393}
394
395static int atmel_tdes_crypt_dma(struct atmel_tdes_dev *dd,
396 dma_addr_t dma_addr_in,
397 dma_addr_t dma_addr_out, int length)
398{
399 struct scatterlist sg[2];
400 struct dma_async_tx_descriptor *in_desc, *out_desc;
401 enum dma_slave_buswidth addr_width;
402
403 dd->dma_size = length;
404
405 if (!(dd->flags & TDES_FLAGS_FAST)) {
406 dma_sync_single_for_device(dd->dev, dma_addr_in, length,
407 DMA_TO_DEVICE);
408 }
409
410 addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
411
412 dd->dma_lch_in.dma_conf.dst_addr_width = addr_width;
413 dd->dma_lch_out.dma_conf.src_addr_width = addr_width;
414
415 dmaengine_slave_config(dd->dma_lch_in.chan, &dd->dma_lch_in.dma_conf);
416 dmaengine_slave_config(dd->dma_lch_out.chan, &dd->dma_lch_out.dma_conf);
417
418 dd->flags |= TDES_FLAGS_DMA;
419
420 sg_init_table(&sg[0], 1);
421 sg_dma_address(&sg[0]) = dma_addr_in;
422 sg_dma_len(&sg[0]) = length;
423
424 sg_init_table(&sg[1], 1);
425 sg_dma_address(&sg[1]) = dma_addr_out;
426 sg_dma_len(&sg[1]) = length;
427
428 in_desc = dmaengine_prep_slave_sg(dd->dma_lch_in.chan, &sg[0],
429 1, DMA_MEM_TO_DEV,
430 DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
431 if (!in_desc)
432 return -EINVAL;
433
434 out_desc = dmaengine_prep_slave_sg(dd->dma_lch_out.chan, &sg[1],
435 1, DMA_DEV_TO_MEM,
436 DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
437 if (!out_desc)
438 return -EINVAL;
439
440 out_desc->callback = atmel_tdes_dma_callback;
441 out_desc->callback_param = dd;
442
443 dmaengine_submit(out_desc);
444 dma_async_issue_pending(dd->dma_lch_out.chan);
445
446 dmaengine_submit(in_desc);
447 dma_async_issue_pending(dd->dma_lch_in.chan);
448
449 return 0;
450}
451
452static int atmel_tdes_crypt_start(struct atmel_tdes_dev *dd)
453{
454 int err, fast = 0, in, out;
455 size_t count;
456 dma_addr_t addr_in, addr_out;
457
458 if ((!dd->in_offset) && (!dd->out_offset)) {
459 /* check for alignment */
460 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) &&
461 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size);
462 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) &&
463 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size);
464 fast = in && out;
465
466 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg))
467 fast = 0;
468 }
469
470
471 if (fast) {
472 count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg));
473 count = min_t(size_t, count, sg_dma_len(dd->out_sg));
474
475 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
476 if (!err) {
477 dev_dbg(dd->dev, "dma_map_sg() error\n");
478 return -EINVAL;
479 }
480
481 err = dma_map_sg(dd->dev, dd->out_sg, 1,
482 DMA_FROM_DEVICE);
483 if (!err) {
484 dev_dbg(dd->dev, "dma_map_sg() error\n");
485 dma_unmap_sg(dd->dev, dd->in_sg, 1,
486 DMA_TO_DEVICE);
487 return -EINVAL;
488 }
489
490 addr_in = sg_dma_address(dd->in_sg);
491 addr_out = sg_dma_address(dd->out_sg);
492
493 dd->flags |= TDES_FLAGS_FAST;
494
495 } else {
496 /* use cache buffers */
497 count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset,
498 dd->buf_in, dd->buflen, dd->total, 0);
499
500 addr_in = dd->dma_addr_in;
501 addr_out = dd->dma_addr_out;
502
503 dd->flags &= ~TDES_FLAGS_FAST;
504 }
505
506 dd->total -= count;
507
508 if (dd->caps.has_dma)
509 err = atmel_tdes_crypt_dma(dd, addr_in, addr_out, count);
510 else
511 err = atmel_tdes_crypt_pdc(dd, addr_in, addr_out, count);
512
513 if (err && (dd->flags & TDES_FLAGS_FAST)) {
514 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
515 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE);
516 }
517
518 return err;
519}
520
521static void
522atmel_tdes_set_iv_as_last_ciphertext_block(struct atmel_tdes_dev *dd)
523{
524 struct skcipher_request *req = dd->req;
525 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
526 struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
527 unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
528
529 if (req->cryptlen < ivsize)
530 return;
531
532 if (rctx->mode & TDES_FLAGS_ENCRYPT)
533 scatterwalk_map_and_copy(req->iv, req->dst,
534 req->cryptlen - ivsize, ivsize, 0);
535 else
536 memcpy(req->iv, rctx->lastc, ivsize);
537
538}
539
540static void atmel_tdes_finish_req(struct atmel_tdes_dev *dd, int err)
541{
542 struct skcipher_request *req = dd->req;
543 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
544
545 clk_disable_unprepare(dd->iclk);
546
547 dd->flags &= ~TDES_FLAGS_BUSY;
548
549 if (!err && (rctx->mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB)
550 atmel_tdes_set_iv_as_last_ciphertext_block(dd);
551
552 skcipher_request_complete(req, err);
553}
554
555static int atmel_tdes_handle_queue(struct atmel_tdes_dev *dd,
556 struct skcipher_request *req)
557{
558 struct crypto_async_request *async_req, *backlog;
559 struct atmel_tdes_ctx *ctx;
560 struct atmel_tdes_reqctx *rctx;
561 unsigned long flags;
562 int err, ret = 0;
563
564 spin_lock_irqsave(&dd->lock, flags);
565 if (req)
566 ret = crypto_enqueue_request(&dd->queue, &req->base);
567 if (dd->flags & TDES_FLAGS_BUSY) {
568 spin_unlock_irqrestore(&dd->lock, flags);
569 return ret;
570 }
571 backlog = crypto_get_backlog(&dd->queue);
572 async_req = crypto_dequeue_request(&dd->queue);
573 if (async_req)
574 dd->flags |= TDES_FLAGS_BUSY;
575 spin_unlock_irqrestore(&dd->lock, flags);
576
577 if (!async_req)
578 return ret;
579
580 if (backlog)
581 crypto_request_complete(backlog, -EINPROGRESS);
582
583 req = skcipher_request_cast(async_req);
584
585 /* assign new request to device */
586 dd->req = req;
587 dd->total = req->cryptlen;
588 dd->in_offset = 0;
589 dd->in_sg = req->src;
590 dd->out_offset = 0;
591 dd->out_sg = req->dst;
592
593 rctx = skcipher_request_ctx(req);
594 ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
595 rctx->mode &= TDES_FLAGS_MODE_MASK;
596 dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode;
597 dd->ctx = ctx;
598
599 err = atmel_tdes_write_ctrl(dd);
600 if (!err)
601 err = atmel_tdes_crypt_start(dd);
602 if (err) {
603 /* des_task will not finish it, so do it here */
604 atmel_tdes_finish_req(dd, err);
605 tasklet_schedule(&dd->queue_task);
606 }
607
608 return ret;
609}
610
611static int atmel_tdes_crypt_dma_stop(struct atmel_tdes_dev *dd)
612{
613 int err = -EINVAL;
614 size_t count;
615
616 if (dd->flags & TDES_FLAGS_DMA) {
617 err = 0;
618 if (dd->flags & TDES_FLAGS_FAST) {
619 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
620 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
621 } else {
622 dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
623 dd->dma_size, DMA_FROM_DEVICE);
624
625 /* copy data */
626 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
627 dd->buf_out, dd->buflen, dd->dma_size, 1);
628 if (count != dd->dma_size) {
629 err = -EINVAL;
630 dev_dbg(dd->dev, "not all data converted: %zu\n", count);
631 }
632 }
633 }
634 return err;
635}
636
637static int atmel_tdes_crypt(struct skcipher_request *req, unsigned long mode)
638{
639 struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
640 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(skcipher);
641 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
642 struct device *dev = ctx->dd->dev;
643
644 if (!req->cryptlen)
645 return 0;
646
647 if (!IS_ALIGNED(req->cryptlen, DES_BLOCK_SIZE)) {
648 dev_dbg(dev, "request size is not exact amount of DES blocks\n");
649 return -EINVAL;
650 }
651 ctx->block_size = DES_BLOCK_SIZE;
652
653 rctx->mode = mode;
654
655 if ((mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB &&
656 !(mode & TDES_FLAGS_ENCRYPT)) {
657 unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
658
659 if (req->cryptlen >= ivsize)
660 scatterwalk_map_and_copy(rctx->lastc, req->src,
661 req->cryptlen - ivsize,
662 ivsize, 0);
663 }
664
665 return atmel_tdes_handle_queue(ctx->dd, req);
666}
667
668static int atmel_tdes_dma_init(struct atmel_tdes_dev *dd)
669{
670 int ret;
671
672 /* Try to grab 2 DMA channels */
673 dd->dma_lch_in.chan = dma_request_chan(dd->dev, "tx");
674 if (IS_ERR(dd->dma_lch_in.chan)) {
675 ret = PTR_ERR(dd->dma_lch_in.chan);
676 goto err_dma_in;
677 }
678
679 dd->dma_lch_in.dma_conf.dst_addr = dd->phys_base +
680 TDES_IDATA1R;
681 dd->dma_lch_in.dma_conf.src_maxburst = 1;
682 dd->dma_lch_in.dma_conf.src_addr_width =
683 DMA_SLAVE_BUSWIDTH_4_BYTES;
684 dd->dma_lch_in.dma_conf.dst_maxburst = 1;
685 dd->dma_lch_in.dma_conf.dst_addr_width =
686 DMA_SLAVE_BUSWIDTH_4_BYTES;
687 dd->dma_lch_in.dma_conf.device_fc = false;
688
689 dd->dma_lch_out.chan = dma_request_chan(dd->dev, "rx");
690 if (IS_ERR(dd->dma_lch_out.chan)) {
691 ret = PTR_ERR(dd->dma_lch_out.chan);
692 goto err_dma_out;
693 }
694
695 dd->dma_lch_out.dma_conf.src_addr = dd->phys_base +
696 TDES_ODATA1R;
697 dd->dma_lch_out.dma_conf.src_maxburst = 1;
698 dd->dma_lch_out.dma_conf.src_addr_width =
699 DMA_SLAVE_BUSWIDTH_4_BYTES;
700 dd->dma_lch_out.dma_conf.dst_maxburst = 1;
701 dd->dma_lch_out.dma_conf.dst_addr_width =
702 DMA_SLAVE_BUSWIDTH_4_BYTES;
703 dd->dma_lch_out.dma_conf.device_fc = false;
704
705 return 0;
706
707err_dma_out:
708 dma_release_channel(dd->dma_lch_in.chan);
709err_dma_in:
710 dev_err(dd->dev, "no DMA channel available\n");
711 return ret;
712}
713
714static void atmel_tdes_dma_cleanup(struct atmel_tdes_dev *dd)
715{
716 dma_release_channel(dd->dma_lch_in.chan);
717 dma_release_channel(dd->dma_lch_out.chan);
718}
719
720static int atmel_des_setkey(struct crypto_skcipher *tfm, const u8 *key,
721 unsigned int keylen)
722{
723 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
724 int err;
725
726 err = verify_skcipher_des_key(tfm, key);
727 if (err)
728 return err;
729
730 memcpy(ctx->key, key, keylen);
731 ctx->keylen = keylen;
732
733 return 0;
734}
735
736static int atmel_tdes_setkey(struct crypto_skcipher *tfm, const u8 *key,
737 unsigned int keylen)
738{
739 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
740 int err;
741
742 err = verify_skcipher_des3_key(tfm, key);
743 if (err)
744 return err;
745
746 memcpy(ctx->key, key, keylen);
747 ctx->keylen = keylen;
748
749 return 0;
750}
751
752static int atmel_tdes_ecb_encrypt(struct skcipher_request *req)
753{
754 return atmel_tdes_crypt(req, TDES_FLAGS_ECB | TDES_FLAGS_ENCRYPT);
755}
756
757static int atmel_tdes_ecb_decrypt(struct skcipher_request *req)
758{
759 return atmel_tdes_crypt(req, TDES_FLAGS_ECB);
760}
761
762static int atmel_tdes_cbc_encrypt(struct skcipher_request *req)
763{
764 return atmel_tdes_crypt(req, TDES_FLAGS_CBC | TDES_FLAGS_ENCRYPT);
765}
766
767static int atmel_tdes_cbc_decrypt(struct skcipher_request *req)
768{
769 return atmel_tdes_crypt(req, TDES_FLAGS_CBC);
770}
771
772static int atmel_tdes_init_tfm(struct crypto_skcipher *tfm)
773{
774 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
775
776 ctx->dd = atmel_tdes_dev_alloc();
777 if (!ctx->dd)
778 return -ENODEV;
779
780 crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_tdes_reqctx));
781
782 return 0;
783}
784
785static void atmel_tdes_skcipher_alg_init(struct skcipher_alg *alg)
786{
787 alg->base.cra_priority = ATMEL_TDES_PRIORITY;
788 alg->base.cra_flags = CRYPTO_ALG_ASYNC;
789 alg->base.cra_ctxsize = sizeof(struct atmel_tdes_ctx);
790 alg->base.cra_module = THIS_MODULE;
791
792 alg->init = atmel_tdes_init_tfm;
793}
794
795static struct skcipher_alg tdes_algs[] = {
796{
797 .base.cra_name = "ecb(des)",
798 .base.cra_driver_name = "atmel-ecb-des",
799 .base.cra_blocksize = DES_BLOCK_SIZE,
800 .base.cra_alignmask = 0x7,
801
802 .min_keysize = DES_KEY_SIZE,
803 .max_keysize = DES_KEY_SIZE,
804 .setkey = atmel_des_setkey,
805 .encrypt = atmel_tdes_ecb_encrypt,
806 .decrypt = atmel_tdes_ecb_decrypt,
807},
808{
809 .base.cra_name = "cbc(des)",
810 .base.cra_driver_name = "atmel-cbc-des",
811 .base.cra_blocksize = DES_BLOCK_SIZE,
812 .base.cra_alignmask = 0x7,
813
814 .min_keysize = DES_KEY_SIZE,
815 .max_keysize = DES_KEY_SIZE,
816 .ivsize = DES_BLOCK_SIZE,
817 .setkey = atmel_des_setkey,
818 .encrypt = atmel_tdes_cbc_encrypt,
819 .decrypt = atmel_tdes_cbc_decrypt,
820},
821{
822 .base.cra_name = "ecb(des3_ede)",
823 .base.cra_driver_name = "atmel-ecb-tdes",
824 .base.cra_blocksize = DES_BLOCK_SIZE,
825 .base.cra_alignmask = 0x7,
826
827 .min_keysize = DES3_EDE_KEY_SIZE,
828 .max_keysize = DES3_EDE_KEY_SIZE,
829 .setkey = atmel_tdes_setkey,
830 .encrypt = atmel_tdes_ecb_encrypt,
831 .decrypt = atmel_tdes_ecb_decrypt,
832},
833{
834 .base.cra_name = "cbc(des3_ede)",
835 .base.cra_driver_name = "atmel-cbc-tdes",
836 .base.cra_blocksize = DES_BLOCK_SIZE,
837 .base.cra_alignmask = 0x7,
838
839 .min_keysize = DES3_EDE_KEY_SIZE,
840 .max_keysize = DES3_EDE_KEY_SIZE,
841 .setkey = atmel_tdes_setkey,
842 .encrypt = atmel_tdes_cbc_encrypt,
843 .decrypt = atmel_tdes_cbc_decrypt,
844 .ivsize = DES_BLOCK_SIZE,
845},
846};
847
848static void atmel_tdes_queue_task(unsigned long data)
849{
850 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *)data;
851
852 atmel_tdes_handle_queue(dd, NULL);
853}
854
855static void atmel_tdes_done_task(unsigned long data)
856{
857 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *) data;
858 int err;
859
860 if (!(dd->flags & TDES_FLAGS_DMA))
861 err = atmel_tdes_crypt_pdc_stop(dd);
862 else
863 err = atmel_tdes_crypt_dma_stop(dd);
864
865 if (dd->total && !err) {
866 if (dd->flags & TDES_FLAGS_FAST) {
867 dd->in_sg = sg_next(dd->in_sg);
868 dd->out_sg = sg_next(dd->out_sg);
869 if (!dd->in_sg || !dd->out_sg)
870 err = -EINVAL;
871 }
872 if (!err)
873 err = atmel_tdes_crypt_start(dd);
874 if (!err)
875 return; /* DMA started. Not fininishing. */
876 }
877
878 atmel_tdes_finish_req(dd, err);
879 atmel_tdes_handle_queue(dd, NULL);
880}
881
882static irqreturn_t atmel_tdes_irq(int irq, void *dev_id)
883{
884 struct atmel_tdes_dev *tdes_dd = dev_id;
885 u32 reg;
886
887 reg = atmel_tdes_read(tdes_dd, TDES_ISR);
888 if (reg & atmel_tdes_read(tdes_dd, TDES_IMR)) {
889 atmel_tdes_write(tdes_dd, TDES_IDR, reg);
890 if (TDES_FLAGS_BUSY & tdes_dd->flags)
891 tasklet_schedule(&tdes_dd->done_task);
892 else
893 dev_warn(tdes_dd->dev, "TDES interrupt when no active requests.\n");
894 return IRQ_HANDLED;
895 }
896
897 return IRQ_NONE;
898}
899
900static void atmel_tdes_unregister_algs(struct atmel_tdes_dev *dd)
901{
902 int i;
903
904 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++)
905 crypto_unregister_skcipher(&tdes_algs[i]);
906}
907
908static int atmel_tdes_register_algs(struct atmel_tdes_dev *dd)
909{
910 int err, i, j;
911
912 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) {
913 atmel_tdes_skcipher_alg_init(&tdes_algs[i]);
914
915 err = crypto_register_skcipher(&tdes_algs[i]);
916 if (err)
917 goto err_tdes_algs;
918 }
919
920 return 0;
921
922err_tdes_algs:
923 for (j = 0; j < i; j++)
924 crypto_unregister_skcipher(&tdes_algs[j]);
925
926 return err;
927}
928
929static void atmel_tdes_get_cap(struct atmel_tdes_dev *dd)
930{
931
932 dd->caps.has_dma = 0;
933
934 /* keep only major version number */
935 switch (dd->hw_version & 0xf00) {
936 case 0x800:
937 case 0x700:
938 dd->caps.has_dma = 1;
939 break;
940 case 0x600:
941 break;
942 default:
943 dev_warn(dd->dev,
944 "Unmanaged tdes version, set minimum capabilities\n");
945 break;
946 }
947}
948
949static const struct of_device_id atmel_tdes_dt_ids[] = {
950 { .compatible = "atmel,at91sam9g46-tdes" },
951 { /* sentinel */ }
952};
953MODULE_DEVICE_TABLE(of, atmel_tdes_dt_ids);
954
955static int atmel_tdes_probe(struct platform_device *pdev)
956{
957 struct atmel_tdes_dev *tdes_dd;
958 struct device *dev = &pdev->dev;
959 struct resource *tdes_res;
960 int err;
961
962 tdes_dd = devm_kmalloc(&pdev->dev, sizeof(*tdes_dd), GFP_KERNEL);
963 if (!tdes_dd)
964 return -ENOMEM;
965
966 tdes_dd->dev = dev;
967
968 platform_set_drvdata(pdev, tdes_dd);
969
970 INIT_LIST_HEAD(&tdes_dd->list);
971 spin_lock_init(&tdes_dd->lock);
972
973 tasklet_init(&tdes_dd->done_task, atmel_tdes_done_task,
974 (unsigned long)tdes_dd);
975 tasklet_init(&tdes_dd->queue_task, atmel_tdes_queue_task,
976 (unsigned long)tdes_dd);
977
978 crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH);
979
980 tdes_dd->io_base = devm_platform_get_and_ioremap_resource(pdev, 0, &tdes_res);
981 if (IS_ERR(tdes_dd->io_base)) {
982 err = PTR_ERR(tdes_dd->io_base);
983 goto err_tasklet_kill;
984 }
985 tdes_dd->phys_base = tdes_res->start;
986
987 /* Get the IRQ */
988 tdes_dd->irq = platform_get_irq(pdev, 0);
989 if (tdes_dd->irq < 0) {
990 err = tdes_dd->irq;
991 goto err_tasklet_kill;
992 }
993
994 err = devm_request_irq(&pdev->dev, tdes_dd->irq, atmel_tdes_irq,
995 IRQF_SHARED, "atmel-tdes", tdes_dd);
996 if (err) {
997 dev_err(dev, "unable to request tdes irq.\n");
998 goto err_tasklet_kill;
999 }
1000
1001 /* Initializing the clock */
1002 tdes_dd->iclk = devm_clk_get(&pdev->dev, "tdes_clk");
1003 if (IS_ERR(tdes_dd->iclk)) {
1004 dev_err(dev, "clock initialization failed.\n");
1005 err = PTR_ERR(tdes_dd->iclk);
1006 goto err_tasklet_kill;
1007 }
1008
1009 err = atmel_tdes_hw_version_init(tdes_dd);
1010 if (err)
1011 goto err_tasklet_kill;
1012
1013 atmel_tdes_get_cap(tdes_dd);
1014
1015 err = atmel_tdes_buff_init(tdes_dd);
1016 if (err)
1017 goto err_tasklet_kill;
1018
1019 if (tdes_dd->caps.has_dma) {
1020 err = atmel_tdes_dma_init(tdes_dd);
1021 if (err)
1022 goto err_buff_cleanup;
1023
1024 dev_info(dev, "using %s, %s for DMA transfers\n",
1025 dma_chan_name(tdes_dd->dma_lch_in.chan),
1026 dma_chan_name(tdes_dd->dma_lch_out.chan));
1027 }
1028
1029 spin_lock(&atmel_tdes.lock);
1030 list_add_tail(&tdes_dd->list, &atmel_tdes.dev_list);
1031 spin_unlock(&atmel_tdes.lock);
1032
1033 err = atmel_tdes_register_algs(tdes_dd);
1034 if (err)
1035 goto err_algs;
1036
1037 dev_info(dev, "Atmel DES/TDES\n");
1038
1039 return 0;
1040
1041err_algs:
1042 spin_lock(&atmel_tdes.lock);
1043 list_del(&tdes_dd->list);
1044 spin_unlock(&atmel_tdes.lock);
1045 if (tdes_dd->caps.has_dma)
1046 atmel_tdes_dma_cleanup(tdes_dd);
1047err_buff_cleanup:
1048 atmel_tdes_buff_cleanup(tdes_dd);
1049err_tasklet_kill:
1050 tasklet_kill(&tdes_dd->done_task);
1051 tasklet_kill(&tdes_dd->queue_task);
1052
1053 return err;
1054}
1055
1056static void atmel_tdes_remove(struct platform_device *pdev)
1057{
1058 struct atmel_tdes_dev *tdes_dd = platform_get_drvdata(pdev);
1059
1060 spin_lock(&atmel_tdes.lock);
1061 list_del(&tdes_dd->list);
1062 spin_unlock(&atmel_tdes.lock);
1063
1064 atmel_tdes_unregister_algs(tdes_dd);
1065
1066 tasklet_kill(&tdes_dd->done_task);
1067 tasklet_kill(&tdes_dd->queue_task);
1068
1069 if (tdes_dd->caps.has_dma)
1070 atmel_tdes_dma_cleanup(tdes_dd);
1071
1072 atmel_tdes_buff_cleanup(tdes_dd);
1073}
1074
1075static struct platform_driver atmel_tdes_driver = {
1076 .probe = atmel_tdes_probe,
1077 .remove_new = atmel_tdes_remove,
1078 .driver = {
1079 .name = "atmel_tdes",
1080 .of_match_table = atmel_tdes_dt_ids,
1081 },
1082};
1083
1084module_platform_driver(atmel_tdes_driver);
1085
1086MODULE_DESCRIPTION("Atmel DES/TDES hw acceleration support.");
1087MODULE_LICENSE("GPL v2");
1088MODULE_AUTHOR("Nicolas Royer - Eukréa Electromatique");
1/*
2 * Cryptographic API.
3 *
4 * Support for ATMEL DES/TDES HW acceleration.
5 *
6 * Copyright (c) 2012 Eukréa Electromatique - ATMEL
7 * Author: Nicolas Royer <nicolas@eukrea.com>
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License version 2 as published
11 * by the Free Software Foundation.
12 *
13 * Some ideas are from omap-aes.c drivers.
14 */
15
16
17#include <linux/kernel.h>
18#include <linux/module.h>
19#include <linux/slab.h>
20#include <linux/err.h>
21#include <linux/clk.h>
22#include <linux/io.h>
23#include <linux/hw_random.h>
24#include <linux/platform_device.h>
25
26#include <linux/device.h>
27#include <linux/init.h>
28#include <linux/errno.h>
29#include <linux/interrupt.h>
30#include <linux/irq.h>
31#include <linux/scatterlist.h>
32#include <linux/dma-mapping.h>
33#include <linux/of_device.h>
34#include <linux/delay.h>
35#include <linux/crypto.h>
36#include <linux/cryptohash.h>
37#include <crypto/scatterwalk.h>
38#include <crypto/algapi.h>
39#include <crypto/des.h>
40#include <crypto/hash.h>
41#include <crypto/internal/hash.h>
42#include <linux/platform_data/crypto-atmel.h>
43#include "atmel-tdes-regs.h"
44
45/* TDES flags */
46#define TDES_FLAGS_MODE_MASK 0x00ff
47#define TDES_FLAGS_ENCRYPT BIT(0)
48#define TDES_FLAGS_CBC BIT(1)
49#define TDES_FLAGS_CFB BIT(2)
50#define TDES_FLAGS_CFB8 BIT(3)
51#define TDES_FLAGS_CFB16 BIT(4)
52#define TDES_FLAGS_CFB32 BIT(5)
53#define TDES_FLAGS_CFB64 BIT(6)
54#define TDES_FLAGS_OFB BIT(7)
55
56#define TDES_FLAGS_INIT BIT(16)
57#define TDES_FLAGS_FAST BIT(17)
58#define TDES_FLAGS_BUSY BIT(18)
59#define TDES_FLAGS_DMA BIT(19)
60
61#define ATMEL_TDES_QUEUE_LENGTH 50
62
63#define CFB8_BLOCK_SIZE 1
64#define CFB16_BLOCK_SIZE 2
65#define CFB32_BLOCK_SIZE 4
66
67struct atmel_tdes_caps {
68 bool has_dma;
69 u32 has_cfb_3keys;
70};
71
72struct atmel_tdes_dev;
73
74struct atmel_tdes_ctx {
75 struct atmel_tdes_dev *dd;
76
77 int keylen;
78 u32 key[3*DES_KEY_SIZE / sizeof(u32)];
79 unsigned long flags;
80
81 u16 block_size;
82};
83
84struct atmel_tdes_reqctx {
85 unsigned long mode;
86};
87
88struct atmel_tdes_dma {
89 struct dma_chan *chan;
90 struct dma_slave_config dma_conf;
91};
92
93struct atmel_tdes_dev {
94 struct list_head list;
95 unsigned long phys_base;
96 void __iomem *io_base;
97
98 struct atmel_tdes_ctx *ctx;
99 struct device *dev;
100 struct clk *iclk;
101 int irq;
102
103 unsigned long flags;
104 int err;
105
106 spinlock_t lock;
107 struct crypto_queue queue;
108
109 struct tasklet_struct done_task;
110 struct tasklet_struct queue_task;
111
112 struct ablkcipher_request *req;
113 size_t total;
114
115 struct scatterlist *in_sg;
116 unsigned int nb_in_sg;
117 size_t in_offset;
118 struct scatterlist *out_sg;
119 unsigned int nb_out_sg;
120 size_t out_offset;
121
122 size_t buflen;
123 size_t dma_size;
124
125 void *buf_in;
126 int dma_in;
127 dma_addr_t dma_addr_in;
128 struct atmel_tdes_dma dma_lch_in;
129
130 void *buf_out;
131 int dma_out;
132 dma_addr_t dma_addr_out;
133 struct atmel_tdes_dma dma_lch_out;
134
135 struct atmel_tdes_caps caps;
136
137 u32 hw_version;
138};
139
140struct atmel_tdes_drv {
141 struct list_head dev_list;
142 spinlock_t lock;
143};
144
145static struct atmel_tdes_drv atmel_tdes = {
146 .dev_list = LIST_HEAD_INIT(atmel_tdes.dev_list),
147 .lock = __SPIN_LOCK_UNLOCKED(atmel_tdes.lock),
148};
149
150static int atmel_tdes_sg_copy(struct scatterlist **sg, size_t *offset,
151 void *buf, size_t buflen, size_t total, int out)
152{
153 unsigned int count, off = 0;
154
155 while (buflen && total) {
156 count = min((*sg)->length - *offset, total);
157 count = min(count, buflen);
158
159 if (!count)
160 return off;
161
162 scatterwalk_map_and_copy(buf + off, *sg, *offset, count, out);
163
164 off += count;
165 buflen -= count;
166 *offset += count;
167 total -= count;
168
169 if (*offset == (*sg)->length) {
170 *sg = sg_next(*sg);
171 if (*sg)
172 *offset = 0;
173 else
174 total = 0;
175 }
176 }
177
178 return off;
179}
180
181static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset)
182{
183 return readl_relaxed(dd->io_base + offset);
184}
185
186static inline void atmel_tdes_write(struct atmel_tdes_dev *dd,
187 u32 offset, u32 value)
188{
189 writel_relaxed(value, dd->io_base + offset);
190}
191
192static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset,
193 u32 *value, int count)
194{
195 for (; count--; value++, offset += 4)
196 atmel_tdes_write(dd, offset, *value);
197}
198
199static struct atmel_tdes_dev *atmel_tdes_find_dev(struct atmel_tdes_ctx *ctx)
200{
201 struct atmel_tdes_dev *tdes_dd = NULL;
202 struct atmel_tdes_dev *tmp;
203
204 spin_lock_bh(&atmel_tdes.lock);
205 if (!ctx->dd) {
206 list_for_each_entry(tmp, &atmel_tdes.dev_list, list) {
207 tdes_dd = tmp;
208 break;
209 }
210 ctx->dd = tdes_dd;
211 } else {
212 tdes_dd = ctx->dd;
213 }
214 spin_unlock_bh(&atmel_tdes.lock);
215
216 return tdes_dd;
217}
218
219static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd)
220{
221 int err;
222
223 err = clk_prepare_enable(dd->iclk);
224 if (err)
225 return err;
226
227 if (!(dd->flags & TDES_FLAGS_INIT)) {
228 atmel_tdes_write(dd, TDES_CR, TDES_CR_SWRST);
229 dd->flags |= TDES_FLAGS_INIT;
230 dd->err = 0;
231 }
232
233 return 0;
234}
235
236static inline unsigned int atmel_tdes_get_version(struct atmel_tdes_dev *dd)
237{
238 return atmel_tdes_read(dd, TDES_HW_VERSION) & 0x00000fff;
239}
240
241static void atmel_tdes_hw_version_init(struct atmel_tdes_dev *dd)
242{
243 atmel_tdes_hw_init(dd);
244
245 dd->hw_version = atmel_tdes_get_version(dd);
246
247 dev_info(dd->dev,
248 "version: 0x%x\n", dd->hw_version);
249
250 clk_disable_unprepare(dd->iclk);
251}
252
253static void atmel_tdes_dma_callback(void *data)
254{
255 struct atmel_tdes_dev *dd = data;
256
257 /* dma_lch_out - completed */
258 tasklet_schedule(&dd->done_task);
259}
260
261static int atmel_tdes_write_ctrl(struct atmel_tdes_dev *dd)
262{
263 int err;
264 u32 valcr = 0, valmr = TDES_MR_SMOD_PDC;
265
266 err = atmel_tdes_hw_init(dd);
267
268 if (err)
269 return err;
270
271 if (!dd->caps.has_dma)
272 atmel_tdes_write(dd, TDES_PTCR,
273 TDES_PTCR_TXTDIS | TDES_PTCR_RXTDIS);
274
275 /* MR register must be set before IV registers */
276 if (dd->ctx->keylen > (DES_KEY_SIZE << 1)) {
277 valmr |= TDES_MR_KEYMOD_3KEY;
278 valmr |= TDES_MR_TDESMOD_TDES;
279 } else if (dd->ctx->keylen > DES_KEY_SIZE) {
280 valmr |= TDES_MR_KEYMOD_2KEY;
281 valmr |= TDES_MR_TDESMOD_TDES;
282 } else {
283 valmr |= TDES_MR_TDESMOD_DES;
284 }
285
286 if (dd->flags & TDES_FLAGS_CBC) {
287 valmr |= TDES_MR_OPMOD_CBC;
288 } else if (dd->flags & TDES_FLAGS_CFB) {
289 valmr |= TDES_MR_OPMOD_CFB;
290
291 if (dd->flags & TDES_FLAGS_CFB8)
292 valmr |= TDES_MR_CFBS_8b;
293 else if (dd->flags & TDES_FLAGS_CFB16)
294 valmr |= TDES_MR_CFBS_16b;
295 else if (dd->flags & TDES_FLAGS_CFB32)
296 valmr |= TDES_MR_CFBS_32b;
297 else if (dd->flags & TDES_FLAGS_CFB64)
298 valmr |= TDES_MR_CFBS_64b;
299 } else if (dd->flags & TDES_FLAGS_OFB) {
300 valmr |= TDES_MR_OPMOD_OFB;
301 }
302
303 if ((dd->flags & TDES_FLAGS_ENCRYPT) || (dd->flags & TDES_FLAGS_OFB))
304 valmr |= TDES_MR_CYPHER_ENC;
305
306 atmel_tdes_write(dd, TDES_CR, valcr);
307 atmel_tdes_write(dd, TDES_MR, valmr);
308
309 atmel_tdes_write_n(dd, TDES_KEY1W1R, dd->ctx->key,
310 dd->ctx->keylen >> 2);
311
312 if (((dd->flags & TDES_FLAGS_CBC) || (dd->flags & TDES_FLAGS_CFB) ||
313 (dd->flags & TDES_FLAGS_OFB)) && dd->req->info) {
314 atmel_tdes_write_n(dd, TDES_IV1R, dd->req->info, 2);
315 }
316
317 return 0;
318}
319
320static int atmel_tdes_crypt_pdc_stop(struct atmel_tdes_dev *dd)
321{
322 int err = 0;
323 size_t count;
324
325 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
326
327 if (dd->flags & TDES_FLAGS_FAST) {
328 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
329 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
330 } else {
331 dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
332 dd->dma_size, DMA_FROM_DEVICE);
333
334 /* copy data */
335 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
336 dd->buf_out, dd->buflen, dd->dma_size, 1);
337 if (count != dd->dma_size) {
338 err = -EINVAL;
339 pr_err("not all data converted: %u\n", count);
340 }
341 }
342
343 return err;
344}
345
346static int atmel_tdes_buff_init(struct atmel_tdes_dev *dd)
347{
348 int err = -ENOMEM;
349
350 dd->buf_in = (void *)__get_free_pages(GFP_KERNEL, 0);
351 dd->buf_out = (void *)__get_free_pages(GFP_KERNEL, 0);
352 dd->buflen = PAGE_SIZE;
353 dd->buflen &= ~(DES_BLOCK_SIZE - 1);
354
355 if (!dd->buf_in || !dd->buf_out) {
356 dev_err(dd->dev, "unable to alloc pages.\n");
357 goto err_alloc;
358 }
359
360 /* MAP here */
361 dd->dma_addr_in = dma_map_single(dd->dev, dd->buf_in,
362 dd->buflen, DMA_TO_DEVICE);
363 if (dma_mapping_error(dd->dev, dd->dma_addr_in)) {
364 dev_err(dd->dev, "dma %d bytes error\n", dd->buflen);
365 err = -EINVAL;
366 goto err_map_in;
367 }
368
369 dd->dma_addr_out = dma_map_single(dd->dev, dd->buf_out,
370 dd->buflen, DMA_FROM_DEVICE);
371 if (dma_mapping_error(dd->dev, dd->dma_addr_out)) {
372 dev_err(dd->dev, "dma %d bytes error\n", dd->buflen);
373 err = -EINVAL;
374 goto err_map_out;
375 }
376
377 return 0;
378
379err_map_out:
380 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
381 DMA_TO_DEVICE);
382err_map_in:
383err_alloc:
384 free_page((unsigned long)dd->buf_out);
385 free_page((unsigned long)dd->buf_in);
386 if (err)
387 pr_err("error: %d\n", err);
388 return err;
389}
390
391static void atmel_tdes_buff_cleanup(struct atmel_tdes_dev *dd)
392{
393 dma_unmap_single(dd->dev, dd->dma_addr_out, dd->buflen,
394 DMA_FROM_DEVICE);
395 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
396 DMA_TO_DEVICE);
397 free_page((unsigned long)dd->buf_out);
398 free_page((unsigned long)dd->buf_in);
399}
400
401static int atmel_tdes_crypt_pdc(struct crypto_tfm *tfm, dma_addr_t dma_addr_in,
402 dma_addr_t dma_addr_out, int length)
403{
404 struct atmel_tdes_ctx *ctx = crypto_tfm_ctx(tfm);
405 struct atmel_tdes_dev *dd = ctx->dd;
406 int len32;
407
408 dd->dma_size = length;
409
410 if (!(dd->flags & TDES_FLAGS_FAST)) {
411 dma_sync_single_for_device(dd->dev, dma_addr_in, length,
412 DMA_TO_DEVICE);
413 }
414
415 if ((dd->flags & TDES_FLAGS_CFB) && (dd->flags & TDES_FLAGS_CFB8))
416 len32 = DIV_ROUND_UP(length, sizeof(u8));
417 else if ((dd->flags & TDES_FLAGS_CFB) && (dd->flags & TDES_FLAGS_CFB16))
418 len32 = DIV_ROUND_UP(length, sizeof(u16));
419 else
420 len32 = DIV_ROUND_UP(length, sizeof(u32));
421
422 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
423 atmel_tdes_write(dd, TDES_TPR, dma_addr_in);
424 atmel_tdes_write(dd, TDES_TCR, len32);
425 atmel_tdes_write(dd, TDES_RPR, dma_addr_out);
426 atmel_tdes_write(dd, TDES_RCR, len32);
427
428 /* Enable Interrupt */
429 atmel_tdes_write(dd, TDES_IER, TDES_INT_ENDRX);
430
431 /* Start DMA transfer */
432 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTEN | TDES_PTCR_RXTEN);
433
434 return 0;
435}
436
437static int atmel_tdes_crypt_dma(struct crypto_tfm *tfm, dma_addr_t dma_addr_in,
438 dma_addr_t dma_addr_out, int length)
439{
440 struct atmel_tdes_ctx *ctx = crypto_tfm_ctx(tfm);
441 struct atmel_tdes_dev *dd = ctx->dd;
442 struct scatterlist sg[2];
443 struct dma_async_tx_descriptor *in_desc, *out_desc;
444
445 dd->dma_size = length;
446
447 if (!(dd->flags & TDES_FLAGS_FAST)) {
448 dma_sync_single_for_device(dd->dev, dma_addr_in, length,
449 DMA_TO_DEVICE);
450 }
451
452 if (dd->flags & TDES_FLAGS_CFB8) {
453 dd->dma_lch_in.dma_conf.dst_addr_width =
454 DMA_SLAVE_BUSWIDTH_1_BYTE;
455 dd->dma_lch_out.dma_conf.src_addr_width =
456 DMA_SLAVE_BUSWIDTH_1_BYTE;
457 } else if (dd->flags & TDES_FLAGS_CFB16) {
458 dd->dma_lch_in.dma_conf.dst_addr_width =
459 DMA_SLAVE_BUSWIDTH_2_BYTES;
460 dd->dma_lch_out.dma_conf.src_addr_width =
461 DMA_SLAVE_BUSWIDTH_2_BYTES;
462 } else {
463 dd->dma_lch_in.dma_conf.dst_addr_width =
464 DMA_SLAVE_BUSWIDTH_4_BYTES;
465 dd->dma_lch_out.dma_conf.src_addr_width =
466 DMA_SLAVE_BUSWIDTH_4_BYTES;
467 }
468
469 dmaengine_slave_config(dd->dma_lch_in.chan, &dd->dma_lch_in.dma_conf);
470 dmaengine_slave_config(dd->dma_lch_out.chan, &dd->dma_lch_out.dma_conf);
471
472 dd->flags |= TDES_FLAGS_DMA;
473
474 sg_init_table(&sg[0], 1);
475 sg_dma_address(&sg[0]) = dma_addr_in;
476 sg_dma_len(&sg[0]) = length;
477
478 sg_init_table(&sg[1], 1);
479 sg_dma_address(&sg[1]) = dma_addr_out;
480 sg_dma_len(&sg[1]) = length;
481
482 in_desc = dmaengine_prep_slave_sg(dd->dma_lch_in.chan, &sg[0],
483 1, DMA_MEM_TO_DEV,
484 DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
485 if (!in_desc)
486 return -EINVAL;
487
488 out_desc = dmaengine_prep_slave_sg(dd->dma_lch_out.chan, &sg[1],
489 1, DMA_DEV_TO_MEM,
490 DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
491 if (!out_desc)
492 return -EINVAL;
493
494 out_desc->callback = atmel_tdes_dma_callback;
495 out_desc->callback_param = dd;
496
497 dmaengine_submit(out_desc);
498 dma_async_issue_pending(dd->dma_lch_out.chan);
499
500 dmaengine_submit(in_desc);
501 dma_async_issue_pending(dd->dma_lch_in.chan);
502
503 return 0;
504}
505
506static int atmel_tdes_crypt_start(struct atmel_tdes_dev *dd)
507{
508 struct crypto_tfm *tfm = crypto_ablkcipher_tfm(
509 crypto_ablkcipher_reqtfm(dd->req));
510 int err, fast = 0, in, out;
511 size_t count;
512 dma_addr_t addr_in, addr_out;
513
514 if ((!dd->in_offset) && (!dd->out_offset)) {
515 /* check for alignment */
516 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) &&
517 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size);
518 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) &&
519 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size);
520 fast = in && out;
521
522 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg))
523 fast = 0;
524 }
525
526
527 if (fast) {
528 count = min(dd->total, sg_dma_len(dd->in_sg));
529 count = min(count, sg_dma_len(dd->out_sg));
530
531 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
532 if (!err) {
533 dev_err(dd->dev, "dma_map_sg() error\n");
534 return -EINVAL;
535 }
536
537 err = dma_map_sg(dd->dev, dd->out_sg, 1,
538 DMA_FROM_DEVICE);
539 if (!err) {
540 dev_err(dd->dev, "dma_map_sg() error\n");
541 dma_unmap_sg(dd->dev, dd->in_sg, 1,
542 DMA_TO_DEVICE);
543 return -EINVAL;
544 }
545
546 addr_in = sg_dma_address(dd->in_sg);
547 addr_out = sg_dma_address(dd->out_sg);
548
549 dd->flags |= TDES_FLAGS_FAST;
550
551 } else {
552 /* use cache buffers */
553 count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset,
554 dd->buf_in, dd->buflen, dd->total, 0);
555
556 addr_in = dd->dma_addr_in;
557 addr_out = dd->dma_addr_out;
558
559 dd->flags &= ~TDES_FLAGS_FAST;
560 }
561
562 dd->total -= count;
563
564 if (dd->caps.has_dma)
565 err = atmel_tdes_crypt_dma(tfm, addr_in, addr_out, count);
566 else
567 err = atmel_tdes_crypt_pdc(tfm, addr_in, addr_out, count);
568
569 if (err && (dd->flags & TDES_FLAGS_FAST)) {
570 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
571 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE);
572 }
573
574 return err;
575}
576
577static void atmel_tdes_finish_req(struct atmel_tdes_dev *dd, int err)
578{
579 struct ablkcipher_request *req = dd->req;
580
581 clk_disable_unprepare(dd->iclk);
582
583 dd->flags &= ~TDES_FLAGS_BUSY;
584
585 req->base.complete(&req->base, err);
586}
587
588static int atmel_tdes_handle_queue(struct atmel_tdes_dev *dd,
589 struct ablkcipher_request *req)
590{
591 struct crypto_async_request *async_req, *backlog;
592 struct atmel_tdes_ctx *ctx;
593 struct atmel_tdes_reqctx *rctx;
594 unsigned long flags;
595 int err, ret = 0;
596
597 spin_lock_irqsave(&dd->lock, flags);
598 if (req)
599 ret = ablkcipher_enqueue_request(&dd->queue, req);
600 if (dd->flags & TDES_FLAGS_BUSY) {
601 spin_unlock_irqrestore(&dd->lock, flags);
602 return ret;
603 }
604 backlog = crypto_get_backlog(&dd->queue);
605 async_req = crypto_dequeue_request(&dd->queue);
606 if (async_req)
607 dd->flags |= TDES_FLAGS_BUSY;
608 spin_unlock_irqrestore(&dd->lock, flags);
609
610 if (!async_req)
611 return ret;
612
613 if (backlog)
614 backlog->complete(backlog, -EINPROGRESS);
615
616 req = ablkcipher_request_cast(async_req);
617
618 /* assign new request to device */
619 dd->req = req;
620 dd->total = req->nbytes;
621 dd->in_offset = 0;
622 dd->in_sg = req->src;
623 dd->out_offset = 0;
624 dd->out_sg = req->dst;
625
626 rctx = ablkcipher_request_ctx(req);
627 ctx = crypto_ablkcipher_ctx(crypto_ablkcipher_reqtfm(req));
628 rctx->mode &= TDES_FLAGS_MODE_MASK;
629 dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode;
630 dd->ctx = ctx;
631 ctx->dd = dd;
632
633 err = atmel_tdes_write_ctrl(dd);
634 if (!err)
635 err = atmel_tdes_crypt_start(dd);
636 if (err) {
637 /* des_task will not finish it, so do it here */
638 atmel_tdes_finish_req(dd, err);
639 tasklet_schedule(&dd->queue_task);
640 }
641
642 return ret;
643}
644
645static int atmel_tdes_crypt_dma_stop(struct atmel_tdes_dev *dd)
646{
647 int err = -EINVAL;
648 size_t count;
649
650 if (dd->flags & TDES_FLAGS_DMA) {
651 err = 0;
652 if (dd->flags & TDES_FLAGS_FAST) {
653 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
654 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
655 } else {
656 dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
657 dd->dma_size, DMA_FROM_DEVICE);
658
659 /* copy data */
660 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
661 dd->buf_out, dd->buflen, dd->dma_size, 1);
662 if (count != dd->dma_size) {
663 err = -EINVAL;
664 pr_err("not all data converted: %u\n", count);
665 }
666 }
667 }
668 return err;
669}
670
671static int atmel_tdes_crypt(struct ablkcipher_request *req, unsigned long mode)
672{
673 struct atmel_tdes_ctx *ctx = crypto_ablkcipher_ctx(
674 crypto_ablkcipher_reqtfm(req));
675 struct atmel_tdes_reqctx *rctx = ablkcipher_request_ctx(req);
676
677 if (mode & TDES_FLAGS_CFB8) {
678 if (!IS_ALIGNED(req->nbytes, CFB8_BLOCK_SIZE)) {
679 pr_err("request size is not exact amount of CFB8 blocks\n");
680 return -EINVAL;
681 }
682 ctx->block_size = CFB8_BLOCK_SIZE;
683 } else if (mode & TDES_FLAGS_CFB16) {
684 if (!IS_ALIGNED(req->nbytes, CFB16_BLOCK_SIZE)) {
685 pr_err("request size is not exact amount of CFB16 blocks\n");
686 return -EINVAL;
687 }
688 ctx->block_size = CFB16_BLOCK_SIZE;
689 } else if (mode & TDES_FLAGS_CFB32) {
690 if (!IS_ALIGNED(req->nbytes, CFB32_BLOCK_SIZE)) {
691 pr_err("request size is not exact amount of CFB32 blocks\n");
692 return -EINVAL;
693 }
694 ctx->block_size = CFB32_BLOCK_SIZE;
695 } else {
696 if (!IS_ALIGNED(req->nbytes, DES_BLOCK_SIZE)) {
697 pr_err("request size is not exact amount of DES blocks\n");
698 return -EINVAL;
699 }
700 ctx->block_size = DES_BLOCK_SIZE;
701 }
702
703 rctx->mode = mode;
704
705 return atmel_tdes_handle_queue(ctx->dd, req);
706}
707
708static bool atmel_tdes_filter(struct dma_chan *chan, void *slave)
709{
710 struct at_dma_slave *sl = slave;
711
712 if (sl && sl->dma_dev == chan->device->dev) {
713 chan->private = sl;
714 return true;
715 } else {
716 return false;
717 }
718}
719
720static int atmel_tdes_dma_init(struct atmel_tdes_dev *dd,
721 struct crypto_platform_data *pdata)
722{
723 int err = -ENOMEM;
724 dma_cap_mask_t mask;
725
726 dma_cap_zero(mask);
727 dma_cap_set(DMA_SLAVE, mask);
728
729 /* Try to grab 2 DMA channels */
730 dd->dma_lch_in.chan = dma_request_slave_channel_compat(mask,
731 atmel_tdes_filter, &pdata->dma_slave->rxdata, dd->dev, "tx");
732 if (!dd->dma_lch_in.chan)
733 goto err_dma_in;
734
735 dd->dma_lch_in.dma_conf.direction = DMA_MEM_TO_DEV;
736 dd->dma_lch_in.dma_conf.dst_addr = dd->phys_base +
737 TDES_IDATA1R;
738 dd->dma_lch_in.dma_conf.src_maxburst = 1;
739 dd->dma_lch_in.dma_conf.src_addr_width =
740 DMA_SLAVE_BUSWIDTH_4_BYTES;
741 dd->dma_lch_in.dma_conf.dst_maxburst = 1;
742 dd->dma_lch_in.dma_conf.dst_addr_width =
743 DMA_SLAVE_BUSWIDTH_4_BYTES;
744 dd->dma_lch_in.dma_conf.device_fc = false;
745
746 dd->dma_lch_out.chan = dma_request_slave_channel_compat(mask,
747 atmel_tdes_filter, &pdata->dma_slave->txdata, dd->dev, "rx");
748 if (!dd->dma_lch_out.chan)
749 goto err_dma_out;
750
751 dd->dma_lch_out.dma_conf.direction = DMA_DEV_TO_MEM;
752 dd->dma_lch_out.dma_conf.src_addr = dd->phys_base +
753 TDES_ODATA1R;
754 dd->dma_lch_out.dma_conf.src_maxburst = 1;
755 dd->dma_lch_out.dma_conf.src_addr_width =
756 DMA_SLAVE_BUSWIDTH_4_BYTES;
757 dd->dma_lch_out.dma_conf.dst_maxburst = 1;
758 dd->dma_lch_out.dma_conf.dst_addr_width =
759 DMA_SLAVE_BUSWIDTH_4_BYTES;
760 dd->dma_lch_out.dma_conf.device_fc = false;
761
762 return 0;
763
764err_dma_out:
765 dma_release_channel(dd->dma_lch_in.chan);
766err_dma_in:
767 dev_warn(dd->dev, "no DMA channel available\n");
768 return err;
769}
770
771static void atmel_tdes_dma_cleanup(struct atmel_tdes_dev *dd)
772{
773 dma_release_channel(dd->dma_lch_in.chan);
774 dma_release_channel(dd->dma_lch_out.chan);
775}
776
777static int atmel_des_setkey(struct crypto_ablkcipher *tfm, const u8 *key,
778 unsigned int keylen)
779{
780 u32 tmp[DES_EXPKEY_WORDS];
781 int err;
782 struct crypto_tfm *ctfm = crypto_ablkcipher_tfm(tfm);
783
784 struct atmel_tdes_ctx *ctx = crypto_ablkcipher_ctx(tfm);
785
786 if (keylen != DES_KEY_SIZE) {
787 crypto_ablkcipher_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
788 return -EINVAL;
789 }
790
791 err = des_ekey(tmp, key);
792 if (err == 0 && (ctfm->crt_flags & CRYPTO_TFM_REQ_WEAK_KEY)) {
793 ctfm->crt_flags |= CRYPTO_TFM_RES_WEAK_KEY;
794 return -EINVAL;
795 }
796
797 memcpy(ctx->key, key, keylen);
798 ctx->keylen = keylen;
799
800 return 0;
801}
802
803static int atmel_tdes_setkey(struct crypto_ablkcipher *tfm, const u8 *key,
804 unsigned int keylen)
805{
806 struct atmel_tdes_ctx *ctx = crypto_ablkcipher_ctx(tfm);
807 const char *alg_name;
808
809 alg_name = crypto_tfm_alg_name(crypto_ablkcipher_tfm(tfm));
810
811 /*
812 * HW bug in cfb 3-keys mode.
813 */
814 if (!ctx->dd->caps.has_cfb_3keys && strstr(alg_name, "cfb")
815 && (keylen != 2*DES_KEY_SIZE)) {
816 crypto_ablkcipher_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
817 return -EINVAL;
818 } else if ((keylen != 2*DES_KEY_SIZE) && (keylen != 3*DES_KEY_SIZE)) {
819 crypto_ablkcipher_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
820 return -EINVAL;
821 }
822
823 memcpy(ctx->key, key, keylen);
824 ctx->keylen = keylen;
825
826 return 0;
827}
828
829static int atmel_tdes_ecb_encrypt(struct ablkcipher_request *req)
830{
831 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT);
832}
833
834static int atmel_tdes_ecb_decrypt(struct ablkcipher_request *req)
835{
836 return atmel_tdes_crypt(req, 0);
837}
838
839static int atmel_tdes_cbc_encrypt(struct ablkcipher_request *req)
840{
841 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT | TDES_FLAGS_CBC);
842}
843
844static int atmel_tdes_cbc_decrypt(struct ablkcipher_request *req)
845{
846 return atmel_tdes_crypt(req, TDES_FLAGS_CBC);
847}
848static int atmel_tdes_cfb_encrypt(struct ablkcipher_request *req)
849{
850 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT | TDES_FLAGS_CFB);
851}
852
853static int atmel_tdes_cfb_decrypt(struct ablkcipher_request *req)
854{
855 return atmel_tdes_crypt(req, TDES_FLAGS_CFB);
856}
857
858static int atmel_tdes_cfb8_encrypt(struct ablkcipher_request *req)
859{
860 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT | TDES_FLAGS_CFB |
861 TDES_FLAGS_CFB8);
862}
863
864static int atmel_tdes_cfb8_decrypt(struct ablkcipher_request *req)
865{
866 return atmel_tdes_crypt(req, TDES_FLAGS_CFB | TDES_FLAGS_CFB8);
867}
868
869static int atmel_tdes_cfb16_encrypt(struct ablkcipher_request *req)
870{
871 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT | TDES_FLAGS_CFB |
872 TDES_FLAGS_CFB16);
873}
874
875static int atmel_tdes_cfb16_decrypt(struct ablkcipher_request *req)
876{
877 return atmel_tdes_crypt(req, TDES_FLAGS_CFB | TDES_FLAGS_CFB16);
878}
879
880static int atmel_tdes_cfb32_encrypt(struct ablkcipher_request *req)
881{
882 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT | TDES_FLAGS_CFB |
883 TDES_FLAGS_CFB32);
884}
885
886static int atmel_tdes_cfb32_decrypt(struct ablkcipher_request *req)
887{
888 return atmel_tdes_crypt(req, TDES_FLAGS_CFB | TDES_FLAGS_CFB32);
889}
890
891static int atmel_tdes_ofb_encrypt(struct ablkcipher_request *req)
892{
893 return atmel_tdes_crypt(req, TDES_FLAGS_ENCRYPT | TDES_FLAGS_OFB);
894}
895
896static int atmel_tdes_ofb_decrypt(struct ablkcipher_request *req)
897{
898 return atmel_tdes_crypt(req, TDES_FLAGS_OFB);
899}
900
901static int atmel_tdes_cra_init(struct crypto_tfm *tfm)
902{
903 struct atmel_tdes_ctx *ctx = crypto_tfm_ctx(tfm);
904 struct atmel_tdes_dev *dd;
905
906 tfm->crt_ablkcipher.reqsize = sizeof(struct atmel_tdes_reqctx);
907
908 dd = atmel_tdes_find_dev(ctx);
909 if (!dd)
910 return -ENODEV;
911
912 return 0;
913}
914
915static void atmel_tdes_cra_exit(struct crypto_tfm *tfm)
916{
917}
918
919static struct crypto_alg tdes_algs[] = {
920{
921 .cra_name = "ecb(des)",
922 .cra_driver_name = "atmel-ecb-des",
923 .cra_priority = 100,
924 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
925 .cra_blocksize = DES_BLOCK_SIZE,
926 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
927 .cra_alignmask = 0x7,
928 .cra_type = &crypto_ablkcipher_type,
929 .cra_module = THIS_MODULE,
930 .cra_init = atmel_tdes_cra_init,
931 .cra_exit = atmel_tdes_cra_exit,
932 .cra_u.ablkcipher = {
933 .min_keysize = DES_KEY_SIZE,
934 .max_keysize = DES_KEY_SIZE,
935 .setkey = atmel_des_setkey,
936 .encrypt = atmel_tdes_ecb_encrypt,
937 .decrypt = atmel_tdes_ecb_decrypt,
938 }
939},
940{
941 .cra_name = "cbc(des)",
942 .cra_driver_name = "atmel-cbc-des",
943 .cra_priority = 100,
944 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
945 .cra_blocksize = DES_BLOCK_SIZE,
946 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
947 .cra_alignmask = 0x7,
948 .cra_type = &crypto_ablkcipher_type,
949 .cra_module = THIS_MODULE,
950 .cra_init = atmel_tdes_cra_init,
951 .cra_exit = atmel_tdes_cra_exit,
952 .cra_u.ablkcipher = {
953 .min_keysize = DES_KEY_SIZE,
954 .max_keysize = DES_KEY_SIZE,
955 .ivsize = DES_BLOCK_SIZE,
956 .setkey = atmel_des_setkey,
957 .encrypt = atmel_tdes_cbc_encrypt,
958 .decrypt = atmel_tdes_cbc_decrypt,
959 }
960},
961{
962 .cra_name = "cfb(des)",
963 .cra_driver_name = "atmel-cfb-des",
964 .cra_priority = 100,
965 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
966 .cra_blocksize = DES_BLOCK_SIZE,
967 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
968 .cra_alignmask = 0x7,
969 .cra_type = &crypto_ablkcipher_type,
970 .cra_module = THIS_MODULE,
971 .cra_init = atmel_tdes_cra_init,
972 .cra_exit = atmel_tdes_cra_exit,
973 .cra_u.ablkcipher = {
974 .min_keysize = DES_KEY_SIZE,
975 .max_keysize = DES_KEY_SIZE,
976 .ivsize = DES_BLOCK_SIZE,
977 .setkey = atmel_des_setkey,
978 .encrypt = atmel_tdes_cfb_encrypt,
979 .decrypt = atmel_tdes_cfb_decrypt,
980 }
981},
982{
983 .cra_name = "cfb8(des)",
984 .cra_driver_name = "atmel-cfb8-des",
985 .cra_priority = 100,
986 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
987 .cra_blocksize = CFB8_BLOCK_SIZE,
988 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
989 .cra_alignmask = 0,
990 .cra_type = &crypto_ablkcipher_type,
991 .cra_module = THIS_MODULE,
992 .cra_init = atmel_tdes_cra_init,
993 .cra_exit = atmel_tdes_cra_exit,
994 .cra_u.ablkcipher = {
995 .min_keysize = DES_KEY_SIZE,
996 .max_keysize = DES_KEY_SIZE,
997 .ivsize = DES_BLOCK_SIZE,
998 .setkey = atmel_des_setkey,
999 .encrypt = atmel_tdes_cfb8_encrypt,
1000 .decrypt = atmel_tdes_cfb8_decrypt,
1001 }
1002},
1003{
1004 .cra_name = "cfb16(des)",
1005 .cra_driver_name = "atmel-cfb16-des",
1006 .cra_priority = 100,
1007 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1008 .cra_blocksize = CFB16_BLOCK_SIZE,
1009 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1010 .cra_alignmask = 0x1,
1011 .cra_type = &crypto_ablkcipher_type,
1012 .cra_module = THIS_MODULE,
1013 .cra_init = atmel_tdes_cra_init,
1014 .cra_exit = atmel_tdes_cra_exit,
1015 .cra_u.ablkcipher = {
1016 .min_keysize = DES_KEY_SIZE,
1017 .max_keysize = DES_KEY_SIZE,
1018 .ivsize = DES_BLOCK_SIZE,
1019 .setkey = atmel_des_setkey,
1020 .encrypt = atmel_tdes_cfb16_encrypt,
1021 .decrypt = atmel_tdes_cfb16_decrypt,
1022 }
1023},
1024{
1025 .cra_name = "cfb32(des)",
1026 .cra_driver_name = "atmel-cfb32-des",
1027 .cra_priority = 100,
1028 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1029 .cra_blocksize = CFB32_BLOCK_SIZE,
1030 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1031 .cra_alignmask = 0x3,
1032 .cra_type = &crypto_ablkcipher_type,
1033 .cra_module = THIS_MODULE,
1034 .cra_init = atmel_tdes_cra_init,
1035 .cra_exit = atmel_tdes_cra_exit,
1036 .cra_u.ablkcipher = {
1037 .min_keysize = DES_KEY_SIZE,
1038 .max_keysize = DES_KEY_SIZE,
1039 .ivsize = DES_BLOCK_SIZE,
1040 .setkey = atmel_des_setkey,
1041 .encrypt = atmel_tdes_cfb32_encrypt,
1042 .decrypt = atmel_tdes_cfb32_decrypt,
1043 }
1044},
1045{
1046 .cra_name = "ofb(des)",
1047 .cra_driver_name = "atmel-ofb-des",
1048 .cra_priority = 100,
1049 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1050 .cra_blocksize = DES_BLOCK_SIZE,
1051 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1052 .cra_alignmask = 0x7,
1053 .cra_type = &crypto_ablkcipher_type,
1054 .cra_module = THIS_MODULE,
1055 .cra_init = atmel_tdes_cra_init,
1056 .cra_exit = atmel_tdes_cra_exit,
1057 .cra_u.ablkcipher = {
1058 .min_keysize = DES_KEY_SIZE,
1059 .max_keysize = DES_KEY_SIZE,
1060 .ivsize = DES_BLOCK_SIZE,
1061 .setkey = atmel_des_setkey,
1062 .encrypt = atmel_tdes_ofb_encrypt,
1063 .decrypt = atmel_tdes_ofb_decrypt,
1064 }
1065},
1066{
1067 .cra_name = "ecb(des3_ede)",
1068 .cra_driver_name = "atmel-ecb-tdes",
1069 .cra_priority = 100,
1070 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1071 .cra_blocksize = DES_BLOCK_SIZE,
1072 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1073 .cra_alignmask = 0x7,
1074 .cra_type = &crypto_ablkcipher_type,
1075 .cra_module = THIS_MODULE,
1076 .cra_init = atmel_tdes_cra_init,
1077 .cra_exit = atmel_tdes_cra_exit,
1078 .cra_u.ablkcipher = {
1079 .min_keysize = 2 * DES_KEY_SIZE,
1080 .max_keysize = 3 * DES_KEY_SIZE,
1081 .setkey = atmel_tdes_setkey,
1082 .encrypt = atmel_tdes_ecb_encrypt,
1083 .decrypt = atmel_tdes_ecb_decrypt,
1084 }
1085},
1086{
1087 .cra_name = "cbc(des3_ede)",
1088 .cra_driver_name = "atmel-cbc-tdes",
1089 .cra_priority = 100,
1090 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1091 .cra_blocksize = DES_BLOCK_SIZE,
1092 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1093 .cra_alignmask = 0x7,
1094 .cra_type = &crypto_ablkcipher_type,
1095 .cra_module = THIS_MODULE,
1096 .cra_init = atmel_tdes_cra_init,
1097 .cra_exit = atmel_tdes_cra_exit,
1098 .cra_u.ablkcipher = {
1099 .min_keysize = 2*DES_KEY_SIZE,
1100 .max_keysize = 3*DES_KEY_SIZE,
1101 .ivsize = DES_BLOCK_SIZE,
1102 .setkey = atmel_tdes_setkey,
1103 .encrypt = atmel_tdes_cbc_encrypt,
1104 .decrypt = atmel_tdes_cbc_decrypt,
1105 }
1106},
1107{
1108 .cra_name = "cfb(des3_ede)",
1109 .cra_driver_name = "atmel-cfb-tdes",
1110 .cra_priority = 100,
1111 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1112 .cra_blocksize = DES_BLOCK_SIZE,
1113 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1114 .cra_alignmask = 0x7,
1115 .cra_type = &crypto_ablkcipher_type,
1116 .cra_module = THIS_MODULE,
1117 .cra_init = atmel_tdes_cra_init,
1118 .cra_exit = atmel_tdes_cra_exit,
1119 .cra_u.ablkcipher = {
1120 .min_keysize = 2*DES_KEY_SIZE,
1121 .max_keysize = 2*DES_KEY_SIZE,
1122 .ivsize = DES_BLOCK_SIZE,
1123 .setkey = atmel_tdes_setkey,
1124 .encrypt = atmel_tdes_cfb_encrypt,
1125 .decrypt = atmel_tdes_cfb_decrypt,
1126 }
1127},
1128{
1129 .cra_name = "cfb8(des3_ede)",
1130 .cra_driver_name = "atmel-cfb8-tdes",
1131 .cra_priority = 100,
1132 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1133 .cra_blocksize = CFB8_BLOCK_SIZE,
1134 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1135 .cra_alignmask = 0,
1136 .cra_type = &crypto_ablkcipher_type,
1137 .cra_module = THIS_MODULE,
1138 .cra_init = atmel_tdes_cra_init,
1139 .cra_exit = atmel_tdes_cra_exit,
1140 .cra_u.ablkcipher = {
1141 .min_keysize = 2*DES_KEY_SIZE,
1142 .max_keysize = 2*DES_KEY_SIZE,
1143 .ivsize = DES_BLOCK_SIZE,
1144 .setkey = atmel_tdes_setkey,
1145 .encrypt = atmel_tdes_cfb8_encrypt,
1146 .decrypt = atmel_tdes_cfb8_decrypt,
1147 }
1148},
1149{
1150 .cra_name = "cfb16(des3_ede)",
1151 .cra_driver_name = "atmel-cfb16-tdes",
1152 .cra_priority = 100,
1153 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1154 .cra_blocksize = CFB16_BLOCK_SIZE,
1155 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1156 .cra_alignmask = 0x1,
1157 .cra_type = &crypto_ablkcipher_type,
1158 .cra_module = THIS_MODULE,
1159 .cra_init = atmel_tdes_cra_init,
1160 .cra_exit = atmel_tdes_cra_exit,
1161 .cra_u.ablkcipher = {
1162 .min_keysize = 2*DES_KEY_SIZE,
1163 .max_keysize = 2*DES_KEY_SIZE,
1164 .ivsize = DES_BLOCK_SIZE,
1165 .setkey = atmel_tdes_setkey,
1166 .encrypt = atmel_tdes_cfb16_encrypt,
1167 .decrypt = atmel_tdes_cfb16_decrypt,
1168 }
1169},
1170{
1171 .cra_name = "cfb32(des3_ede)",
1172 .cra_driver_name = "atmel-cfb32-tdes",
1173 .cra_priority = 100,
1174 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1175 .cra_blocksize = CFB32_BLOCK_SIZE,
1176 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1177 .cra_alignmask = 0x3,
1178 .cra_type = &crypto_ablkcipher_type,
1179 .cra_module = THIS_MODULE,
1180 .cra_init = atmel_tdes_cra_init,
1181 .cra_exit = atmel_tdes_cra_exit,
1182 .cra_u.ablkcipher = {
1183 .min_keysize = 2*DES_KEY_SIZE,
1184 .max_keysize = 2*DES_KEY_SIZE,
1185 .ivsize = DES_BLOCK_SIZE,
1186 .setkey = atmel_tdes_setkey,
1187 .encrypt = atmel_tdes_cfb32_encrypt,
1188 .decrypt = atmel_tdes_cfb32_decrypt,
1189 }
1190},
1191{
1192 .cra_name = "ofb(des3_ede)",
1193 .cra_driver_name = "atmel-ofb-tdes",
1194 .cra_priority = 100,
1195 .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
1196 .cra_blocksize = DES_BLOCK_SIZE,
1197 .cra_ctxsize = sizeof(struct atmel_tdes_ctx),
1198 .cra_alignmask = 0x7,
1199 .cra_type = &crypto_ablkcipher_type,
1200 .cra_module = THIS_MODULE,
1201 .cra_init = atmel_tdes_cra_init,
1202 .cra_exit = atmel_tdes_cra_exit,
1203 .cra_u.ablkcipher = {
1204 .min_keysize = 2*DES_KEY_SIZE,
1205 .max_keysize = 3*DES_KEY_SIZE,
1206 .ivsize = DES_BLOCK_SIZE,
1207 .setkey = atmel_tdes_setkey,
1208 .encrypt = atmel_tdes_ofb_encrypt,
1209 .decrypt = atmel_tdes_ofb_decrypt,
1210 }
1211},
1212};
1213
1214static void atmel_tdes_queue_task(unsigned long data)
1215{
1216 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *)data;
1217
1218 atmel_tdes_handle_queue(dd, NULL);
1219}
1220
1221static void atmel_tdes_done_task(unsigned long data)
1222{
1223 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *) data;
1224 int err;
1225
1226 if (!(dd->flags & TDES_FLAGS_DMA))
1227 err = atmel_tdes_crypt_pdc_stop(dd);
1228 else
1229 err = atmel_tdes_crypt_dma_stop(dd);
1230
1231 err = dd->err ? : err;
1232
1233 if (dd->total && !err) {
1234 if (dd->flags & TDES_FLAGS_FAST) {
1235 dd->in_sg = sg_next(dd->in_sg);
1236 dd->out_sg = sg_next(dd->out_sg);
1237 if (!dd->in_sg || !dd->out_sg)
1238 err = -EINVAL;
1239 }
1240 if (!err)
1241 err = atmel_tdes_crypt_start(dd);
1242 if (!err)
1243 return; /* DMA started. Not fininishing. */
1244 }
1245
1246 atmel_tdes_finish_req(dd, err);
1247 atmel_tdes_handle_queue(dd, NULL);
1248}
1249
1250static irqreturn_t atmel_tdes_irq(int irq, void *dev_id)
1251{
1252 struct atmel_tdes_dev *tdes_dd = dev_id;
1253 u32 reg;
1254
1255 reg = atmel_tdes_read(tdes_dd, TDES_ISR);
1256 if (reg & atmel_tdes_read(tdes_dd, TDES_IMR)) {
1257 atmel_tdes_write(tdes_dd, TDES_IDR, reg);
1258 if (TDES_FLAGS_BUSY & tdes_dd->flags)
1259 tasklet_schedule(&tdes_dd->done_task);
1260 else
1261 dev_warn(tdes_dd->dev, "TDES interrupt when no active requests.\n");
1262 return IRQ_HANDLED;
1263 }
1264
1265 return IRQ_NONE;
1266}
1267
1268static void atmel_tdes_unregister_algs(struct atmel_tdes_dev *dd)
1269{
1270 int i;
1271
1272 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++)
1273 crypto_unregister_alg(&tdes_algs[i]);
1274}
1275
1276static int atmel_tdes_register_algs(struct atmel_tdes_dev *dd)
1277{
1278 int err, i, j;
1279
1280 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) {
1281 err = crypto_register_alg(&tdes_algs[i]);
1282 if (err)
1283 goto err_tdes_algs;
1284 }
1285
1286 return 0;
1287
1288err_tdes_algs:
1289 for (j = 0; j < i; j++)
1290 crypto_unregister_alg(&tdes_algs[j]);
1291
1292 return err;
1293}
1294
1295static void atmel_tdes_get_cap(struct atmel_tdes_dev *dd)
1296{
1297
1298 dd->caps.has_dma = 0;
1299 dd->caps.has_cfb_3keys = 0;
1300
1301 /* keep only major version number */
1302 switch (dd->hw_version & 0xf00) {
1303 case 0x700:
1304 dd->caps.has_dma = 1;
1305 dd->caps.has_cfb_3keys = 1;
1306 break;
1307 case 0x600:
1308 break;
1309 default:
1310 dev_warn(dd->dev,
1311 "Unmanaged tdes version, set minimum capabilities\n");
1312 break;
1313 }
1314}
1315
1316#if defined(CONFIG_OF)
1317static const struct of_device_id atmel_tdes_dt_ids[] = {
1318 { .compatible = "atmel,at91sam9g46-tdes" },
1319 { /* sentinel */ }
1320};
1321MODULE_DEVICE_TABLE(of, atmel_tdes_dt_ids);
1322
1323static struct crypto_platform_data *atmel_tdes_of_init(struct platform_device *pdev)
1324{
1325 struct device_node *np = pdev->dev.of_node;
1326 struct crypto_platform_data *pdata;
1327
1328 if (!np) {
1329 dev_err(&pdev->dev, "device node not found\n");
1330 return ERR_PTR(-EINVAL);
1331 }
1332
1333 pdata = devm_kzalloc(&pdev->dev, sizeof(*pdata), GFP_KERNEL);
1334 if (!pdata) {
1335 dev_err(&pdev->dev, "could not allocate memory for pdata\n");
1336 return ERR_PTR(-ENOMEM);
1337 }
1338
1339 pdata->dma_slave = devm_kzalloc(&pdev->dev,
1340 sizeof(*(pdata->dma_slave)),
1341 GFP_KERNEL);
1342 if (!pdata->dma_slave) {
1343 dev_err(&pdev->dev, "could not allocate memory for dma_slave\n");
1344 return ERR_PTR(-ENOMEM);
1345 }
1346
1347 return pdata;
1348}
1349#else /* CONFIG_OF */
1350static inline struct crypto_platform_data *atmel_tdes_of_init(struct platform_device *pdev)
1351{
1352 return ERR_PTR(-EINVAL);
1353}
1354#endif
1355
1356static int atmel_tdes_probe(struct platform_device *pdev)
1357{
1358 struct atmel_tdes_dev *tdes_dd;
1359 struct crypto_platform_data *pdata;
1360 struct device *dev = &pdev->dev;
1361 struct resource *tdes_res;
1362 int err;
1363
1364 tdes_dd = devm_kmalloc(&pdev->dev, sizeof(*tdes_dd), GFP_KERNEL);
1365 if (tdes_dd == NULL) {
1366 dev_err(dev, "unable to alloc data struct.\n");
1367 err = -ENOMEM;
1368 goto tdes_dd_err;
1369 }
1370
1371 tdes_dd->dev = dev;
1372
1373 platform_set_drvdata(pdev, tdes_dd);
1374
1375 INIT_LIST_HEAD(&tdes_dd->list);
1376 spin_lock_init(&tdes_dd->lock);
1377
1378 tasklet_init(&tdes_dd->done_task, atmel_tdes_done_task,
1379 (unsigned long)tdes_dd);
1380 tasklet_init(&tdes_dd->queue_task, atmel_tdes_queue_task,
1381 (unsigned long)tdes_dd);
1382
1383 crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH);
1384
1385 tdes_dd->irq = -1;
1386
1387 /* Get the base address */
1388 tdes_res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1389 if (!tdes_res) {
1390 dev_err(dev, "no MEM resource info\n");
1391 err = -ENODEV;
1392 goto res_err;
1393 }
1394 tdes_dd->phys_base = tdes_res->start;
1395
1396 /* Get the IRQ */
1397 tdes_dd->irq = platform_get_irq(pdev, 0);
1398 if (tdes_dd->irq < 0) {
1399 dev_err(dev, "no IRQ resource info\n");
1400 err = tdes_dd->irq;
1401 goto res_err;
1402 }
1403
1404 err = devm_request_irq(&pdev->dev, tdes_dd->irq, atmel_tdes_irq,
1405 IRQF_SHARED, "atmel-tdes", tdes_dd);
1406 if (err) {
1407 dev_err(dev, "unable to request tdes irq.\n");
1408 goto res_err;
1409 }
1410
1411 /* Initializing the clock */
1412 tdes_dd->iclk = devm_clk_get(&pdev->dev, "tdes_clk");
1413 if (IS_ERR(tdes_dd->iclk)) {
1414 dev_err(dev, "clock initialization failed.\n");
1415 err = PTR_ERR(tdes_dd->iclk);
1416 goto res_err;
1417 }
1418
1419 tdes_dd->io_base = devm_ioremap_resource(&pdev->dev, tdes_res);
1420 if (IS_ERR(tdes_dd->io_base)) {
1421 dev_err(dev, "can't ioremap\n");
1422 err = PTR_ERR(tdes_dd->io_base);
1423 goto res_err;
1424 }
1425
1426 atmel_tdes_hw_version_init(tdes_dd);
1427
1428 atmel_tdes_get_cap(tdes_dd);
1429
1430 err = atmel_tdes_buff_init(tdes_dd);
1431 if (err)
1432 goto err_tdes_buff;
1433
1434 if (tdes_dd->caps.has_dma) {
1435 pdata = pdev->dev.platform_data;
1436 if (!pdata) {
1437 pdata = atmel_tdes_of_init(pdev);
1438 if (IS_ERR(pdata)) {
1439 dev_err(&pdev->dev, "platform data not available\n");
1440 err = PTR_ERR(pdata);
1441 goto err_pdata;
1442 }
1443 }
1444 if (!pdata->dma_slave) {
1445 err = -ENXIO;
1446 goto err_pdata;
1447 }
1448 err = atmel_tdes_dma_init(tdes_dd, pdata);
1449 if (err)
1450 goto err_tdes_dma;
1451
1452 dev_info(dev, "using %s, %s for DMA transfers\n",
1453 dma_chan_name(tdes_dd->dma_lch_in.chan),
1454 dma_chan_name(tdes_dd->dma_lch_out.chan));
1455 }
1456
1457 spin_lock(&atmel_tdes.lock);
1458 list_add_tail(&tdes_dd->list, &atmel_tdes.dev_list);
1459 spin_unlock(&atmel_tdes.lock);
1460
1461 err = atmel_tdes_register_algs(tdes_dd);
1462 if (err)
1463 goto err_algs;
1464
1465 dev_info(dev, "Atmel DES/TDES\n");
1466
1467 return 0;
1468
1469err_algs:
1470 spin_lock(&atmel_tdes.lock);
1471 list_del(&tdes_dd->list);
1472 spin_unlock(&atmel_tdes.lock);
1473 if (tdes_dd->caps.has_dma)
1474 atmel_tdes_dma_cleanup(tdes_dd);
1475err_tdes_dma:
1476err_pdata:
1477 atmel_tdes_buff_cleanup(tdes_dd);
1478err_tdes_buff:
1479res_err:
1480 tasklet_kill(&tdes_dd->done_task);
1481 tasklet_kill(&tdes_dd->queue_task);
1482tdes_dd_err:
1483 dev_err(dev, "initialization failed.\n");
1484
1485 return err;
1486}
1487
1488static int atmel_tdes_remove(struct platform_device *pdev)
1489{
1490 static struct atmel_tdes_dev *tdes_dd;
1491
1492 tdes_dd = platform_get_drvdata(pdev);
1493 if (!tdes_dd)
1494 return -ENODEV;
1495 spin_lock(&atmel_tdes.lock);
1496 list_del(&tdes_dd->list);
1497 spin_unlock(&atmel_tdes.lock);
1498
1499 atmel_tdes_unregister_algs(tdes_dd);
1500
1501 tasklet_kill(&tdes_dd->done_task);
1502 tasklet_kill(&tdes_dd->queue_task);
1503
1504 if (tdes_dd->caps.has_dma)
1505 atmel_tdes_dma_cleanup(tdes_dd);
1506
1507 atmel_tdes_buff_cleanup(tdes_dd);
1508
1509 return 0;
1510}
1511
1512static struct platform_driver atmel_tdes_driver = {
1513 .probe = atmel_tdes_probe,
1514 .remove = atmel_tdes_remove,
1515 .driver = {
1516 .name = "atmel_tdes",
1517 .of_match_table = of_match_ptr(atmel_tdes_dt_ids),
1518 },
1519};
1520
1521module_platform_driver(atmel_tdes_driver);
1522
1523MODULE_DESCRIPTION("Atmel DES/TDES hw acceleration support.");
1524MODULE_LICENSE("GPL v2");
1525MODULE_AUTHOR("Nicolas Royer - Eukréa Electromatique");