Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * SPI-Engine SPI controller driver
4 * Copyright 2015 Analog Devices Inc.
5 * Author: Lars-Peter Clausen <lars@metafoo.de>
6 */
7
8#include <linux/clk.h>
9#include <linux/idr.h>
10#include <linux/interrupt.h>
11#include <linux/io.h>
12#include <linux/of.h>
13#include <linux/module.h>
14#include <linux/platform_device.h>
15#include <linux/spi/spi.h>
16#include <linux/timer.h>
17
18#define SPI_ENGINE_VERSION_MAJOR(x) ((x >> 16) & 0xff)
19#define SPI_ENGINE_VERSION_MINOR(x) ((x >> 8) & 0xff)
20#define SPI_ENGINE_VERSION_PATCH(x) (x & 0xff)
21
22#define SPI_ENGINE_REG_VERSION 0x00
23
24#define SPI_ENGINE_REG_RESET 0x40
25
26#define SPI_ENGINE_REG_INT_ENABLE 0x80
27#define SPI_ENGINE_REG_INT_PENDING 0x84
28#define SPI_ENGINE_REG_INT_SOURCE 0x88
29
30#define SPI_ENGINE_REG_SYNC_ID 0xc0
31
32#define SPI_ENGINE_REG_CMD_FIFO_ROOM 0xd0
33#define SPI_ENGINE_REG_SDO_FIFO_ROOM 0xd4
34#define SPI_ENGINE_REG_SDI_FIFO_LEVEL 0xd8
35
36#define SPI_ENGINE_REG_CMD_FIFO 0xe0
37#define SPI_ENGINE_REG_SDO_DATA_FIFO 0xe4
38#define SPI_ENGINE_REG_SDI_DATA_FIFO 0xe8
39#define SPI_ENGINE_REG_SDI_DATA_FIFO_PEEK 0xec
40
41#define SPI_ENGINE_INT_CMD_ALMOST_EMPTY BIT(0)
42#define SPI_ENGINE_INT_SDO_ALMOST_EMPTY BIT(1)
43#define SPI_ENGINE_INT_SDI_ALMOST_FULL BIT(2)
44#define SPI_ENGINE_INT_SYNC BIT(3)
45
46#define SPI_ENGINE_CONFIG_CPHA BIT(0)
47#define SPI_ENGINE_CONFIG_CPOL BIT(1)
48#define SPI_ENGINE_CONFIG_3WIRE BIT(2)
49
50#define SPI_ENGINE_INST_TRANSFER 0x0
51#define SPI_ENGINE_INST_ASSERT 0x1
52#define SPI_ENGINE_INST_WRITE 0x2
53#define SPI_ENGINE_INST_MISC 0x3
54
55#define SPI_ENGINE_CMD_REG_CLK_DIV 0x0
56#define SPI_ENGINE_CMD_REG_CONFIG 0x1
57#define SPI_ENGINE_CMD_REG_XFER_BITS 0x2
58
59#define SPI_ENGINE_MISC_SYNC 0x0
60#define SPI_ENGINE_MISC_SLEEP 0x1
61
62#define SPI_ENGINE_TRANSFER_WRITE 0x1
63#define SPI_ENGINE_TRANSFER_READ 0x2
64
65#define SPI_ENGINE_CMD(inst, arg1, arg2) \
66 (((inst) << 12) | ((arg1) << 8) | (arg2))
67
68#define SPI_ENGINE_CMD_TRANSFER(flags, n) \
69 SPI_ENGINE_CMD(SPI_ENGINE_INST_TRANSFER, (flags), (n))
70#define SPI_ENGINE_CMD_ASSERT(delay, cs) \
71 SPI_ENGINE_CMD(SPI_ENGINE_INST_ASSERT, (delay), (cs))
72#define SPI_ENGINE_CMD_WRITE(reg, val) \
73 SPI_ENGINE_CMD(SPI_ENGINE_INST_WRITE, (reg), (val))
74#define SPI_ENGINE_CMD_SLEEP(delay) \
75 SPI_ENGINE_CMD(SPI_ENGINE_INST_MISC, SPI_ENGINE_MISC_SLEEP, (delay))
76#define SPI_ENGINE_CMD_SYNC(id) \
77 SPI_ENGINE_CMD(SPI_ENGINE_INST_MISC, SPI_ENGINE_MISC_SYNC, (id))
78
79struct spi_engine_program {
80 unsigned int length;
81 uint16_t instructions[];
82};
83
84/**
85 * struct spi_engine_message_state - SPI engine per-message state
86 */
87struct spi_engine_message_state {
88 /** @p: Instructions for executing this message. */
89 struct spi_engine_program *p;
90 /** @cmd_length: Number of elements in cmd_buf array. */
91 unsigned cmd_length;
92 /** @cmd_buf: Array of commands not yet written to CMD FIFO. */
93 const uint16_t *cmd_buf;
94 /** @tx_xfer: Next xfer with tx_buf not yet fully written to TX FIFO. */
95 struct spi_transfer *tx_xfer;
96 /** @tx_length: Size of tx_buf in bytes. */
97 unsigned int tx_length;
98 /** @tx_buf: Bytes not yet written to TX FIFO. */
99 const uint8_t *tx_buf;
100 /** @rx_xfer: Next xfer with rx_buf not yet fully written to RX FIFO. */
101 struct spi_transfer *rx_xfer;
102 /** @rx_length: Size of tx_buf in bytes. */
103 unsigned int rx_length;
104 /** @rx_buf: Bytes not yet written to the RX FIFO. */
105 uint8_t *rx_buf;
106 /** @sync_id: ID to correlate SYNC interrupts with this message. */
107 u8 sync_id;
108};
109
110struct spi_engine {
111 struct clk *clk;
112 struct clk *ref_clk;
113
114 spinlock_t lock;
115
116 void __iomem *base;
117 struct ida sync_ida;
118 struct timer_list watchdog_timer;
119 struct spi_controller *controller;
120
121 unsigned int int_enable;
122};
123
124static void spi_engine_program_add_cmd(struct spi_engine_program *p,
125 bool dry, uint16_t cmd)
126{
127 if (!dry)
128 p->instructions[p->length] = cmd;
129 p->length++;
130}
131
132static unsigned int spi_engine_get_config(struct spi_device *spi)
133{
134 unsigned int config = 0;
135
136 if (spi->mode & SPI_CPOL)
137 config |= SPI_ENGINE_CONFIG_CPOL;
138 if (spi->mode & SPI_CPHA)
139 config |= SPI_ENGINE_CONFIG_CPHA;
140 if (spi->mode & SPI_3WIRE)
141 config |= SPI_ENGINE_CONFIG_3WIRE;
142
143 return config;
144}
145
146static void spi_engine_gen_xfer(struct spi_engine_program *p, bool dry,
147 struct spi_transfer *xfer)
148{
149 unsigned int len;
150
151 if (xfer->bits_per_word <= 8)
152 len = xfer->len;
153 else if (xfer->bits_per_word <= 16)
154 len = xfer->len / 2;
155 else
156 len = xfer->len / 4;
157
158 while (len) {
159 unsigned int n = min(len, 256U);
160 unsigned int flags = 0;
161
162 if (xfer->tx_buf)
163 flags |= SPI_ENGINE_TRANSFER_WRITE;
164 if (xfer->rx_buf)
165 flags |= SPI_ENGINE_TRANSFER_READ;
166
167 spi_engine_program_add_cmd(p, dry,
168 SPI_ENGINE_CMD_TRANSFER(flags, n - 1));
169 len -= n;
170 }
171}
172
173static void spi_engine_gen_sleep(struct spi_engine_program *p, bool dry,
174 int delay_ns, u32 sclk_hz)
175{
176 unsigned int t;
177
178 /* negative delay indicates error, e.g. from spi_delay_to_ns() */
179 if (delay_ns <= 0)
180 return;
181
182 /* rounding down since executing the instruction adds a couple of ticks delay */
183 t = DIV_ROUND_DOWN_ULL((u64)delay_ns * sclk_hz, NSEC_PER_SEC);
184 while (t) {
185 unsigned int n = min(t, 256U);
186
187 spi_engine_program_add_cmd(p, dry, SPI_ENGINE_CMD_SLEEP(n - 1));
188 t -= n;
189 }
190}
191
192static void spi_engine_gen_cs(struct spi_engine_program *p, bool dry,
193 struct spi_device *spi, bool assert)
194{
195 unsigned int mask = 0xff;
196
197 if (assert)
198 mask ^= BIT(spi_get_chipselect(spi, 0));
199
200 spi_engine_program_add_cmd(p, dry, SPI_ENGINE_CMD_ASSERT(0, mask));
201}
202
203/*
204 * Performs precompile steps on the message.
205 *
206 * The SPI core does most of the message/transfer validation and filling in
207 * fields for us via __spi_validate(). This fixes up anything remaining not
208 * done there.
209 *
210 * NB: This is separate from spi_engine_compile_message() because the latter
211 * is called twice and would otherwise result in double-evaluation.
212 */
213static void spi_engine_precompile_message(struct spi_message *msg)
214{
215 unsigned int clk_div, max_hz = msg->spi->controller->max_speed_hz;
216 struct spi_transfer *xfer;
217
218 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
219 clk_div = DIV_ROUND_UP(max_hz, xfer->speed_hz);
220 xfer->effective_speed_hz = max_hz / min(clk_div, 256U);
221 }
222}
223
224static void spi_engine_compile_message(struct spi_message *msg, bool dry,
225 struct spi_engine_program *p)
226{
227 struct spi_device *spi = msg->spi;
228 struct spi_controller *host = spi->controller;
229 struct spi_transfer *xfer;
230 int clk_div, new_clk_div;
231 bool keep_cs = false;
232 u8 bits_per_word = 0;
233
234 clk_div = 1;
235
236 spi_engine_program_add_cmd(p, dry,
237 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CONFIG,
238 spi_engine_get_config(spi)));
239
240 xfer = list_first_entry(&msg->transfers, struct spi_transfer, transfer_list);
241 spi_engine_gen_cs(p, dry, spi, !xfer->cs_off);
242
243 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
244 new_clk_div = host->max_speed_hz / xfer->effective_speed_hz;
245 if (new_clk_div != clk_div) {
246 clk_div = new_clk_div;
247 /* actual divider used is register value + 1 */
248 spi_engine_program_add_cmd(p, dry,
249 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CLK_DIV,
250 clk_div - 1));
251 }
252
253 if (bits_per_word != xfer->bits_per_word) {
254 bits_per_word = xfer->bits_per_word;
255 spi_engine_program_add_cmd(p, dry,
256 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_XFER_BITS,
257 bits_per_word));
258 }
259
260 spi_engine_gen_xfer(p, dry, xfer);
261 spi_engine_gen_sleep(p, dry, spi_delay_to_ns(&xfer->delay, xfer),
262 xfer->effective_speed_hz);
263
264 if (xfer->cs_change) {
265 if (list_is_last(&xfer->transfer_list, &msg->transfers)) {
266 keep_cs = true;
267 } else {
268 if (!xfer->cs_off)
269 spi_engine_gen_cs(p, dry, spi, false);
270
271 spi_engine_gen_sleep(p, dry, spi_delay_to_ns(
272 &xfer->cs_change_delay, xfer),
273 xfer->effective_speed_hz);
274
275 if (!list_next_entry(xfer, transfer_list)->cs_off)
276 spi_engine_gen_cs(p, dry, spi, true);
277 }
278 } else if (!list_is_last(&xfer->transfer_list, &msg->transfers) &&
279 xfer->cs_off != list_next_entry(xfer, transfer_list)->cs_off) {
280 spi_engine_gen_cs(p, dry, spi, xfer->cs_off);
281 }
282 }
283
284 if (!keep_cs)
285 spi_engine_gen_cs(p, dry, spi, false);
286
287 /*
288 * Restore clockdiv to default so that future gen_sleep commands don't
289 * have to be aware of the current register state.
290 */
291 if (clk_div != 1)
292 spi_engine_program_add_cmd(p, dry,
293 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CLK_DIV, 0));
294}
295
296static void spi_engine_xfer_next(struct spi_message *msg,
297 struct spi_transfer **_xfer)
298{
299 struct spi_transfer *xfer = *_xfer;
300
301 if (!xfer) {
302 xfer = list_first_entry(&msg->transfers,
303 struct spi_transfer, transfer_list);
304 } else if (list_is_last(&xfer->transfer_list, &msg->transfers)) {
305 xfer = NULL;
306 } else {
307 xfer = list_next_entry(xfer, transfer_list);
308 }
309
310 *_xfer = xfer;
311}
312
313static void spi_engine_tx_next(struct spi_message *msg)
314{
315 struct spi_engine_message_state *st = msg->state;
316 struct spi_transfer *xfer = st->tx_xfer;
317
318 do {
319 spi_engine_xfer_next(msg, &xfer);
320 } while (xfer && !xfer->tx_buf);
321
322 st->tx_xfer = xfer;
323 if (xfer) {
324 st->tx_length = xfer->len;
325 st->tx_buf = xfer->tx_buf;
326 } else {
327 st->tx_buf = NULL;
328 }
329}
330
331static void spi_engine_rx_next(struct spi_message *msg)
332{
333 struct spi_engine_message_state *st = msg->state;
334 struct spi_transfer *xfer = st->rx_xfer;
335
336 do {
337 spi_engine_xfer_next(msg, &xfer);
338 } while (xfer && !xfer->rx_buf);
339
340 st->rx_xfer = xfer;
341 if (xfer) {
342 st->rx_length = xfer->len;
343 st->rx_buf = xfer->rx_buf;
344 } else {
345 st->rx_buf = NULL;
346 }
347}
348
349static bool spi_engine_write_cmd_fifo(struct spi_engine *spi_engine,
350 struct spi_message *msg)
351{
352 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_CMD_FIFO;
353 struct spi_engine_message_state *st = msg->state;
354 unsigned int n, m, i;
355 const uint16_t *buf;
356
357 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_CMD_FIFO_ROOM);
358 while (n && st->cmd_length) {
359 m = min(n, st->cmd_length);
360 buf = st->cmd_buf;
361 for (i = 0; i < m; i++)
362 writel_relaxed(buf[i], addr);
363 st->cmd_buf += m;
364 st->cmd_length -= m;
365 n -= m;
366 }
367
368 return st->cmd_length != 0;
369}
370
371static bool spi_engine_write_tx_fifo(struct spi_engine *spi_engine,
372 struct spi_message *msg)
373{
374 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_SDO_DATA_FIFO;
375 struct spi_engine_message_state *st = msg->state;
376 unsigned int n, m, i;
377
378 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_SDO_FIFO_ROOM);
379 while (n && st->tx_length) {
380 if (st->tx_xfer->bits_per_word <= 8) {
381 const u8 *buf = st->tx_buf;
382
383 m = min(n, st->tx_length);
384 for (i = 0; i < m; i++)
385 writel_relaxed(buf[i], addr);
386 st->tx_buf += m;
387 st->tx_length -= m;
388 } else if (st->tx_xfer->bits_per_word <= 16) {
389 const u16 *buf = (const u16 *)st->tx_buf;
390
391 m = min(n, st->tx_length / 2);
392 for (i = 0; i < m; i++)
393 writel_relaxed(buf[i], addr);
394 st->tx_buf += m * 2;
395 st->tx_length -= m * 2;
396 } else {
397 const u32 *buf = (const u32 *)st->tx_buf;
398
399 m = min(n, st->tx_length / 4);
400 for (i = 0; i < m; i++)
401 writel_relaxed(buf[i], addr);
402 st->tx_buf += m * 4;
403 st->tx_length -= m * 4;
404 }
405 n -= m;
406 if (st->tx_length == 0)
407 spi_engine_tx_next(msg);
408 }
409
410 return st->tx_length != 0;
411}
412
413static bool spi_engine_read_rx_fifo(struct spi_engine *spi_engine,
414 struct spi_message *msg)
415{
416 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_SDI_DATA_FIFO;
417 struct spi_engine_message_state *st = msg->state;
418 unsigned int n, m, i;
419
420 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_SDI_FIFO_LEVEL);
421 while (n && st->rx_length) {
422 if (st->rx_xfer->bits_per_word <= 8) {
423 u8 *buf = st->rx_buf;
424
425 m = min(n, st->rx_length);
426 for (i = 0; i < m; i++)
427 buf[i] = readl_relaxed(addr);
428 st->rx_buf += m;
429 st->rx_length -= m;
430 } else if (st->rx_xfer->bits_per_word <= 16) {
431 u16 *buf = (u16 *)st->rx_buf;
432
433 m = min(n, st->rx_length / 2);
434 for (i = 0; i < m; i++)
435 buf[i] = readl_relaxed(addr);
436 st->rx_buf += m * 2;
437 st->rx_length -= m * 2;
438 } else {
439 u32 *buf = (u32 *)st->rx_buf;
440
441 m = min(n, st->rx_length / 4);
442 for (i = 0; i < m; i++)
443 buf[i] = readl_relaxed(addr);
444 st->rx_buf += m * 4;
445 st->rx_length -= m * 4;
446 }
447 n -= m;
448 if (st->rx_length == 0)
449 spi_engine_rx_next(msg);
450 }
451
452 return st->rx_length != 0;
453}
454
455static irqreturn_t spi_engine_irq(int irq, void *devid)
456{
457 struct spi_controller *host = devid;
458 struct spi_message *msg = host->cur_msg;
459 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
460 unsigned int disable_int = 0;
461 unsigned int pending;
462 int completed_id = -1;
463
464 pending = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
465
466 if (pending & SPI_ENGINE_INT_SYNC) {
467 writel_relaxed(SPI_ENGINE_INT_SYNC,
468 spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
469 completed_id = readl_relaxed(
470 spi_engine->base + SPI_ENGINE_REG_SYNC_ID);
471 }
472
473 spin_lock(&spi_engine->lock);
474
475 if (pending & SPI_ENGINE_INT_CMD_ALMOST_EMPTY) {
476 if (!spi_engine_write_cmd_fifo(spi_engine, msg))
477 disable_int |= SPI_ENGINE_INT_CMD_ALMOST_EMPTY;
478 }
479
480 if (pending & SPI_ENGINE_INT_SDO_ALMOST_EMPTY) {
481 if (!spi_engine_write_tx_fifo(spi_engine, msg))
482 disable_int |= SPI_ENGINE_INT_SDO_ALMOST_EMPTY;
483 }
484
485 if (pending & (SPI_ENGINE_INT_SDI_ALMOST_FULL | SPI_ENGINE_INT_SYNC)) {
486 if (!spi_engine_read_rx_fifo(spi_engine, msg))
487 disable_int |= SPI_ENGINE_INT_SDI_ALMOST_FULL;
488 }
489
490 if (pending & SPI_ENGINE_INT_SYNC && msg) {
491 struct spi_engine_message_state *st = msg->state;
492
493 if (completed_id == st->sync_id) {
494 if (timer_delete_sync(&spi_engine->watchdog_timer)) {
495 msg->status = 0;
496 msg->actual_length = msg->frame_length;
497 spi_finalize_current_message(host);
498 }
499 disable_int |= SPI_ENGINE_INT_SYNC;
500 }
501 }
502
503 if (disable_int) {
504 spi_engine->int_enable &= ~disable_int;
505 writel_relaxed(spi_engine->int_enable,
506 spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
507 }
508
509 spin_unlock(&spi_engine->lock);
510
511 return IRQ_HANDLED;
512}
513
514static int spi_engine_prepare_message(struct spi_controller *host,
515 struct spi_message *msg)
516{
517 struct spi_engine_program p_dry, *p;
518 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
519 struct spi_engine_message_state *st;
520 size_t size;
521 int ret;
522
523 st = kzalloc(sizeof(*st), GFP_KERNEL);
524 if (!st)
525 return -ENOMEM;
526
527 spi_engine_precompile_message(msg);
528
529 p_dry.length = 0;
530 spi_engine_compile_message(msg, true, &p_dry);
531
532 size = sizeof(*p->instructions) * (p_dry.length + 1);
533 p = kzalloc(sizeof(*p) + size, GFP_KERNEL);
534 if (!p) {
535 kfree(st);
536 return -ENOMEM;
537 }
538
539 ret = ida_alloc_range(&spi_engine->sync_ida, 0, U8_MAX, GFP_KERNEL);
540 if (ret < 0) {
541 kfree(p);
542 kfree(st);
543 return ret;
544 }
545
546 st->sync_id = ret;
547
548 spi_engine_compile_message(msg, false, p);
549
550 spi_engine_program_add_cmd(p, false, SPI_ENGINE_CMD_SYNC(st->sync_id));
551
552 st->p = p;
553 st->cmd_buf = p->instructions;
554 st->cmd_length = p->length;
555 msg->state = st;
556
557 return 0;
558}
559
560static int spi_engine_unprepare_message(struct spi_controller *host,
561 struct spi_message *msg)
562{
563 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
564 struct spi_engine_message_state *st = msg->state;
565
566 ida_free(&spi_engine->sync_ida, st->sync_id);
567 kfree(st->p);
568 kfree(st);
569
570 return 0;
571}
572
573static int spi_engine_transfer_one_message(struct spi_controller *host,
574 struct spi_message *msg)
575{
576 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
577 struct spi_engine_message_state *st = msg->state;
578 unsigned int int_enable = 0;
579 unsigned long flags;
580
581 mod_timer(&spi_engine->watchdog_timer, jiffies + msecs_to_jiffies(5000));
582
583 spin_lock_irqsave(&spi_engine->lock, flags);
584
585 if (spi_engine_write_cmd_fifo(spi_engine, msg))
586 int_enable |= SPI_ENGINE_INT_CMD_ALMOST_EMPTY;
587
588 spi_engine_tx_next(msg);
589 if (spi_engine_write_tx_fifo(spi_engine, msg))
590 int_enable |= SPI_ENGINE_INT_SDO_ALMOST_EMPTY;
591
592 spi_engine_rx_next(msg);
593 if (st->rx_length != 0)
594 int_enable |= SPI_ENGINE_INT_SDI_ALMOST_FULL;
595
596 int_enable |= SPI_ENGINE_INT_SYNC;
597
598 writel_relaxed(int_enable,
599 spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
600 spi_engine->int_enable = int_enable;
601 spin_unlock_irqrestore(&spi_engine->lock, flags);
602
603 return 0;
604}
605
606static void spi_engine_timeout(struct timer_list *timer)
607{
608 struct spi_engine *spi_engine = from_timer(spi_engine, timer, watchdog_timer);
609 struct spi_controller *host = spi_engine->controller;
610
611 if (WARN_ON(!host->cur_msg))
612 return;
613
614 dev_err(&host->dev,
615 "Timeout occurred while waiting for transfer to complete. Hardware is probably broken.\n");
616 host->cur_msg->status = -ETIMEDOUT;
617 spi_finalize_current_message(host);
618}
619
620static void spi_engine_release_hw(void *p)
621{
622 struct spi_engine *spi_engine = p;
623
624 writel_relaxed(0xff, spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
625 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
626 writel_relaxed(0x01, spi_engine->base + SPI_ENGINE_REG_RESET);
627}
628
629static int spi_engine_probe(struct platform_device *pdev)
630{
631 struct spi_engine *spi_engine;
632 struct spi_controller *host;
633 unsigned int version;
634 int irq;
635 int ret;
636
637 irq = platform_get_irq(pdev, 0);
638 if (irq < 0)
639 return irq;
640
641 host = devm_spi_alloc_host(&pdev->dev, sizeof(*spi_engine));
642 if (!host)
643 return -ENOMEM;
644
645 spi_engine = spi_controller_get_devdata(host);
646
647 spin_lock_init(&spi_engine->lock);
648 ida_init(&spi_engine->sync_ida);
649 timer_setup(&spi_engine->watchdog_timer, spi_engine_timeout, TIMER_IRQSAFE);
650 spi_engine->controller = host;
651
652 spi_engine->clk = devm_clk_get_enabled(&pdev->dev, "s_axi_aclk");
653 if (IS_ERR(spi_engine->clk))
654 return PTR_ERR(spi_engine->clk);
655
656 spi_engine->ref_clk = devm_clk_get_enabled(&pdev->dev, "spi_clk");
657 if (IS_ERR(spi_engine->ref_clk))
658 return PTR_ERR(spi_engine->ref_clk);
659
660 spi_engine->base = devm_platform_ioremap_resource(pdev, 0);
661 if (IS_ERR(spi_engine->base))
662 return PTR_ERR(spi_engine->base);
663
664 version = readl(spi_engine->base + SPI_ENGINE_REG_VERSION);
665 if (SPI_ENGINE_VERSION_MAJOR(version) != 1) {
666 dev_err(&pdev->dev, "Unsupported peripheral version %u.%u.%c\n",
667 SPI_ENGINE_VERSION_MAJOR(version),
668 SPI_ENGINE_VERSION_MINOR(version),
669 SPI_ENGINE_VERSION_PATCH(version));
670 return -ENODEV;
671 }
672
673 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_RESET);
674 writel_relaxed(0xff, spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
675 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
676
677 ret = devm_add_action_or_reset(&pdev->dev, spi_engine_release_hw,
678 spi_engine);
679 if (ret)
680 return ret;
681
682 ret = devm_request_irq(&pdev->dev, irq, spi_engine_irq, 0, pdev->name,
683 host);
684 if (ret)
685 return ret;
686
687 host->dev.of_node = pdev->dev.of_node;
688 host->mode_bits = SPI_CPOL | SPI_CPHA | SPI_3WIRE;
689 host->bits_per_word_mask = SPI_BPW_RANGE_MASK(1, 32);
690 host->max_speed_hz = clk_get_rate(spi_engine->ref_clk) / 2;
691 host->transfer_one_message = spi_engine_transfer_one_message;
692 host->prepare_message = spi_engine_prepare_message;
693 host->unprepare_message = spi_engine_unprepare_message;
694 host->num_chipselect = 8;
695
696 if (host->max_speed_hz == 0)
697 return dev_err_probe(&pdev->dev, -EINVAL, "spi_clk rate is 0");
698
699 ret = devm_spi_register_controller(&pdev->dev, host);
700 if (ret)
701 return ret;
702
703 platform_set_drvdata(pdev, host);
704
705 return 0;
706}
707
708static const struct of_device_id spi_engine_match_table[] = {
709 { .compatible = "adi,axi-spi-engine-1.00.a" },
710 { },
711};
712MODULE_DEVICE_TABLE(of, spi_engine_match_table);
713
714static struct platform_driver spi_engine_driver = {
715 .probe = spi_engine_probe,
716 .driver = {
717 .name = "spi-engine",
718 .of_match_table = spi_engine_match_table,
719 },
720};
721module_platform_driver(spi_engine_driver);
722
723MODULE_AUTHOR("Lars-Peter Clausen <lars@metafoo.de>");
724MODULE_DESCRIPTION("Analog Devices SPI engine peripheral driver");
725MODULE_LICENSE("GPL");
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * SPI-Engine SPI controller driver
4 * Copyright 2015 Analog Devices Inc.
5 * Author: Lars-Peter Clausen <lars@metafoo.de>
6 */
7
8#include <linux/clk.h>
9#include <linux/interrupt.h>
10#include <linux/io.h>
11#include <linux/of.h>
12#include <linux/module.h>
13#include <linux/platform_device.h>
14#include <linux/spi/spi.h>
15
16#define SPI_ENGINE_VERSION_MAJOR(x) ((x >> 16) & 0xff)
17#define SPI_ENGINE_VERSION_MINOR(x) ((x >> 8) & 0xff)
18#define SPI_ENGINE_VERSION_PATCH(x) (x & 0xff)
19
20#define SPI_ENGINE_REG_VERSION 0x00
21
22#define SPI_ENGINE_REG_RESET 0x40
23
24#define SPI_ENGINE_REG_INT_ENABLE 0x80
25#define SPI_ENGINE_REG_INT_PENDING 0x84
26#define SPI_ENGINE_REG_INT_SOURCE 0x88
27
28#define SPI_ENGINE_REG_SYNC_ID 0xc0
29
30#define SPI_ENGINE_REG_CMD_FIFO_ROOM 0xd0
31#define SPI_ENGINE_REG_SDO_FIFO_ROOM 0xd4
32#define SPI_ENGINE_REG_SDI_FIFO_LEVEL 0xd8
33
34#define SPI_ENGINE_REG_CMD_FIFO 0xe0
35#define SPI_ENGINE_REG_SDO_DATA_FIFO 0xe4
36#define SPI_ENGINE_REG_SDI_DATA_FIFO 0xe8
37#define SPI_ENGINE_REG_SDI_DATA_FIFO_PEEK 0xec
38
39#define SPI_ENGINE_INT_CMD_ALMOST_EMPTY BIT(0)
40#define SPI_ENGINE_INT_SDO_ALMOST_EMPTY BIT(1)
41#define SPI_ENGINE_INT_SDI_ALMOST_FULL BIT(2)
42#define SPI_ENGINE_INT_SYNC BIT(3)
43
44#define SPI_ENGINE_CONFIG_CPHA BIT(0)
45#define SPI_ENGINE_CONFIG_CPOL BIT(1)
46#define SPI_ENGINE_CONFIG_3WIRE BIT(2)
47
48#define SPI_ENGINE_INST_TRANSFER 0x0
49#define SPI_ENGINE_INST_ASSERT 0x1
50#define SPI_ENGINE_INST_WRITE 0x2
51#define SPI_ENGINE_INST_MISC 0x3
52
53#define SPI_ENGINE_CMD_REG_CLK_DIV 0x0
54#define SPI_ENGINE_CMD_REG_CONFIG 0x1
55
56#define SPI_ENGINE_MISC_SYNC 0x0
57#define SPI_ENGINE_MISC_SLEEP 0x1
58
59#define SPI_ENGINE_TRANSFER_WRITE 0x1
60#define SPI_ENGINE_TRANSFER_READ 0x2
61
62#define SPI_ENGINE_CMD(inst, arg1, arg2) \
63 (((inst) << 12) | ((arg1) << 8) | (arg2))
64
65#define SPI_ENGINE_CMD_TRANSFER(flags, n) \
66 SPI_ENGINE_CMD(SPI_ENGINE_INST_TRANSFER, (flags), (n))
67#define SPI_ENGINE_CMD_ASSERT(delay, cs) \
68 SPI_ENGINE_CMD(SPI_ENGINE_INST_ASSERT, (delay), (cs))
69#define SPI_ENGINE_CMD_WRITE(reg, val) \
70 SPI_ENGINE_CMD(SPI_ENGINE_INST_WRITE, (reg), (val))
71#define SPI_ENGINE_CMD_SLEEP(delay) \
72 SPI_ENGINE_CMD(SPI_ENGINE_INST_MISC, SPI_ENGINE_MISC_SLEEP, (delay))
73#define SPI_ENGINE_CMD_SYNC(id) \
74 SPI_ENGINE_CMD(SPI_ENGINE_INST_MISC, SPI_ENGINE_MISC_SYNC, (id))
75
76struct spi_engine_program {
77 unsigned int length;
78 uint16_t instructions[];
79};
80
81struct spi_engine {
82 struct clk *clk;
83 struct clk *ref_clk;
84
85 spinlock_t lock;
86
87 void __iomem *base;
88
89 struct spi_message *msg;
90 struct spi_engine_program *p;
91 unsigned cmd_length;
92 const uint16_t *cmd_buf;
93
94 struct spi_transfer *tx_xfer;
95 unsigned int tx_length;
96 const uint8_t *tx_buf;
97
98 struct spi_transfer *rx_xfer;
99 unsigned int rx_length;
100 uint8_t *rx_buf;
101
102 unsigned int sync_id;
103 unsigned int completed_id;
104
105 unsigned int int_enable;
106};
107
108static void spi_engine_program_add_cmd(struct spi_engine_program *p,
109 bool dry, uint16_t cmd)
110{
111 if (!dry)
112 p->instructions[p->length] = cmd;
113 p->length++;
114}
115
116static unsigned int spi_engine_get_config(struct spi_device *spi)
117{
118 unsigned int config = 0;
119
120 if (spi->mode & SPI_CPOL)
121 config |= SPI_ENGINE_CONFIG_CPOL;
122 if (spi->mode & SPI_CPHA)
123 config |= SPI_ENGINE_CONFIG_CPHA;
124 if (spi->mode & SPI_3WIRE)
125 config |= SPI_ENGINE_CONFIG_3WIRE;
126
127 return config;
128}
129
130static unsigned int spi_engine_get_clk_div(struct spi_engine *spi_engine,
131 struct spi_device *spi, struct spi_transfer *xfer)
132{
133 unsigned int clk_div;
134
135 clk_div = DIV_ROUND_UP(clk_get_rate(spi_engine->ref_clk),
136 xfer->speed_hz * 2);
137 if (clk_div > 255)
138 clk_div = 255;
139 else if (clk_div > 0)
140 clk_div -= 1;
141
142 return clk_div;
143}
144
145static void spi_engine_gen_xfer(struct spi_engine_program *p, bool dry,
146 struct spi_transfer *xfer)
147{
148 unsigned int len = xfer->len;
149
150 while (len) {
151 unsigned int n = min(len, 256U);
152 unsigned int flags = 0;
153
154 if (xfer->tx_buf)
155 flags |= SPI_ENGINE_TRANSFER_WRITE;
156 if (xfer->rx_buf)
157 flags |= SPI_ENGINE_TRANSFER_READ;
158
159 spi_engine_program_add_cmd(p, dry,
160 SPI_ENGINE_CMD_TRANSFER(flags, n - 1));
161 len -= n;
162 }
163}
164
165static void spi_engine_gen_sleep(struct spi_engine_program *p, bool dry,
166 struct spi_engine *spi_engine, unsigned int clk_div,
167 struct spi_transfer *xfer)
168{
169 unsigned int spi_clk = clk_get_rate(spi_engine->ref_clk);
170 unsigned int t;
171 int delay;
172
173 delay = spi_delay_to_ns(&xfer->delay, xfer);
174 if (delay < 0)
175 return;
176 delay /= 1000;
177
178 if (delay == 0)
179 return;
180
181 t = DIV_ROUND_UP(delay * spi_clk, (clk_div + 1) * 2);
182 while (t) {
183 unsigned int n = min(t, 256U);
184
185 spi_engine_program_add_cmd(p, dry, SPI_ENGINE_CMD_SLEEP(n - 1));
186 t -= n;
187 }
188}
189
190static void spi_engine_gen_cs(struct spi_engine_program *p, bool dry,
191 struct spi_device *spi, bool assert)
192{
193 unsigned int mask = 0xff;
194
195 if (assert)
196 mask ^= BIT(spi->chip_select);
197
198 spi_engine_program_add_cmd(p, dry, SPI_ENGINE_CMD_ASSERT(1, mask));
199}
200
201static int spi_engine_compile_message(struct spi_engine *spi_engine,
202 struct spi_message *msg, bool dry, struct spi_engine_program *p)
203{
204 struct spi_device *spi = msg->spi;
205 struct spi_transfer *xfer;
206 int clk_div, new_clk_div;
207 bool cs_change = true;
208
209 clk_div = -1;
210
211 spi_engine_program_add_cmd(p, dry,
212 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CONFIG,
213 spi_engine_get_config(spi)));
214
215 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
216 new_clk_div = spi_engine_get_clk_div(spi_engine, spi, xfer);
217 if (new_clk_div != clk_div) {
218 clk_div = new_clk_div;
219 spi_engine_program_add_cmd(p, dry,
220 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CLK_DIV,
221 clk_div));
222 }
223
224 if (cs_change)
225 spi_engine_gen_cs(p, dry, spi, true);
226
227 spi_engine_gen_xfer(p, dry, xfer);
228 spi_engine_gen_sleep(p, dry, spi_engine, clk_div, xfer);
229
230 cs_change = xfer->cs_change;
231 if (list_is_last(&xfer->transfer_list, &msg->transfers))
232 cs_change = !cs_change;
233
234 if (cs_change)
235 spi_engine_gen_cs(p, dry, spi, false);
236 }
237
238 return 0;
239}
240
241static void spi_engine_xfer_next(struct spi_engine *spi_engine,
242 struct spi_transfer **_xfer)
243{
244 struct spi_message *msg = spi_engine->msg;
245 struct spi_transfer *xfer = *_xfer;
246
247 if (!xfer) {
248 xfer = list_first_entry(&msg->transfers,
249 struct spi_transfer, transfer_list);
250 } else if (list_is_last(&xfer->transfer_list, &msg->transfers)) {
251 xfer = NULL;
252 } else {
253 xfer = list_next_entry(xfer, transfer_list);
254 }
255
256 *_xfer = xfer;
257}
258
259static void spi_engine_tx_next(struct spi_engine *spi_engine)
260{
261 struct spi_transfer *xfer = spi_engine->tx_xfer;
262
263 do {
264 spi_engine_xfer_next(spi_engine, &xfer);
265 } while (xfer && !xfer->tx_buf);
266
267 spi_engine->tx_xfer = xfer;
268 if (xfer) {
269 spi_engine->tx_length = xfer->len;
270 spi_engine->tx_buf = xfer->tx_buf;
271 } else {
272 spi_engine->tx_buf = NULL;
273 }
274}
275
276static void spi_engine_rx_next(struct spi_engine *spi_engine)
277{
278 struct spi_transfer *xfer = spi_engine->rx_xfer;
279
280 do {
281 spi_engine_xfer_next(spi_engine, &xfer);
282 } while (xfer && !xfer->rx_buf);
283
284 spi_engine->rx_xfer = xfer;
285 if (xfer) {
286 spi_engine->rx_length = xfer->len;
287 spi_engine->rx_buf = xfer->rx_buf;
288 } else {
289 spi_engine->rx_buf = NULL;
290 }
291}
292
293static bool spi_engine_write_cmd_fifo(struct spi_engine *spi_engine)
294{
295 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_CMD_FIFO;
296 unsigned int n, m, i;
297 const uint16_t *buf;
298
299 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_CMD_FIFO_ROOM);
300 while (n && spi_engine->cmd_length) {
301 m = min(n, spi_engine->cmd_length);
302 buf = spi_engine->cmd_buf;
303 for (i = 0; i < m; i++)
304 writel_relaxed(buf[i], addr);
305 spi_engine->cmd_buf += m;
306 spi_engine->cmd_length -= m;
307 n -= m;
308 }
309
310 return spi_engine->cmd_length != 0;
311}
312
313static bool spi_engine_write_tx_fifo(struct spi_engine *spi_engine)
314{
315 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_SDO_DATA_FIFO;
316 unsigned int n, m, i;
317 const uint8_t *buf;
318
319 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_SDO_FIFO_ROOM);
320 while (n && spi_engine->tx_length) {
321 m = min(n, spi_engine->tx_length);
322 buf = spi_engine->tx_buf;
323 for (i = 0; i < m; i++)
324 writel_relaxed(buf[i], addr);
325 spi_engine->tx_buf += m;
326 spi_engine->tx_length -= m;
327 n -= m;
328 if (spi_engine->tx_length == 0)
329 spi_engine_tx_next(spi_engine);
330 }
331
332 return spi_engine->tx_length != 0;
333}
334
335static bool spi_engine_read_rx_fifo(struct spi_engine *spi_engine)
336{
337 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_SDI_DATA_FIFO;
338 unsigned int n, m, i;
339 uint8_t *buf;
340
341 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_SDI_FIFO_LEVEL);
342 while (n && spi_engine->rx_length) {
343 m = min(n, spi_engine->rx_length);
344 buf = spi_engine->rx_buf;
345 for (i = 0; i < m; i++)
346 buf[i] = readl_relaxed(addr);
347 spi_engine->rx_buf += m;
348 spi_engine->rx_length -= m;
349 n -= m;
350 if (spi_engine->rx_length == 0)
351 spi_engine_rx_next(spi_engine);
352 }
353
354 return spi_engine->rx_length != 0;
355}
356
357static irqreturn_t spi_engine_irq(int irq, void *devid)
358{
359 struct spi_master *master = devid;
360 struct spi_engine *spi_engine = spi_master_get_devdata(master);
361 unsigned int disable_int = 0;
362 unsigned int pending;
363
364 pending = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
365
366 if (pending & SPI_ENGINE_INT_SYNC) {
367 writel_relaxed(SPI_ENGINE_INT_SYNC,
368 spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
369 spi_engine->completed_id = readl_relaxed(
370 spi_engine->base + SPI_ENGINE_REG_SYNC_ID);
371 }
372
373 spin_lock(&spi_engine->lock);
374
375 if (pending & SPI_ENGINE_INT_CMD_ALMOST_EMPTY) {
376 if (!spi_engine_write_cmd_fifo(spi_engine))
377 disable_int |= SPI_ENGINE_INT_CMD_ALMOST_EMPTY;
378 }
379
380 if (pending & SPI_ENGINE_INT_SDO_ALMOST_EMPTY) {
381 if (!spi_engine_write_tx_fifo(spi_engine))
382 disable_int |= SPI_ENGINE_INT_SDO_ALMOST_EMPTY;
383 }
384
385 if (pending & (SPI_ENGINE_INT_SDI_ALMOST_FULL | SPI_ENGINE_INT_SYNC)) {
386 if (!spi_engine_read_rx_fifo(spi_engine))
387 disable_int |= SPI_ENGINE_INT_SDI_ALMOST_FULL;
388 }
389
390 if (pending & SPI_ENGINE_INT_SYNC) {
391 if (spi_engine->msg &&
392 spi_engine->completed_id == spi_engine->sync_id) {
393 struct spi_message *msg = spi_engine->msg;
394
395 kfree(spi_engine->p);
396 msg->status = 0;
397 msg->actual_length = msg->frame_length;
398 spi_engine->msg = NULL;
399 spi_finalize_current_message(master);
400 disable_int |= SPI_ENGINE_INT_SYNC;
401 }
402 }
403
404 if (disable_int) {
405 spi_engine->int_enable &= ~disable_int;
406 writel_relaxed(spi_engine->int_enable,
407 spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
408 }
409
410 spin_unlock(&spi_engine->lock);
411
412 return IRQ_HANDLED;
413}
414
415static int spi_engine_transfer_one_message(struct spi_master *master,
416 struct spi_message *msg)
417{
418 struct spi_engine_program p_dry, *p;
419 struct spi_engine *spi_engine = spi_master_get_devdata(master);
420 unsigned int int_enable = 0;
421 unsigned long flags;
422 size_t size;
423
424 p_dry.length = 0;
425 spi_engine_compile_message(spi_engine, msg, true, &p_dry);
426
427 size = sizeof(*p->instructions) * (p_dry.length + 1);
428 p = kzalloc(sizeof(*p) + size, GFP_KERNEL);
429 if (!p)
430 return -ENOMEM;
431 spi_engine_compile_message(spi_engine, msg, false, p);
432
433 spin_lock_irqsave(&spi_engine->lock, flags);
434 spi_engine->sync_id = (spi_engine->sync_id + 1) & 0xff;
435 spi_engine_program_add_cmd(p, false,
436 SPI_ENGINE_CMD_SYNC(spi_engine->sync_id));
437
438 spi_engine->msg = msg;
439 spi_engine->p = p;
440
441 spi_engine->cmd_buf = p->instructions;
442 spi_engine->cmd_length = p->length;
443 if (spi_engine_write_cmd_fifo(spi_engine))
444 int_enable |= SPI_ENGINE_INT_CMD_ALMOST_EMPTY;
445
446 spi_engine_tx_next(spi_engine);
447 if (spi_engine_write_tx_fifo(spi_engine))
448 int_enable |= SPI_ENGINE_INT_SDO_ALMOST_EMPTY;
449
450 spi_engine_rx_next(spi_engine);
451 if (spi_engine->rx_length != 0)
452 int_enable |= SPI_ENGINE_INT_SDI_ALMOST_FULL;
453
454 int_enable |= SPI_ENGINE_INT_SYNC;
455
456 writel_relaxed(int_enable,
457 spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
458 spi_engine->int_enable = int_enable;
459 spin_unlock_irqrestore(&spi_engine->lock, flags);
460
461 return 0;
462}
463
464static int spi_engine_probe(struct platform_device *pdev)
465{
466 struct spi_engine *spi_engine;
467 struct spi_master *master;
468 unsigned int version;
469 int irq;
470 int ret;
471
472 irq = platform_get_irq(pdev, 0);
473 if (irq <= 0)
474 return -ENXIO;
475
476 spi_engine = devm_kzalloc(&pdev->dev, sizeof(*spi_engine), GFP_KERNEL);
477 if (!spi_engine)
478 return -ENOMEM;
479
480 master = spi_alloc_master(&pdev->dev, 0);
481 if (!master)
482 return -ENOMEM;
483
484 spi_master_set_devdata(master, spi_engine);
485
486 spin_lock_init(&spi_engine->lock);
487
488 spi_engine->clk = devm_clk_get(&pdev->dev, "s_axi_aclk");
489 if (IS_ERR(spi_engine->clk)) {
490 ret = PTR_ERR(spi_engine->clk);
491 goto err_put_master;
492 }
493
494 spi_engine->ref_clk = devm_clk_get(&pdev->dev, "spi_clk");
495 if (IS_ERR(spi_engine->ref_clk)) {
496 ret = PTR_ERR(spi_engine->ref_clk);
497 goto err_put_master;
498 }
499
500 ret = clk_prepare_enable(spi_engine->clk);
501 if (ret)
502 goto err_put_master;
503
504 ret = clk_prepare_enable(spi_engine->ref_clk);
505 if (ret)
506 goto err_clk_disable;
507
508 spi_engine->base = devm_platform_ioremap_resource(pdev, 0);
509 if (IS_ERR(spi_engine->base)) {
510 ret = PTR_ERR(spi_engine->base);
511 goto err_ref_clk_disable;
512 }
513
514 version = readl(spi_engine->base + SPI_ENGINE_REG_VERSION);
515 if (SPI_ENGINE_VERSION_MAJOR(version) != 1) {
516 dev_err(&pdev->dev, "Unsupported peripheral version %u.%u.%c\n",
517 SPI_ENGINE_VERSION_MAJOR(version),
518 SPI_ENGINE_VERSION_MINOR(version),
519 SPI_ENGINE_VERSION_PATCH(version));
520 ret = -ENODEV;
521 goto err_ref_clk_disable;
522 }
523
524 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_RESET);
525 writel_relaxed(0xff, spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
526 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
527
528 ret = request_irq(irq, spi_engine_irq, 0, pdev->name, master);
529 if (ret)
530 goto err_ref_clk_disable;
531
532 master->dev.of_node = pdev->dev.of_node;
533 master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_3WIRE;
534 master->bits_per_word_mask = SPI_BPW_MASK(8);
535 master->max_speed_hz = clk_get_rate(spi_engine->ref_clk) / 2;
536 master->transfer_one_message = spi_engine_transfer_one_message;
537 master->num_chipselect = 8;
538
539 ret = spi_register_master(master);
540 if (ret)
541 goto err_free_irq;
542
543 platform_set_drvdata(pdev, master);
544
545 return 0;
546err_free_irq:
547 free_irq(irq, master);
548err_ref_clk_disable:
549 clk_disable_unprepare(spi_engine->ref_clk);
550err_clk_disable:
551 clk_disable_unprepare(spi_engine->clk);
552err_put_master:
553 spi_master_put(master);
554 return ret;
555}
556
557static int spi_engine_remove(struct platform_device *pdev)
558{
559 struct spi_master *master = spi_master_get(platform_get_drvdata(pdev));
560 struct spi_engine *spi_engine = spi_master_get_devdata(master);
561 int irq = platform_get_irq(pdev, 0);
562
563 spi_unregister_master(master);
564
565 free_irq(irq, master);
566
567 spi_master_put(master);
568
569 writel_relaxed(0xff, spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
570 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
571 writel_relaxed(0x01, spi_engine->base + SPI_ENGINE_REG_RESET);
572
573 clk_disable_unprepare(spi_engine->ref_clk);
574 clk_disable_unprepare(spi_engine->clk);
575
576 return 0;
577}
578
579static const struct of_device_id spi_engine_match_table[] = {
580 { .compatible = "adi,axi-spi-engine-1.00.a" },
581 { },
582};
583MODULE_DEVICE_TABLE(of, spi_engine_match_table);
584
585static struct platform_driver spi_engine_driver = {
586 .probe = spi_engine_probe,
587 .remove = spi_engine_remove,
588 .driver = {
589 .name = "spi-engine",
590 .of_match_table = spi_engine_match_table,
591 },
592};
593module_platform_driver(spi_engine_driver);
594
595MODULE_AUTHOR("Lars-Peter Clausen <lars@metafoo.de>");
596MODULE_DESCRIPTION("Analog Devices SPI engine peripheral driver");
597MODULE_LICENSE("GPL");