1 // SPDX-License-Identifier: GPL-2.0-only
3 * SPI-Engine SPI controller driver
4 * Copyright 2015 Analog Devices Inc.
10 #include <linux/interrupt.h>
13 #include <linux/module.h>
14 #include <linux/platform_device.h>
15 #include <linux/spi/spi.h>
16 #include <linux/timer.h>
18 #define SPI_ENGINE_VERSION_MAJOR(x) ((x >> 16) & 0xff)
19 #define SPI_ENGINE_VERSION_MINOR(x) ((x >> 8) & 0xff)
20 #define SPI_ENGINE_VERSION_PATCH(x) (x & 0xff)
22 #define SPI_ENGINE_REG_VERSION 0x00
24 #define SPI_ENGINE_REG_RESET 0x40
26 #define SPI_ENGINE_REG_INT_ENABLE 0x80
27 #define SPI_ENGINE_REG_INT_PENDING 0x84
28 #define SPI_ENGINE_REG_INT_SOURCE 0x88
30 #define SPI_ENGINE_REG_SYNC_ID 0xc0
32 #define SPI_ENGINE_REG_CMD_FIFO_ROOM 0xd0
33 #define SPI_ENGINE_REG_SDO_FIFO_ROOM 0xd4
34 #define SPI_ENGINE_REG_SDI_FIFO_LEVEL 0xd8
36 #define SPI_ENGINE_REG_CMD_FIFO 0xe0
37 #define SPI_ENGINE_REG_SDO_DATA_FIFO 0xe4
38 #define SPI_ENGINE_REG_SDI_DATA_FIFO 0xe8
39 #define SPI_ENGINE_REG_SDI_DATA_FIFO_PEEK 0xec
41 #define SPI_ENGINE_INT_CMD_ALMOST_EMPTY BIT(0)
42 #define SPI_ENGINE_INT_SDO_ALMOST_EMPTY BIT(1)
43 #define SPI_ENGINE_INT_SDI_ALMOST_FULL BIT(2)
44 #define SPI_ENGINE_INT_SYNC BIT(3)
46 #define SPI_ENGINE_CONFIG_CPHA BIT(0)
47 #define SPI_ENGINE_CONFIG_CPOL BIT(1)
48 #define SPI_ENGINE_CONFIG_3WIRE BIT(2)
50 #define SPI_ENGINE_INST_TRANSFER 0x0
51 #define SPI_ENGINE_INST_ASSERT 0x1
52 #define SPI_ENGINE_INST_WRITE 0x2
53 #define SPI_ENGINE_INST_MISC 0x3
55 #define SPI_ENGINE_CMD_REG_CLK_DIV 0x0
56 #define SPI_ENGINE_CMD_REG_CONFIG 0x1
57 #define SPI_ENGINE_CMD_REG_XFER_BITS 0x2
59 #define SPI_ENGINE_MISC_SYNC 0x0
60 #define SPI_ENGINE_MISC_SLEEP 0x1
62 #define SPI_ENGINE_TRANSFER_WRITE 0x1
63 #define SPI_ENGINE_TRANSFER_READ 0x2
65 #define SPI_ENGINE_CMD(inst, arg1, arg2) \
66 (((inst) << 12) | ((arg1) << 8) | (arg2))
68 #define SPI_ENGINE_CMD_TRANSFER(flags, n) \
69 SPI_ENGINE_CMD(SPI_ENGINE_INST_TRANSFER, (flags), (n))
70 #define SPI_ENGINE_CMD_ASSERT(delay, cs) \
71 SPI_ENGINE_CMD(SPI_ENGINE_INST_ASSERT, (delay), (cs))
72 #define SPI_ENGINE_CMD_WRITE(reg, val) \
73 SPI_ENGINE_CMD(SPI_ENGINE_INST_WRITE, (reg), (val))
74 #define SPI_ENGINE_CMD_SLEEP(delay) \
75 SPI_ENGINE_CMD(SPI_ENGINE_INST_MISC, SPI_ENGINE_MISC_SLEEP, (delay))
76 #define SPI_ENGINE_CMD_SYNC(id) \
77 SPI_ENGINE_CMD(SPI_ENGINE_INST_MISC, SPI_ENGINE_MISC_SYNC, (id))
79 struct spi_engine_program {
81 uint16_t instructions[];
85 * struct spi_engine_message_state - SPI engine per-message state
87 struct spi_engine_message_state {
88 /** @p: Instructions for executing this message. */
89 struct spi_engine_program *p;
90 /** @cmd_length: Number of elements in cmd_buf array. */
92 /** @cmd_buf: Array of commands not yet written to CMD FIFO. */
93 const uint16_t *cmd_buf;
94 /** @tx_xfer: Next xfer with tx_buf not yet fully written to TX FIFO. */
95 struct spi_transfer *tx_xfer;
96 /** @tx_length: Size of tx_buf in bytes. */
97 unsigned int tx_length;
98 /** @tx_buf: Bytes not yet written to TX FIFO. */
99 const uint8_t *tx_buf;
100 /** @rx_xfer: Next xfer with rx_buf not yet fully written to RX FIFO. */
101 struct spi_transfer *rx_xfer;
102 /** @rx_length: Size of tx_buf in bytes. */
103 unsigned int rx_length;
104 /** @rx_buf: Bytes not yet written to the RX FIFO. */
106 /** @sync_id: ID to correlate SYNC interrupts with this message. */
118 struct timer_list watchdog_timer;
119 struct spi_controller *controller;
121 unsigned int int_enable;
124 static void spi_engine_program_add_cmd(struct spi_engine_program *p,
125 bool dry, uint16_t cmd)
128 p->instructions[p->length] = cmd;
132 static unsigned int spi_engine_get_config(struct spi_device *spi)
134 unsigned int config = 0;
136 if (spi->mode & SPI_CPOL)
137 config |= SPI_ENGINE_CONFIG_CPOL;
138 if (spi->mode & SPI_CPHA)
139 config |= SPI_ENGINE_CONFIG_CPHA;
140 if (spi->mode & SPI_3WIRE)
141 config |= SPI_ENGINE_CONFIG_3WIRE;
146 static void spi_engine_gen_xfer(struct spi_engine_program *p, bool dry,
147 struct spi_transfer *xfer)
151 if (xfer->bits_per_word <= 8)
153 else if (xfer->bits_per_word <= 16)
159 unsigned int n = min(len, 256U);
160 unsigned int flags = 0;
163 flags |= SPI_ENGINE_TRANSFER_WRITE;
165 flags |= SPI_ENGINE_TRANSFER_READ;
167 spi_engine_program_add_cmd(p, dry,
168 SPI_ENGINE_CMD_TRANSFER(flags, n - 1));
173 static void spi_engine_gen_sleep(struct spi_engine_program *p, bool dry,
174 int delay_ns, u32 sclk_hz)
178 /* negative delay indicates error, e.g. from spi_delay_to_ns() */
182 /* rounding down since executing the instruction adds a couple of ticks delay */
183 t = DIV_ROUND_DOWN_ULL((u64)delay_ns * sclk_hz, NSEC_PER_SEC);
185 unsigned int n = min(t, 256U);
187 spi_engine_program_add_cmd(p, dry, SPI_ENGINE_CMD_SLEEP(n - 1));
192 static void spi_engine_gen_cs(struct spi_engine_program *p, bool dry,
193 struct spi_device *spi, bool assert)
195 unsigned int mask = 0xff;
198 mask ^= BIT(spi_get_chipselect(spi, 0));
200 spi_engine_program_add_cmd(p, dry, SPI_ENGINE_CMD_ASSERT(0, mask));
204 * Performs precompile steps on the message.
206 * The SPI core does most of the message/transfer validation and filling in
207 * fields for us via __spi_validate(). This fixes up anything remaining not
210 * NB: This is separate from spi_engine_compile_message() because the latter
211 * is called twice and would otherwise result in double-evaluation.
213 static void spi_engine_precompile_message(struct spi_message *msg)
215 unsigned int clk_div, max_hz = msg->spi->controller->max_speed_hz;
216 struct spi_transfer *xfer;
218 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
219 clk_div = DIV_ROUND_UP(max_hz, xfer->speed_hz);
220 xfer->effective_speed_hz = max_hz / min(clk_div, 256U);
224 static void spi_engine_compile_message(struct spi_message *msg, bool dry,
225 struct spi_engine_program *p)
227 struct spi_device *spi = msg->spi;
228 struct spi_controller *host = spi->controller;
229 struct spi_transfer *xfer;
230 int clk_div, new_clk_div;
231 bool keep_cs = false;
232 u8 bits_per_word = 0;
236 spi_engine_program_add_cmd(p, dry,
237 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CONFIG,
238 spi_engine_get_config(spi)));
240 xfer = list_first_entry(&msg->transfers, struct spi_transfer, transfer_list);
241 spi_engine_gen_cs(p, dry, spi, !xfer->cs_off);
243 list_for_each_entry(xfer, &msg->transfers, transfer_list) {
244 new_clk_div = host->max_speed_hz / xfer->effective_speed_hz;
245 if (new_clk_div != clk_div) {
246 clk_div = new_clk_div;
247 /* actual divider used is register value + 1 */
248 spi_engine_program_add_cmd(p, dry,
249 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CLK_DIV,
253 if (bits_per_word != xfer->bits_per_word) {
254 bits_per_word = xfer->bits_per_word;
255 spi_engine_program_add_cmd(p, dry,
256 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_XFER_BITS,
260 spi_engine_gen_xfer(p, dry, xfer);
261 spi_engine_gen_sleep(p, dry, spi_delay_to_ns(&xfer->delay, xfer),
262 xfer->effective_speed_hz);
264 if (xfer->cs_change) {
265 if (list_is_last(&xfer->transfer_list, &msg->transfers)) {
269 spi_engine_gen_cs(p, dry, spi, false);
271 spi_engine_gen_sleep(p, dry, spi_delay_to_ns(
272 &xfer->cs_change_delay, xfer),
273 xfer->effective_speed_hz);
275 if (!list_next_entry(xfer, transfer_list)->cs_off)
276 spi_engine_gen_cs(p, dry, spi, true);
278 } else if (!list_is_last(&xfer->transfer_list, &msg->transfers) &&
279 xfer->cs_off != list_next_entry(xfer, transfer_list)->cs_off) {
280 spi_engine_gen_cs(p, dry, spi, xfer->cs_off);
285 spi_engine_gen_cs(p, dry, spi, false);
288 * Restore clockdiv to default so that future gen_sleep commands don't
289 * have to be aware of the current register state.
292 spi_engine_program_add_cmd(p, dry,
293 SPI_ENGINE_CMD_WRITE(SPI_ENGINE_CMD_REG_CLK_DIV, 0));
296 static void spi_engine_xfer_next(struct spi_message *msg,
297 struct spi_transfer **_xfer)
299 struct spi_transfer *xfer = *_xfer;
302 xfer = list_first_entry(&msg->transfers,
303 struct spi_transfer, transfer_list);
304 } else if (list_is_last(&xfer->transfer_list, &msg->transfers)) {
307 xfer = list_next_entry(xfer, transfer_list);
313 static void spi_engine_tx_next(struct spi_message *msg)
315 struct spi_engine_message_state *st = msg->state;
316 struct spi_transfer *xfer = st->tx_xfer;
319 spi_engine_xfer_next(msg, &xfer);
320 } while (xfer && !xfer->tx_buf);
324 st->tx_length = xfer->len;
325 st->tx_buf = xfer->tx_buf;
331 static void spi_engine_rx_next(struct spi_message *msg)
333 struct spi_engine_message_state *st = msg->state;
334 struct spi_transfer *xfer = st->rx_xfer;
337 spi_engine_xfer_next(msg, &xfer);
338 } while (xfer && !xfer->rx_buf);
342 st->rx_length = xfer->len;
343 st->rx_buf = xfer->rx_buf;
349 static bool spi_engine_write_cmd_fifo(struct spi_engine *spi_engine,
350 struct spi_message *msg)
352 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_CMD_FIFO;
353 struct spi_engine_message_state *st = msg->state;
354 unsigned int n, m, i;
357 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_CMD_FIFO_ROOM);
358 while (n && st->cmd_length) {
359 m = min(n, st->cmd_length);
361 for (i = 0; i < m; i++)
362 writel_relaxed(buf[i], addr);
368 return st->cmd_length != 0;
371 static bool spi_engine_write_tx_fifo(struct spi_engine *spi_engine,
372 struct spi_message *msg)
374 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_SDO_DATA_FIFO;
375 struct spi_engine_message_state *st = msg->state;
376 unsigned int n, m, i;
378 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_SDO_FIFO_ROOM);
379 while (n && st->tx_length) {
380 if (st->tx_xfer->bits_per_word <= 8) {
381 const u8 *buf = st->tx_buf;
383 m = min(n, st->tx_length);
384 for (i = 0; i < m; i++)
385 writel_relaxed(buf[i], addr);
388 } else if (st->tx_xfer->bits_per_word <= 16) {
389 const u16 *buf = (const u16 *)st->tx_buf;
391 m = min(n, st->tx_length / 2);
392 for (i = 0; i < m; i++)
393 writel_relaxed(buf[i], addr);
395 st->tx_length -= m * 2;
397 const u32 *buf = (const u32 *)st->tx_buf;
399 m = min(n, st->tx_length / 4);
400 for (i = 0; i < m; i++)
401 writel_relaxed(buf[i], addr);
403 st->tx_length -= m * 4;
406 if (st->tx_length == 0)
407 spi_engine_tx_next(msg);
410 return st->tx_length != 0;
413 static bool spi_engine_read_rx_fifo(struct spi_engine *spi_engine,
414 struct spi_message *msg)
416 void __iomem *addr = spi_engine->base + SPI_ENGINE_REG_SDI_DATA_FIFO;
417 struct spi_engine_message_state *st = msg->state;
418 unsigned int n, m, i;
420 n = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_SDI_FIFO_LEVEL);
421 while (n && st->rx_length) {
422 if (st->rx_xfer->bits_per_word <= 8) {
423 u8 *buf = st->rx_buf;
425 m = min(n, st->rx_length);
426 for (i = 0; i < m; i++)
427 buf[i] = readl_relaxed(addr);
430 } else if (st->rx_xfer->bits_per_word <= 16) {
431 u16 *buf = (u16 *)st->rx_buf;
433 m = min(n, st->rx_length / 2);
434 for (i = 0; i < m; i++)
435 buf[i] = readl_relaxed(addr);
437 st->rx_length -= m * 2;
439 u32 *buf = (u32 *)st->rx_buf;
441 m = min(n, st->rx_length / 4);
442 for (i = 0; i < m; i++)
443 buf[i] = readl_relaxed(addr);
445 st->rx_length -= m * 4;
448 if (st->rx_length == 0)
449 spi_engine_rx_next(msg);
452 return st->rx_length != 0;
455 static irqreturn_t spi_engine_irq(int irq, void *devid)
457 struct spi_controller *host = devid;
458 struct spi_message *msg = host->cur_msg;
459 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
460 unsigned int disable_int = 0;
461 unsigned int pending;
462 int completed_id = -1;
464 pending = readl_relaxed(spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
466 if (pending & SPI_ENGINE_INT_SYNC) {
467 writel_relaxed(SPI_ENGINE_INT_SYNC,
468 spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
469 completed_id = readl_relaxed(
470 spi_engine->base + SPI_ENGINE_REG_SYNC_ID);
473 spin_lock(&spi_engine->lock);
475 if (pending & SPI_ENGINE_INT_CMD_ALMOST_EMPTY) {
476 if (!spi_engine_write_cmd_fifo(spi_engine, msg))
477 disable_int |= SPI_ENGINE_INT_CMD_ALMOST_EMPTY;
480 if (pending & SPI_ENGINE_INT_SDO_ALMOST_EMPTY) {
481 if (!spi_engine_write_tx_fifo(spi_engine, msg))
482 disable_int |= SPI_ENGINE_INT_SDO_ALMOST_EMPTY;
485 if (pending & (SPI_ENGINE_INT_SDI_ALMOST_FULL | SPI_ENGINE_INT_SYNC)) {
486 if (!spi_engine_read_rx_fifo(spi_engine, msg))
487 disable_int |= SPI_ENGINE_INT_SDI_ALMOST_FULL;
490 if (pending & SPI_ENGINE_INT_SYNC && msg) {
491 struct spi_engine_message_state *st = msg->state;
493 if (completed_id == st->sync_id) {
494 if (timer_delete_sync(&spi_engine->watchdog_timer)) {
496 msg->actual_length = msg->frame_length;
497 spi_finalize_current_message(host);
499 disable_int |= SPI_ENGINE_INT_SYNC;
504 spi_engine->int_enable &= ~disable_int;
505 writel_relaxed(spi_engine->int_enable,
506 spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
509 spin_unlock(&spi_engine->lock);
514 static int spi_engine_prepare_message(struct spi_controller *host,
515 struct spi_message *msg)
517 struct spi_engine_program p_dry, *p;
518 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
519 struct spi_engine_message_state *st;
523 st = kzalloc(sizeof(*st), GFP_KERNEL);
527 spi_engine_precompile_message(msg);
530 spi_engine_compile_message(msg, true, &p_dry);
532 size = sizeof(*p->instructions) * (p_dry.length + 1);
533 p = kzalloc(sizeof(*p) + size, GFP_KERNEL);
539 ret = ida_alloc_range(&spi_engine->sync_ida, 0, U8_MAX, GFP_KERNEL);
548 spi_engine_compile_message(msg, false, p);
550 spi_engine_program_add_cmd(p, false, SPI_ENGINE_CMD_SYNC(st->sync_id));
553 st->cmd_buf = p->instructions;
554 st->cmd_length = p->length;
560 static int spi_engine_unprepare_message(struct spi_controller *host,
561 struct spi_message *msg)
563 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
564 struct spi_engine_message_state *st = msg->state;
566 ida_free(&spi_engine->sync_ida, st->sync_id);
573 static int spi_engine_transfer_one_message(struct spi_controller *host,
574 struct spi_message *msg)
576 struct spi_engine *spi_engine = spi_controller_get_devdata(host);
577 struct spi_engine_message_state *st = msg->state;
578 unsigned int int_enable = 0;
581 mod_timer(&spi_engine->watchdog_timer, jiffies + msecs_to_jiffies(5000));
583 spin_lock_irqsave(&spi_engine->lock, flags);
585 if (spi_engine_write_cmd_fifo(spi_engine, msg))
586 int_enable |= SPI_ENGINE_INT_CMD_ALMOST_EMPTY;
588 spi_engine_tx_next(msg);
589 if (spi_engine_write_tx_fifo(spi_engine, msg))
590 int_enable |= SPI_ENGINE_INT_SDO_ALMOST_EMPTY;
592 spi_engine_rx_next(msg);
593 if (st->rx_length != 0)
594 int_enable |= SPI_ENGINE_INT_SDI_ALMOST_FULL;
596 int_enable |= SPI_ENGINE_INT_SYNC;
598 writel_relaxed(int_enable,
599 spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
600 spi_engine->int_enable = int_enable;
601 spin_unlock_irqrestore(&spi_engine->lock, flags);
606 static void spi_engine_timeout(struct timer_list *timer)
608 struct spi_engine *spi_engine = from_timer(spi_engine, timer, watchdog_timer);
609 struct spi_controller *host = spi_engine->controller;
611 if (WARN_ON(!host->cur_msg))
615 "Timeout occurred while waiting for transfer to complete. Hardware is probably broken.\n");
616 host->cur_msg->status = -ETIMEDOUT;
617 spi_finalize_current_message(host);
620 static void spi_engine_release_hw(void *p)
622 struct spi_engine *spi_engine = p;
624 writel_relaxed(0xff, spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
625 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
626 writel_relaxed(0x01, spi_engine->base + SPI_ENGINE_REG_RESET);
629 static int spi_engine_probe(struct platform_device *pdev)
631 struct spi_engine *spi_engine;
632 struct spi_controller *host;
633 unsigned int version;
637 irq = platform_get_irq(pdev, 0);
641 host = devm_spi_alloc_host(&pdev->dev, sizeof(*spi_engine));
645 spi_engine = spi_controller_get_devdata(host);
647 spin_lock_init(&spi_engine->lock);
648 ida_init(&spi_engine->sync_ida);
649 timer_setup(&spi_engine->watchdog_timer, spi_engine_timeout, TIMER_IRQSAFE);
650 spi_engine->controller = host;
652 spi_engine->clk = devm_clk_get_enabled(&pdev->dev, "s_axi_aclk");
653 if (IS_ERR(spi_engine->clk))
654 return PTR_ERR(spi_engine->clk);
656 spi_engine->ref_clk = devm_clk_get_enabled(&pdev->dev, "spi_clk");
657 if (IS_ERR(spi_engine->ref_clk))
658 return PTR_ERR(spi_engine->ref_clk);
660 spi_engine->base = devm_platform_ioremap_resource(pdev, 0);
661 if (IS_ERR(spi_engine->base))
662 return PTR_ERR(spi_engine->base);
664 version = readl(spi_engine->base + SPI_ENGINE_REG_VERSION);
665 if (SPI_ENGINE_VERSION_MAJOR(version) != 1) {
666 dev_err(&pdev->dev, "Unsupported peripheral version %u.%u.%c\n",
667 SPI_ENGINE_VERSION_MAJOR(version),
668 SPI_ENGINE_VERSION_MINOR(version),
669 SPI_ENGINE_VERSION_PATCH(version));
673 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_RESET);
674 writel_relaxed(0xff, spi_engine->base + SPI_ENGINE_REG_INT_PENDING);
675 writel_relaxed(0x00, spi_engine->base + SPI_ENGINE_REG_INT_ENABLE);
677 ret = devm_add_action_or_reset(&pdev->dev, spi_engine_release_hw,
682 ret = devm_request_irq(&pdev->dev, irq, spi_engine_irq, 0, pdev->name,
687 host->dev.of_node = pdev->dev.of_node;
688 host->mode_bits = SPI_CPOL | SPI_CPHA | SPI_3WIRE;
689 host->bits_per_word_mask = SPI_BPW_RANGE_MASK(1, 32);
690 host->max_speed_hz = clk_get_rate(spi_engine->ref_clk) / 2;
691 host->transfer_one_message = spi_engine_transfer_one_message;
692 host->prepare_message = spi_engine_prepare_message;
693 host->unprepare_message = spi_engine_unprepare_message;
694 host->num_chipselect = 8;
696 if (host->max_speed_hz == 0)
697 return dev_err_probe(&pdev->dev, -EINVAL, "spi_clk rate is 0");
699 ret = devm_spi_register_controller(&pdev->dev, host);
703 platform_set_drvdata(pdev, host);
708 static const struct of_device_id spi_engine_match_table[] = {
709 { .compatible = "adi,axi-spi-engine-1.00.a" },
712 MODULE_DEVICE_TABLE(of, spi_engine_match_table);
714 static struct platform_driver spi_engine_driver = {
715 .probe = spi_engine_probe,
717 .name = "spi-engine",
718 .of_match_table = spi_engine_match_table,
721 module_platform_driver(spi_engine_driver);
724 MODULE_DESCRIPTION("Analog Devices SPI engine peripheral driver");
725 MODULE_LICENSE("GPL");