// SPDX-License-Identifier: GPL-2.0 // // Driver for AT91 USART Controllers as SPI // // Copyright (C) 2018 Microchip Technology Inc. // // Author: Radu Pirea #include #include #include #include #include #include #include #include #include #include #include #include #include #define US_CR 0x00 #define US_MR 0x04 #define US_IER 0x08 #define US_IDR 0x0C #define US_CSR 0x14 #define US_RHR 0x18 #define US_THR 0x1C #define US_BRGR 0x20 #define US_VERSION 0xFC #define US_CR_RSTRX BIT(2) #define US_CR_RSTTX BIT(3) #define US_CR_RXEN BIT(4) #define US_CR_RXDIS BIT(5) #define US_CR_TXEN BIT(6) #define US_CR_TXDIS BIT(7) #define US_MR_SPI_MASTER 0x0E #define US_MR_CHRL GENMASK(7, 6) #define US_MR_CPHA BIT(8) #define US_MR_CPOL BIT(16) #define US_MR_CLKO BIT(18) #define US_MR_WRDBT BIT(20) #define US_MR_LOOP BIT(15) #define US_IR_RXRDY BIT(0) #define US_IR_TXRDY BIT(1) #define US_IR_OVRE BIT(5) #define US_BRGR_SIZE BIT(16) #define US_MIN_CLK_DIV 0x06 #define US_MAX_CLK_DIV BIT(16) #define US_RESET (US_CR_RSTRX | US_CR_RSTTX) #define US_DISABLE (US_CR_RXDIS | US_CR_TXDIS) #define US_ENABLE (US_CR_RXEN | US_CR_TXEN) #define US_OVRE_RXRDY_IRQS (US_IR_OVRE | US_IR_RXRDY) #define US_INIT \ (US_MR_SPI_MASTER | US_MR_CHRL | US_MR_CLKO | US_MR_WRDBT) #define US_DMA_MIN_BYTES 16 #define US_DMA_TIMEOUT (msecs_to_jiffies(1000)) /* Register access macros */ #define at91_usart_spi_readl(port, reg) \ readl_relaxed((port)->regs + US_##reg) #define at91_usart_spi_writel(port, reg, value) \ writel_relaxed((value), (port)->regs + US_##reg) #define at91_usart_spi_readb(port, reg) \ readb_relaxed((port)->regs + US_##reg) #define at91_usart_spi_writeb(port, reg, value) \ writeb_relaxed((value), (port)->regs + US_##reg) struct at91_usart_spi { struct platform_device *mpdev; struct spi_transfer *current_transfer; void __iomem *regs; struct device *dev; struct clk *clk; struct completion xfer_completion; /*used in interrupt to protect data reading*/ spinlock_t lock; phys_addr_t phybase; int irq; unsigned int current_tx_remaining_bytes; unsigned int current_rx_remaining_bytes; u32 spi_clk; u32 status; bool xfer_failed; bool use_dma; }; static void dma_callback(void *data) { struct spi_controller *ctlr = data; struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); at91_usart_spi_writel(aus, IER, US_IR_RXRDY); aus->current_rx_remaining_bytes = 0; complete(&aus->xfer_completion); } static bool at91_usart_spi_can_dma(struct spi_controller *ctrl, struct spi_device *spi, struct spi_transfer *xfer) { struct at91_usart_spi *aus = spi_master_get_devdata(ctrl); return aus->use_dma && xfer->len >= US_DMA_MIN_BYTES; } static int at91_usart_spi_configure_dma(struct spi_controller *ctlr, struct at91_usart_spi *aus) { struct dma_slave_config slave_config; struct device *dev = &aus->mpdev->dev; phys_addr_t phybase = aus->phybase; dma_cap_mask_t mask; int err = 0; dma_cap_zero(mask); dma_cap_set(DMA_SLAVE, mask); ctlr->dma_tx = dma_request_chan(dev, "tx"); if (IS_ERR_OR_NULL(ctlr->dma_tx)) { if (IS_ERR(ctlr->dma_tx)) { err = PTR_ERR(ctlr->dma_tx); goto at91_usart_spi_error_clear; } dev_dbg(dev, "DMA TX channel not available, SPI unable to use DMA\n"); err = -EBUSY; goto at91_usart_spi_error_clear; } ctlr->dma_rx = dma_request_chan(dev, "rx"); if (IS_ERR_OR_NULL(ctlr->dma_rx)) { if (IS_ERR(ctlr->dma_rx)) { err = PTR_ERR(ctlr->dma_rx); goto at91_usart_spi_error; } dev_dbg(dev, "DMA RX channel not available, SPI unable to use DMA\n"); err = -EBUSY; goto at91_usart_spi_error; } slave_config.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; slave_config.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; slave_config.dst_addr = (dma_addr_t)phybase + US_THR; slave_config.src_addr = (dma_addr_t)phybase + US_RHR; slave_config.src_maxburst = 1; slave_config.dst_maxburst = 1; slave_config.device_fc = false; slave_config.direction = DMA_DEV_TO_MEM; if (dmaengine_slave_config(ctlr->dma_rx, &slave_config)) { dev_err(&ctlr->dev, "failed to configure rx dma channel\n"); err = -EINVAL; goto at91_usart_spi_error; } slave_config.direction = DMA_MEM_TO_DEV; if (dmaengine_slave_config(ctlr->dma_tx, &slave_config)) { dev_err(&ctlr->dev, "failed to configure tx dma channel\n"); err = -EINVAL; goto at91_usart_spi_error; } aus->use_dma = true; return 0; at91_usart_spi_error: if (!IS_ERR_OR_NULL(ctlr->dma_tx)) dma_release_channel(ctlr->dma_tx); if (!IS_ERR_OR_NULL(ctlr->dma_rx)) dma_release_channel(ctlr->dma_rx); ctlr->dma_tx = NULL; ctlr->dma_rx = NULL; at91_usart_spi_error_clear: return err; } static void at91_usart_spi_release_dma(struct spi_controller *ctlr) { if (ctlr->dma_rx) dma_release_channel(ctlr->dma_rx); if (ctlr->dma_tx) dma_release_channel(ctlr->dma_tx); } static void at91_usart_spi_stop_dma(struct spi_controller *ctlr) { if (ctlr->dma_rx) dmaengine_terminate_all(ctlr->dma_rx); if (ctlr->dma_tx) dmaengine_terminate_all(ctlr->dma_tx); } static int at91_usart_spi_dma_transfer(struct spi_controller *ctlr, struct spi_transfer *xfer) { struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); struct dma_chan *rxchan = ctlr->dma_rx; struct dma_chan *txchan = ctlr->dma_tx; struct dma_async_tx_descriptor *rxdesc; struct dma_async_tx_descriptor *txdesc; dma_cookie_t cookie; /* Disable RX interrupt */ at91_usart_spi_writel(aus, IDR, US_IR_RXRDY); rxdesc = dmaengine_prep_slave_sg(rxchan, xfer->rx_sg.sgl, xfer->rx_sg.nents, DMA_DEV_TO_MEM, DMA_PREP_INTERRUPT | DMA_CTRL_ACK); if (!rxdesc) goto at91_usart_spi_err_dma; txdesc = dmaengine_prep_slave_sg(txchan, xfer->tx_sg.sgl, xfer->tx_sg.nents, DMA_MEM_TO_DEV, DMA_PREP_INTERRUPT | DMA_CTRL_ACK); if (!txdesc) goto at91_usart_spi_err_dma; rxdesc->callback = dma_callback; rxdesc->callback_param = ctlr; cookie = rxdesc->tx_submit(rxdesc); if (dma_submit_error(cookie)) goto at91_usart_spi_err_dma; cookie = txdesc->tx_submit(txdesc); if (dma_submit_error(cookie)) goto at91_usart_spi_err_dma; rxchan->device->device_issue_pending(rxchan); txchan->device->device_issue_pending(txchan); return 0; at91_usart_spi_err_dma: /* Enable RX interrupt if something fails and fallback to PIO */ at91_usart_spi_writel(aus, IER, US_IR_RXRDY); at91_usart_spi_stop_dma(ctlr); return -ENOMEM; } static unsigned long at91_usart_spi_dma_timeout(struct at91_usart_spi *aus) { return wait_for_completion_timeout(&aus->xfer_completion, US_DMA_TIMEOUT); } static inline u32 at91_usart_spi_tx_ready(struct at91_usart_spi *aus) { return aus->status & US_IR_TXRDY; } static inline u32 at91_usart_spi_rx_ready(struct at91_usart_spi *aus) { return aus->status & US_IR_RXRDY; } static inline u32 at91_usart_spi_check_overrun(struct at91_usart_spi *aus) { return aus->status & US_IR_OVRE; } static inline u32 at91_usart_spi_read_status(struct at91_usart_spi *aus) { aus->status = at91_usart_spi_readl(aus, CSR); return aus->status; } static inline void at91_usart_spi_tx(struct at91_usart_spi *aus) { unsigned int len = aus->current_transfer->len; unsigned int remaining = aus->current_tx_remaining_bytes; const u8 *tx_buf = aus->current_transfer->tx_buf; if (!remaining) return; if (at91_usart_spi_tx_ready(aus)) { at91_usart_spi_writeb(aus, THR, tx_buf[len - remaining]); aus->current_tx_remaining_bytes--; } } static inline void at91_usart_spi_rx(struct at91_usart_spi *aus) { int len = aus->current_transfer->len; int remaining = aus->current_rx_remaining_bytes; u8 *rx_buf = aus->current_transfer->rx_buf; if (!remaining) return; rx_buf[len - remaining] = at91_usart_spi_readb(aus, RHR); aus->current_rx_remaining_bytes--; } static inline void at91_usart_spi_set_xfer_speed(struct at91_usart_spi *aus, struct spi_transfer *xfer) { at91_usart_spi_writel(aus, BRGR, DIV_ROUND_UP(aus->spi_clk, xfer->speed_hz)); } static irqreturn_t at91_usart_spi_interrupt(int irq, void *dev_id) { struct spi_controller *controller = dev_id; struct at91_usart_spi *aus = spi_master_get_devdata(controller); spin_lock(&aus->lock); at91_usart_spi_read_status(aus); if (at91_usart_spi_check_overrun(aus)) { aus->xfer_failed = true; at91_usart_spi_writel(aus, IDR, US_IR_OVRE | US_IR_RXRDY); spin_unlock(&aus->lock); return IRQ_HANDLED; } if (at91_usart_spi_rx_ready(aus)) { at91_usart_spi_rx(aus); spin_unlock(&aus->lock); return IRQ_HANDLED; } spin_unlock(&aus->lock); return IRQ_NONE; } static int at91_usart_spi_setup(struct spi_device *spi) { struct at91_usart_spi *aus = spi_master_get_devdata(spi->controller); u32 *ausd = spi->controller_state; unsigned int mr = at91_usart_spi_readl(aus, MR); if (spi->mode & SPI_CPOL) mr |= US_MR_CPOL; else mr &= ~US_MR_CPOL; if (spi->mode & SPI_CPHA) mr |= US_MR_CPHA; else mr &= ~US_MR_CPHA; if (spi->mode & SPI_LOOP) mr |= US_MR_LOOP; else mr &= ~US_MR_LOOP; if (!ausd) { ausd = kzalloc(sizeof(*ausd), GFP_KERNEL); if (!ausd) return -ENOMEM; spi->controller_state = ausd; } *ausd = mr; dev_dbg(&spi->dev, "setup: bpw %u mode 0x%x -> mr %d %08x\n", spi->bits_per_word, spi->mode, spi->chip_select, mr); return 0; } static int at91_usart_spi_transfer_one(struct spi_controller *ctlr, struct spi_device *spi, struct spi_transfer *xfer) { struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); unsigned long dma_timeout = 0; int ret = 0; at91_usart_spi_set_xfer_speed(aus, xfer); aus->xfer_failed = false; aus->current_transfer = xfer; aus->current_tx_remaining_bytes = xfer->len; aus->current_rx_remaining_bytes = xfer->len; while ((aus->current_tx_remaining_bytes || aus->current_rx_remaining_bytes) && !aus->xfer_failed) { reinit_completion(&aus->xfer_completion); if (at91_usart_spi_can_dma(ctlr, spi, xfer) && !ret) { ret = at91_usart_spi_dma_transfer(ctlr, xfer); if (ret) continue; dma_timeout = at91_usart_spi_dma_timeout(aus); if (WARN_ON(dma_timeout == 0)) { dev_err(&spi->dev, "DMA transfer timeout\n"); return -EIO; } aus->current_tx_remaining_bytes = 0; } else { at91_usart_spi_read_status(aus); at91_usart_spi_tx(aus); } cpu_relax(); } if (aus->xfer_failed) { dev_err(aus->dev, "Overrun!\n"); return -EIO; } return 0; } static int at91_usart_spi_prepare_message(struct spi_controller *ctlr, struct spi_message *message) { struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); struct spi_device *spi = message->spi; u32 *ausd = spi->controller_state; at91_usart_spi_writel(aus, CR, US_ENABLE); at91_usart_spi_writel(aus, IER, US_OVRE_RXRDY_IRQS); at91_usart_spi_writel(aus, MR, *ausd); return 0; } static int at91_usart_spi_unprepare_message(struct spi_controller *ctlr, struct spi_message *message) { struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); at91_usart_spi_writel(aus, CR, US_RESET | US_DISABLE); at91_usart_spi_writel(aus, IDR, US_OVRE_RXRDY_IRQS); return 0; } static void at91_usart_spi_cleanup(struct spi_device *spi) { struct at91_usart_spi_device *ausd = spi->controller_state; spi->controller_state = NULL; kfree(ausd); } static void at91_usart_spi_init(struct at91_usart_spi *aus) { at91_usart_spi_writel(aus, MR, US_INIT); at91_usart_spi_writel(aus, CR, US_RESET | US_DISABLE); } static int at91_usart_gpio_setup(struct platform_device *pdev) { struct device_node *np = pdev->dev.parent->of_node; int i; int ret; int nb; if (!np) return -EINVAL; nb = of_gpio_named_count(np, "cs-gpios"); for (i = 0; i < nb; i++) { int cs_gpio = of_get_named_gpio(np, "cs-gpios", i); if (cs_gpio < 0) return cs_gpio; if (gpio_is_valid(cs_gpio)) { ret = devm_gpio_request_one(&pdev->dev, cs_gpio, GPIOF_DIR_OUT, dev_name(&pdev->dev)); if (ret) return ret; } } return 0; } static int at91_usart_spi_probe(struct platform_device *pdev) { struct resource *regs; struct spi_controller *controller; struct at91_usart_spi *aus; struct clk *clk; int irq; int ret; regs = platform_get_resource(to_platform_device(pdev->dev.parent), IORESOURCE_MEM, 0); if (!regs) return -EINVAL; irq = platform_get_irq(to_platform_device(pdev->dev.parent), 0); if (irq < 0) return irq; clk = devm_clk_get(pdev->dev.parent, "usart"); if (IS_ERR(clk)) return PTR_ERR(clk); ret = -ENOMEM; controller = spi_alloc_master(&pdev->dev, sizeof(*aus)); if (!controller) goto at91_usart_spi_probe_fail; ret = at91_usart_gpio_setup(pdev); if (ret) goto at91_usart_spi_probe_fail; controller->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LOOP | SPI_CS_HIGH; controller->dev.of_node = pdev->dev.parent->of_node; controller->bits_per_word_mask = SPI_BPW_MASK(8); controller->setup = at91_usart_spi_setup; controller->flags = SPI_MASTER_MUST_RX | SPI_MASTER_MUST_TX; controller->transfer_one = at91_usart_spi_transfer_one; controller->prepare_message = at91_usart_spi_prepare_message; controller->unprepare_message = at91_usart_spi_unprepare_message; controller->can_dma = at91_usart_spi_can_dma; controller->cleanup = at91_usart_spi_cleanup; controller->max_speed_hz = DIV_ROUND_UP(clk_get_rate(clk), US_MIN_CLK_DIV); controller->min_speed_hz = DIV_ROUND_UP(clk_get_rate(clk), US_MAX_CLK_DIV); platform_set_drvdata(pdev, controller); aus = spi_master_get_devdata(controller); aus->dev = &pdev->dev; aus->regs = devm_ioremap_resource(&pdev->dev, regs); if (IS_ERR(aus->regs)) { ret = PTR_ERR(aus->regs); goto at91_usart_spi_probe_fail; } aus->irq = irq; aus->clk = clk; ret = devm_request_irq(&pdev->dev, irq, at91_usart_spi_interrupt, 0, dev_name(&pdev->dev), controller); if (ret) goto at91_usart_spi_probe_fail; ret = clk_prepare_enable(clk); if (ret) goto at91_usart_spi_probe_fail; aus->spi_clk = clk_get_rate(clk); at91_usart_spi_init(aus); aus->phybase = regs->start; aus->mpdev = to_platform_device(pdev->dev.parent); ret = at91_usart_spi_configure_dma(controller, aus); if (ret) goto at91_usart_fail_dma; spin_lock_init(&aus->lock); init_completion(&aus->xfer_completion); ret = devm_spi_register_master(&pdev->dev, controller); if (ret) goto at91_usart_fail_register_master; dev_info(&pdev->dev, "AT91 USART SPI Controller version 0x%x at %pa (irq %d)\n", at91_usart_spi_readl(aus, VERSION), ®s->start, irq); return 0; at91_usart_fail_register_master: at91_usart_spi_release_dma(controller); at91_usart_fail_dma: clk_disable_unprepare(clk); at91_usart_spi_probe_fail: spi_master_put(controller); return ret; } __maybe_unused static int at91_usart_spi_runtime_suspend(struct device *dev) { struct spi_controller *ctlr = dev_get_drvdata(dev); struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); clk_disable_unprepare(aus->clk); pinctrl_pm_select_sleep_state(dev); return 0; } __maybe_unused static int at91_usart_spi_runtime_resume(struct device *dev) { struct spi_controller *ctrl = dev_get_drvdata(dev); struct at91_usart_spi *aus = spi_master_get_devdata(ctrl); pinctrl_pm_select_default_state(dev); return clk_prepare_enable(aus->clk); } __maybe_unused static int at91_usart_spi_suspend(struct device *dev) { struct spi_controller *ctrl = dev_get_drvdata(dev); int ret; ret = spi_controller_suspend(ctrl); if (ret) return ret; if (!pm_runtime_suspended(dev)) at91_usart_spi_runtime_suspend(dev); return 0; } __maybe_unused static int at91_usart_spi_resume(struct device *dev) { struct spi_controller *ctrl = dev_get_drvdata(dev); struct at91_usart_spi *aus = spi_master_get_devdata(ctrl); int ret; if (!pm_runtime_suspended(dev)) { ret = at91_usart_spi_runtime_resume(dev); if (ret) return ret; } at91_usart_spi_init(aus); return spi_controller_resume(ctrl); } static int at91_usart_spi_remove(struct platform_device *pdev) { struct spi_controller *ctlr = platform_get_drvdata(pdev); struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); at91_usart_spi_release_dma(ctlr); clk_disable_unprepare(aus->clk); return 0; } static const struct dev_pm_ops at91_usart_spi_pm_ops = { SET_SYSTEM_SLEEP_PM_OPS(at91_usart_spi_suspend, at91_usart_spi_resume) SET_RUNTIME_PM_OPS(at91_usart_spi_runtime_suspend, at91_usart_spi_runtime_resume, NULL) }; static const struct of_device_id at91_usart_spi_dt_ids[] = { { .compatible = "microchip,at91sam9g45-usart-spi"}, { /* sentinel */} }; MODULE_DEVICE_TABLE(of, at91_usart_spi_dt_ids); static struct platform_driver at91_usart_spi_driver = { .driver = { .name = "at91_usart_spi", .pm = &at91_usart_spi_pm_ops, }, .probe = at91_usart_spi_probe, .remove = at91_usart_spi_remove, }; module_platform_driver(at91_usart_spi_driver); MODULE_DESCRIPTION("Microchip AT91 USART SPI Controller driver"); MODULE_AUTHOR("Radu Pirea "); MODULE_LICENSE("GPL v2"); MODULE_ALIAS("platform:at91_usart_spi");