/* * The driver for Freescale MPC512x LocalPlus Bus FIFO * (called SCLPC in the Reference Manual). * * Copyright (C) 2013-2015 Alexander Popov . * * This file is released under the GPLv2. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #define DRV_NAME "mpc512x_lpbfifo" struct cs_range { u32 csnum; u32 base; /* must be zero */ u32 addr; u32 size; }; static struct lpbfifo_data { spinlock_t lock; /* for protecting lpbfifo_data */ phys_addr_t regs_phys; resource_size_t regs_size; struct mpc512x_lpbfifo __iomem *regs; int irq; struct cs_range *cs_ranges; size_t cs_n; struct dma_chan *chan; struct mpc512x_lpbfifo_request *req; dma_addr_t ram_bus_addr; bool wait_lpbfifo_irq; bool wait_lpbfifo_callback; } lpbfifo; /* * A data transfer from RAM to some device on LPB is finished * when both mpc512x_lpbfifo_irq() and mpc512x_lpbfifo_callback() * have been called. We execute the callback registered in * mpc512x_lpbfifo_request just after that. * But for a data transfer from some device on LPB to RAM we don't enable * LPBFIFO interrupt because clearing MPC512X_SCLPC_SUCCESS interrupt flag * automatically disables LPBFIFO reading request to the DMA controller * and the data transfer hangs. So the callback registered in * mpc512x_lpbfifo_request is executed at the end of mpc512x_lpbfifo_callback(). */ /* * mpc512x_lpbfifo_irq - IRQ handler for LPB FIFO */ static irqreturn_t mpc512x_lpbfifo_irq(int irq, void *param) { struct device *dev = (struct device *)param; struct mpc512x_lpbfifo_request *req = NULL; unsigned long flags; u32 status; spin_lock_irqsave(&lpbfifo.lock, flags); if (!lpbfifo.regs) goto end; req = lpbfifo.req; if (!req || req->dir == MPC512X_LPBFIFO_REQ_DIR_READ) { dev_err(dev, "bogus LPBFIFO IRQ\n"); goto end; } status = in_be32(&lpbfifo.regs->status); if (status != MPC512X_SCLPC_SUCCESS) { dev_err(dev, "DMA transfer from RAM to peripheral failed\n"); out_be32(&lpbfifo.regs->enable, MPC512X_SCLPC_RESET | MPC512X_SCLPC_FIFO_RESET); goto end; } /* Clear the interrupt flag */ out_be32(&lpbfifo.regs->status, MPC512X_SCLPC_SUCCESS); lpbfifo.wait_lpbfifo_irq = false; if (lpbfifo.wait_lpbfifo_callback) goto end; /* Transfer is finished, set the FIFO as idle */ lpbfifo.req = NULL; spin_unlock_irqrestore(&lpbfifo.lock, flags); if (req->callback) req->callback(req); return IRQ_HANDLED; end: spin_unlock_irqrestore(&lpbfifo.lock, flags); return IRQ_HANDLED; } /* * mpc512x_lpbfifo_callback is called by DMA driver when * DMA transaction is finished. */ static void mpc512x_lpbfifo_callback(void *param) { unsigned long flags; struct mpc512x_lpbfifo_request *req = NULL; enum dma_data_direction dir; spin_lock_irqsave(&lpbfifo.lock, flags); if (!lpbfifo.regs) { spin_unlock_irqrestore(&lpbfifo.lock, flags); return; } req = lpbfifo.req; if (!req) { pr_err("bogus LPBFIFO callback\n"); spin_unlock_irqrestore(&lpbfifo.lock, flags); return; } /* Release the mapping */ if (req->dir == MPC512X_LPBFIFO_REQ_DIR_WRITE) dir = DMA_TO_DEVICE; else dir = DMA_FROM_DEVICE; dma_unmap_single(lpbfifo.chan->device->dev, lpbfifo.ram_bus_addr, req->size, dir); lpbfifo.wait_lpbfifo_callback = false; if (!lpbfifo.wait_lpbfifo_irq) { /* Transfer is finished, set the FIFO as idle */ lpbfifo.req = NULL; spin_unlock_irqrestore(&lpbfifo.lock, flags); if (req->callback) req->callback(req); } else { spin_unlock_irqrestore(&lpbfifo.lock, flags); } } static int mpc512x_lpbfifo_kick(void) { u32 bits; bool no_incr = false; u32 bpt = 32; /* max bytes per LPBFIFO transaction involving DMA */ u32 cs = 0; size_t i; struct dma_device *dma_dev = NULL; struct scatterlist sg; enum dma_data_direction dir; struct dma_slave_config dma_conf = {}; struct dma_async_tx_descriptor *dma_tx = NULL; dma_cookie_t cookie; int ret; /* * 1. Fit the requirements: * - the packet size must be a multiple of 4 since FIFO Data Word * Register allows only full-word access according the Reference * Manual; * - the physical address of the device on LPB and the packet size * must be aligned on BPT (bytes per transaction) or 8-bytes * boundary according the Reference Manual; * - but we choose DMA maxburst equal (or very close to) BPT to prevent * DMA controller from overtaking FIFO and causing FIFO underflow * error. So we force the packet size to be aligned on BPT boundary * not to confuse DMA driver which requires the packet size to be * aligned on maxburst boundary; * - BPT should be set to the LPB device port size for operation with * disabled auto-incrementing according Reference Manual. */ if (lpbfifo.req->size == 0 || !IS_ALIGNED(lpbfifo.req->size, 4)) return -EINVAL; if (lpbfifo.req->portsize != LPB_DEV_PORTSIZE_UNDEFINED) { bpt = lpbfifo.req->portsize; no_incr = true; } while (bpt > 1) { if (IS_ALIGNED(lpbfifo.req->dev_phys_addr, min(bpt, 0x8u)) && IS_ALIGNED(lpbfifo.req->size, bpt)) { break; } if (no_incr) return -EINVAL; bpt >>= 1; } dma_conf.dst_maxburst = max(bpt, 0x4u) / 4; dma_conf.src_maxburst = max(bpt, 0x4u) / 4; for (i = 0; i < lpbfifo.cs_n; i++) { phys_addr_t cs_start = lpbfifo.cs_ranges[i].addr; phys_addr_t cs_end = cs_start + lpbfifo.cs_ranges[i].size; phys_addr_t access_start = lpbfifo.req->dev_phys_addr; phys_addr_t access_end = access_start + lpbfifo.req->size; if (access_start >= cs_start && access_end <= cs_end) { cs = lpbfifo.cs_ranges[i].csnum; break; } } if (i == lpbfifo.cs_n) return -EFAULT; /* 2. Prepare DMA */ dma_dev = lpbfifo.chan->device; if (lpbfifo.req->dir == MPC512X_LPBFIFO_REQ_DIR_WRITE) { dir = DMA_TO_DEVICE; dma_conf.direction = DMA_MEM_TO_DEV; dma_conf.dst_addr = lpbfifo.regs_phys + offsetof(struct mpc512x_lpbfifo, data_word); } else { dir = DMA_FROM_DEVICE; dma_conf.direction = DMA_DEV_TO_MEM; dma_conf.src_addr = lpbfifo.regs_phys + offsetof(struct mpc512x_lpbfifo, data_word); } dma_conf.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; dma_conf.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; /* Make DMA channel work with LPB FIFO data register */ if (dma_dev->device_config(lpbfifo.chan, &dma_conf)) { ret = -EINVAL; goto err_dma_prep; } sg_init_table(&sg, 1); sg_dma_address(&sg) = dma_map_single(dma_dev->dev, lpbfifo.req->ram_virt_addr, lpbfifo.req->size, dir); if (dma_mapping_error(dma_dev->dev, sg_dma_address(&sg))) return -EFAULT; lpbfifo.ram_bus_addr = sg_dma_address(&sg); /* For freeing later */ sg_dma_len(&sg) = lpbfifo.req->size; dma_tx = dmaengine_prep_slave_sg(lpbfifo.chan, &sg, 1, dma_conf.direction, 0); if (!dma_tx) { ret = -ENOSPC; goto err_dma_prep; } dma_tx->callback = mpc512x_lpbfifo_callback; dma_tx->callback_param = NULL; /* 3. Prepare FIFO */ out_be32(&lpbfifo.regs->enable, MPC512X_SCLPC_RESET | MPC512X_SCLPC_FIFO_RESET); out_be32(&lpbfifo.regs->enable, 0x0); /* * Configure the watermarks for write operation (RAM->DMA->FIFO->dev): * - high watermark 7 words according the Reference Manual, * - low watermark 512 bytes (half of the FIFO). * These watermarks don't work for read operation since the * MPC512X_SCLPC_FLUSH bit is set (according the Reference Manual). */ out_be32(&lpbfifo.regs->fifo_ctrl, MPC512X_SCLPC_FIFO_CTRL(0x7)); out_be32(&lpbfifo.regs->fifo_alarm, MPC512X_SCLPC_FIFO_ALARM(0x200)); /* * Start address is a physical address of the region which belongs * to the device on the LocalPlus Bus */ out_be32(&lpbfifo.regs->start_addr, lpbfifo.req->dev_phys_addr); /* * Configure chip select, transfer direction, address increment option * and bytes per transaction option */ bits = MPC512X_SCLPC_CS(cs); if (lpbfifo.req->dir == MPC512X_LPBFIFO_REQ_DIR_READ) bits |= MPC512X_SCLPC_READ | MPC512X_SCLPC_FLUSH; if (no_incr) bits |= MPC512X_SCLPC_DAI; bits |= MPC512X_SCLPC_BPT(bpt); out_be32(&lpbfifo.regs->ctrl, bits); /* Unmask irqs */ bits = MPC512X_SCLPC_ENABLE | MPC512X_SCLPC_ABORT_INT_ENABLE; if (lpbfifo.req->dir == MPC512X_LPBFIFO_REQ_DIR_WRITE) bits |= MPC512X_SCLPC_NORM_INT_ENABLE; else lpbfifo.wait_lpbfifo_irq = false; out_be32(&lpbfifo.regs->enable, bits); /* 4. Set packet size and kick FIFO off */ bits = lpbfifo.req->size | MPC512X_SCLPC_START; out_be32(&lpbfifo.regs->pkt_size, bits); /* 5. Finally kick DMA off */ cookie = dma_tx->tx_submit(dma_tx); if (dma_submit_error(cookie)) { ret = -ENOSPC; goto err_dma_submit; } return 0; err_dma_submit: out_be32(&lpbfifo.regs->enable, MPC512X_SCLPC_RESET | MPC512X_SCLPC_FIFO_RESET); err_dma_prep: dma_unmap_single(dma_dev->dev, sg_dma_address(&sg), lpbfifo.req->size, dir); return ret; } static int mpc512x_lpbfifo_submit_locked(struct mpc512x_lpbfifo_request *req) { int ret = 0; if (!lpbfifo.regs) return -ENODEV; /* Check whether a transfer is in progress */ if (lpbfifo.req) return -EBUSY; lpbfifo.wait_lpbfifo_irq = true; lpbfifo.wait_lpbfifo_callback = true; lpbfifo.req = req; ret = mpc512x_lpbfifo_kick(); if (ret != 0) lpbfifo.req = NULL; /* Set the FIFO as idle */ return ret; } int mpc512x_lpbfifo_submit(struct mpc512x_lpbfifo_request *req) { unsigned long flags; int ret = 0; spin_lock_irqsave(&lpbfifo.lock, flags); ret = mpc512x_lpbfifo_submit_locked(req); spin_unlock_irqrestore(&lpbfifo.lock, flags); return ret; } EXPORT_SYMBOL(mpc512x_lpbfifo_submit); /* * LPBFIFO driver uses "ranges" property of "localbus" device tree node * for being able to determine the chip select number of a client device * ordering a DMA transfer. */ static int get_cs_ranges(struct device *dev) { int ret = -ENODEV; struct device_node *lb_node; const u32 *addr_cells_p; const u32 *size_cells_p; int proplen; size_t i; lb_node = of_find_compatible_node(NULL, NULL, "fsl,mpc5121-localbus"); if (!lb_node) return ret; /* * The node defined as compatible with 'fsl,mpc5121-localbus' * should have two address cells and one size cell. * Every item of its ranges property should consist of: * - the first address cell which is the chipselect number; * - the second address cell which is the offset in the chipselect, * must be zero. * - CPU address of the beginning of an access window; * - the only size cell which is the size of an access window. */ addr_cells_p = of_get_property(lb_node, "#address-cells", NULL); size_cells_p = of_get_property(lb_node, "#size-cells", NULL); if (addr_cells_p == NULL || *addr_cells_p != 2 || size_cells_p == NULL || *size_cells_p != 1) { goto end; } proplen = of_property_count_u32_elems(lb_node, "ranges"); if (proplen <= 0 || proplen % 4 != 0) goto end; lpbfifo.cs_n = proplen / 4; lpbfifo.cs_ranges = devm_kcalloc(dev, lpbfifo.cs_n, sizeof(struct cs_range), GFP_KERNEL); if (!lpbfifo.cs_ranges) goto end; if (of_property_read_u32_array(lb_node, "ranges", (u32 *)lpbfifo.cs_ranges, proplen) != 0) { goto end; } for (i = 0; i < lpbfifo.cs_n; i++) { if (lpbfifo.cs_ranges[i].base != 0) goto end; } ret = 0; end: of_node_put(lb_node); return ret; } static int mpc512x_lpbfifo_probe(struct platform_device *pdev) { struct resource r; int ret = 0; memset(&lpbfifo, 0, sizeof(struct lpbfifo_data)); spin_lock_init(&lpbfifo.lock); lpbfifo.chan = dma_request_slave_channel(&pdev->dev, "rx-tx"); if (lpbfifo.chan == NULL) return -EPROBE_DEFER; if (of_address_to_resource(pdev->dev.of_node, 0, &r) != 0) { dev_err(&pdev->dev, "bad 'reg' in 'sclpc' device tree node\n"); ret = -ENODEV; goto err0; } lpbfifo.regs_phys = r.start; lpbfifo.regs_size = resource_size(&r); if (!devm_request_mem_region(&pdev->dev, lpbfifo.regs_phys, lpbfifo.regs_size, DRV_NAME)) { dev_err(&pdev->dev, "unable to request region\n"); ret = -EBUSY; goto err0; } lpbfifo.regs = devm_ioremap(&pdev->dev, lpbfifo.regs_phys, lpbfifo.regs_size); if (!lpbfifo.regs) { dev_err(&pdev->dev, "mapping registers failed\n"); ret = -ENOMEM; goto err0; } out_be32(&lpbfifo.regs->enable, MPC512X_SCLPC_RESET | MPC512X_SCLPC_FIFO_RESET); if (get_cs_ranges(&pdev->dev) != 0) { dev_err(&pdev->dev, "bad '/localbus' device tree node\n"); ret = -ENODEV; goto err0; } lpbfifo.irq = irq_of_parse_and_map(pdev->dev.of_node, 0); if (!lpbfifo.irq) { dev_err(&pdev->dev, "mapping irq failed\n"); ret = -ENODEV; goto err0; } if (request_irq(lpbfifo.irq, mpc512x_lpbfifo_irq, 0, DRV_NAME, &pdev->dev) != 0) { dev_err(&pdev->dev, "requesting irq failed\n"); ret = -ENODEV; goto err1; } dev_info(&pdev->dev, "probe succeeded\n"); return 0; err1: irq_dispose_mapping(lpbfifo.irq); err0: dma_release_channel(lpbfifo.chan); return ret; } static int mpc512x_lpbfifo_remove(struct platform_device *pdev) { unsigned long flags; struct dma_device *dma_dev = lpbfifo.chan->device; struct mpc512x_lpbfifo __iomem *regs = NULL; spin_lock_irqsave(&lpbfifo.lock, flags); regs = lpbfifo.regs; lpbfifo.regs = NULL; spin_unlock_irqrestore(&lpbfifo.lock, flags); dma_dev->device_terminate_all(lpbfifo.chan); out_be32(®s->enable, MPC512X_SCLPC_RESET | MPC512X_SCLPC_FIFO_RESET); free_irq(lpbfifo.irq, &pdev->dev); irq_dispose_mapping(lpbfifo.irq); dma_release_channel(lpbfifo.chan); return 0; } static const struct of_device_id mpc512x_lpbfifo_match[] = { { .compatible = "fsl,mpc512x-lpbfifo", }, {}, }; MODULE_DEVICE_TABLE(of, mpc512x_lpbfifo_match); static struct platform_driver mpc512x_lpbfifo_driver = { .probe = mpc512x_lpbfifo_probe, .remove = mpc512x_lpbfifo_remove, .driver = { .name = DRV_NAME, .of_match_table = mpc512x_lpbfifo_match, }, }; module_platform_driver(mpc512x_lpbfifo_driver); MODULE_AUTHOR("Alexander Popov "); MODULE_DESCRIPTION("MPC512x LocalPlus Bus FIFO device driver"); MODULE_LICENSE("GPL v2");