diff options
author | 2020-05-22 08:48:49 +0000 | |
---|---|---|
committer | 2020-05-22 08:48:49 +0000 | |
commit | a69c39e4dda6bcd002b2110ac97ebdb484b5624b (patch) | |
tree | 15c111bf59f7087b470c0785bdeda82e535a7234 /sys | |
parent | Add test covering revision 1.64 of c_sh, fix exit code of compound lists (diff) | |
download | wireguard-openbsd-a69c39e4dda6bcd002b2110ac97ebdb484b5624b.tar.xz wireguard-openbsd-a69c39e4dda6bcd002b2110ac97ebdb484b5624b.zip |
Add mvkpcie(4), a driver for the Aardvark PCIe controller found on
the Armada 3700 SoC. So far only MSI is supported, since it was
easier to implement. Both MSI and INTx share a single interrupt
pin, so there's nothing to gain anyway, apart from legacy device
support.
With this I can push traffic through a bwfm(4) in my Turris Mox.
ok kettenis@
Diffstat (limited to 'sys')
-rw-r--r-- | sys/dev/fdt/files.fdt | 6 | ||||
-rw-r--r-- | sys/dev/fdt/mvkpcie.c | 1026 |
2 files changed, 1031 insertions, 1 deletions
diff --git a/sys/dev/fdt/files.fdt b/sys/dev/fdt/files.fdt index 364e6a28e1c..46f1cd15449 100644 --- a/sys/dev/fdt/files.fdt +++ b/sys/dev/fdt/files.fdt @@ -1,4 +1,4 @@ -# $OpenBSD: files.fdt,v 1.133 2020/05/18 10:39:33 kettenis Exp $ +# $OpenBSD: files.fdt,v 1.134 2020/05/22 08:48:49 patrick Exp $ # # Config file and device description for machine-independent FDT code. # Included by ports that need it. @@ -374,6 +374,10 @@ device mviic: i2cbus attach mviic at fdt file dev/fdt/mviic.c mviic +device mvkpcie: pcibus +attach mvkpcie at fdt +file dev/fdt/mvkpcie.c mvkpcie + device mvpinctrl attach mvpinctrl at fdt file dev/fdt/mvpinctrl.c mvpinctrl diff --git a/sys/dev/fdt/mvkpcie.c b/sys/dev/fdt/mvkpcie.c new file mode 100644 index 00000000000..387d57cec01 --- /dev/null +++ b/sys/dev/fdt/mvkpcie.c @@ -0,0 +1,1026 @@ +/* $OpenBSD: mvkpcie.c,v 1.1 2020/05/22 08:48:49 patrick Exp $ */ +/* + * Copyright (c) 2018 Mark Kettenis <kettenis@openbsd.org> + * Copyright (c) 2020 Patrick Wildt <patrick@blueri.se> + * + * Permission to use, copy, modify, and distribute this software for any + * purpose with or without fee is hereby granted, provided that the above + * copyright notice and this permission notice appear in all copies. + * + * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES + * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF + * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR + * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES + * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN + * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF + * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. + */ + +#include <sys/param.h> +#include <sys/systm.h> +#include <sys/device.h> +#include <sys/extent.h> +#include <sys/malloc.h> +#include <sys/evcount.h> + +#include <machine/intr.h> +#include <machine/bus.h> +#include <machine/fdt.h> + +#include <dev/pci/pcidevs.h> +#include <dev/pci/pcireg.h> +#include <dev/pci/pcivar.h> +#include <dev/pci/ppbreg.h> + +#include <dev/ofw/openfirm.h> +#include <dev/ofw/ofw_clock.h> +#include <dev/ofw/ofw_gpio.h> +#include <dev/ofw/ofw_misc.h> +#include <dev/ofw/ofw_pinctrl.h> +#include <dev/ofw/ofw_power.h> +#include <dev/ofw/fdt.h> + +/* Registers */ +#define PCIE_DEV_ID 0x0000 +#define PCIE_CMD 0x0004 +#define PCIE_DEV_REV 0x0008 +#define PCIE_DEV_CTRL_STATS 0x00c8 +#define PCIE_DEV_CTRL_STATS_SNOOP (1 << 1) +#define PCIE_DEV_CTRL_STATS_RELAX_ORDER (1 << 4) +#define PCIE_DEV_CTRL_STATS_MAX_PAYLOAD_7 (0x7 << 5) +#define PCIE_DEV_CTRL_STATS_MAX_RD_REQ_SZ (0x2 << 12) +#define PCIE_LINK_CTRL_STAT 0x00d0 +#define PCIE_LINK_CTRL_STAT_LINK_L0S_ENTRY (1 << 0) +#define PCIE_LINK_CTRL_STAT_LINK_TRAINING (1 << 5) +#define PCIE_LINK_CTRL_STAT_LINK_WIDTH_1 (1 << 20) +#define PCIE_ERR_CAPCTL 0x0118 +#define PCIE_ERR_CAPCTL_ECRC_CHK_TX (1 << 5) +#define PCIE_ERR_CAPCTL_ECRC_CHK_TX_EN (1 << 6) +#define PCIE_ERR_CAPCTL_ECRC_CHCK (1 << 7) +#define PCIE_ERR_CAPCTL_ECRC_CHCK_RCV (1 << 8) +#define PIO_CTRL 0x4000 +#define PIO_CTRL_TYPE_MASK (0xf << 0) +#define PIO_CTRL_TYPE_RD0 (0x8 << 0) +#define PIO_CTRL_TYPE_RD1 (0x9 << 0) +#define PIO_CTRL_TYPE_WR0 (0xa << 0) +#define PIO_CTRL_TYPE_WR1 (0xb << 0) +#define PIO_CTRL_ADDR_WIN_DISABLE (1 << 24) +#define PIO_STAT 0x4004 +#define PIO_STAT_COMP_STATUS (0x7 << 7) +#define PIO_ADDR_LS 0x4008 +#define PIO_ADDR_MS 0x400c +#define PIO_WR_DATA 0x4010 +#define PIO_WR_DATA_STRB 0x4014 +#define PIO_WR_DATA_STRB_VALUE 0xf +#define PIO_RD_DATA 0x4018 +#define PIO_START 0x401c +#define PIO_START_STOP (0 << 0) +#define PIO_START_START (1 << 0) +#define PIO_ISR 0x4020 +#define PIO_ISR_CLEAR (1 << 0) +#define PIO_ISRM 0x4024 +#define PCIE_CORE_CTRL0 0x4800 +#define PCIE_CORE_CTRL0_GEN_1 (0 << 0) +#define PCIE_CORE_CTRL0_GEN_2 (1 << 0) +#define PCIE_CORE_CTRL0_GEN_3 (2 << 0) +#define PCIE_CORE_CTRL0_GEN_MASK (0x3 << 0) +#define PCIE_CORE_CTRL0_IS_RC (1 << 2) +#define PCIE_CORE_CTRL0_LANE_1 (0 << 3) +#define PCIE_CORE_CTRL0_LANE_2 (1 << 3) +#define PCIE_CORE_CTRL0_LANE_4 (2 << 3) +#define PCIE_CORE_CTRL0_LANE_8 (3 << 3) +#define PCIE_CORE_CTRL0_LANE_MASK (0x3 << 3) +#define PCIE_CORE_CTRL0_LINK_TRAINING (1 << 6) +#define PCIE_CORE_CTRL2 0x4808 +#define PCIE_CORE_CTRL2_RESERVED (0x7 << 0) +#define PCIE_CORE_CTRL2_TD_ENABLE (1 << 4) +#define PCIE_CORE_CTRL2_STRICT_ORDER_ENABLE (1 << 5) +#define PCIE_CORE_CTRL2_OB_WIN_ENABLE (1 << 6) +#define PCIE_CORE_CTRL2_MSI_ENABLE (1 << 10) +#define PCIE_CORE_ISR0_STATUS 0x4840 +#define PCIE_CORE_ISR0_MASK 0x4844 +#define PCIE_CORE_ISR0_MASK_MSI_INT (1 << 24) +#define PCIE_CORE_ISR0_MASK_ALL 0x07ffffff +#define PCIE_CORE_ISR1_STATUS 0x4848 +#define PCIE_CORE_ISR1_MASK 0x484c +#define PCIE_CORE_ISR1_MASK_ALL 0x00000ff0 +#define PCIE_CORE_ISR1_MASK_INTX(x) (1 << (x + 8)) +#define PCIE_CORE_MSI_ADDR_LOW 0x4850 +#define PCIE_CORE_MSI_ADDR_HIGH 0x4854 +#define PCIE_CORE_MSI_STATUS 0x4858 +#define PCIE_CORE_MSI_MASK 0x485c +#define PCIE_CORE_MSI_PAYLOAD 0x489c +#define LMI_CFG 0x6000 +#define LMI_CFG_LTSSM_VAL(x) (((x) >> 24) & 0x3f) +#define LMI_CFG_LTSSM_L0 0x10 +#define CTRL_CORE_CONFIG 0x18000 +#define CTRL_CORE_CONFIG_MODE_DIRECT (0 << 0) +#define CTRL_CORE_CONFIG_MODE_COMMAND (1 << 0) +#define CTRL_CORE_CONFIG_MODE_MASK (1 << 0) +#define HOST_CTRL_INT_STATUS 0x1b000 +#define HOST_CTRL_INT_MASK 0x1b004 +#define HOST_CTRL_INT_MASK_CORE_INT (1 << 16) +#define HOST_CTRL_INT_MASK_ALL 0xfff0fb + +#define HREAD4(sc, reg) \ + (bus_space_read_4((sc)->sc_iot, (sc)->sc_ioh, (reg))) +#define HWRITE4(sc, reg, val) \ + bus_space_write_4((sc)->sc_iot, (sc)->sc_ioh, (reg), (val)) +#define HSET4(sc, reg, bits) \ + HWRITE4((sc), (reg), HREAD4((sc), (reg)) | (bits)) +#define HCLR4(sc, reg, bits) \ + HWRITE4((sc), (reg), HREAD4((sc), (reg)) & ~(bits)) + +struct mvkpcie_dmamem { + bus_dmamap_t mdm_map; + bus_dma_segment_t mdm_seg; + size_t mdm_size; + caddr_t mdm_kva; +}; + +#define MVKPCIE_DMA_MAP(_mdm) ((_mdm)->mdm_map) +#define MVKPCIE_DMA_LEN(_mdm) ((_mdm)->mdm_size) +#define MVKPCIE_DMA_DVA(_mdm) ((uint64_t)(_mdm)->mdm_map->dm_segs[0].ds_addr) +#define MVKPCIE_DMA_KVA(_mdm) ((void *)(_mdm)->mdm_kva) + +struct intrhand { + int (*ih_func)(void *); /* handler */ + void *ih_arg; /* arg for handler */ + int ih_ipl; /* IPL_* */ + int ih_irq; /* IRQ number */ + struct evcount ih_count; + char *ih_name; + void *ih_sc; +}; + +struct mvkpcie_range { + uint32_t flags; + uint64_t pci_base; + uint64_t phys_base; + uint64_t size; +}; + +struct mvkpcie_softc { + struct device sc_dev; + bus_space_tag_t sc_iot; + bus_space_handle_t sc_ioh; + bus_dma_tag_t sc_dmat; + + bus_addr_t sc_io_base; + bus_addr_t sc_io_bus_addr; + bus_size_t sc_io_size; + bus_addr_t sc_mem_base; + bus_addr_t sc_mem_bus_addr; + bus_size_t sc_mem_size; + + int sc_node; + int sc_acells; + int sc_scells; + int sc_pacells; + int sc_pscells; + struct mvkpcie_range *sc_ranges; + int sc_nranges; + + struct bus_space sc_bus_iot; + struct bus_space sc_bus_memt; + + struct arm64_pci_chipset sc_pc; + int sc_bus; + + uint32_t sc_bridge_command; + uint32_t sc_bridge_bar0; + uint32_t sc_bridge_bar1; + uint32_t sc_bridge_businfo; + uint32_t sc_bridge_iostatus; + uint32_t sc_bridge_io_hi; + uint32_t sc_bridge_mem; + + struct interrupt_controller sc_msi_ic; + struct intrhand *sc_msi_handlers[32]; + struct mvkpcie_dmamem *sc_msi_addr; + void *sc_ih; + int sc_ipl; +}; + +int mvkpcie_match(struct device *, void *, void *); +void mvkpcie_attach(struct device *, struct device *, void *); + +struct cfattach mvkpcie_ca = { + sizeof (struct mvkpcie_softc), mvkpcie_match, mvkpcie_attach +}; + +struct cfdriver mvkpcie_cd = { + NULL, "mvkpcie", DV_DULL +}; + +int +mvkpcie_match(struct device *parent, void *match, void *aux) +{ + struct fdt_attach_args *faa = aux; + + return OF_is_compatible(faa->fa_node, "marvell,armada-3700-pcie"); +} + +int mvkpcie_link_up(struct mvkpcie_softc *); + +void mvkpcie_attach_hook(struct device *, struct device *, + struct pcibus_attach_args *); +int mvkpcie_bus_maxdevs(void *, int); +pcitag_t mvkpcie_make_tag(void *, int, int, int); +void mvkpcie_decompose_tag(void *, pcitag_t, int *, int *, int *); +int mvkpcie_conf_size(void *, pcitag_t); +pcireg_t mvkpcie_conf_read(void *, pcitag_t, int); +void mvkpcie_conf_write(void *, pcitag_t, int, pcireg_t); + +int mvkpcie_intr_map(struct pci_attach_args *, pci_intr_handle_t *); +const char *mvkpcie_intr_string(void *, pci_intr_handle_t); +void *mvkpcie_intr_establish(void *, pci_intr_handle_t, int, + int (*)(void *), void *, char *); +void mvkpcie_intr_disestablish(void *, void *); + +void *mvkpcie_intr_establish(void *, pci_intr_handle_t, int, + int (*)(void *), void *, char *); + +int mvkpcie_bs_iomap(bus_space_tag_t, bus_addr_t, bus_size_t, int, + bus_space_handle_t *); +int mvkpcie_bs_memmap(bus_space_tag_t, bus_addr_t, bus_size_t, int, + bus_space_handle_t *); + +int mvkpcie_intc_intr(void *); +void *mvkpcie_intc_intr_establish_msi(void *, uint64_t *, uint64_t *, + int , int (*)(void *), void *, char *); +void mvkpcie_intc_intr_disestablish_msi(void *); +void mvkpcie_intc_recalc_ipl(struct mvkpcie_softc *); + +struct mvkpcie_dmamem *mvkpcie_dmamem_alloc(struct mvkpcie_softc *, bus_size_t, + bus_size_t); +void mvkpcie_dmamem_free(struct mvkpcie_softc *, struct mvkpcie_dmamem *); + +void +mvkpcie_attach(struct device *parent, struct device *self, void *aux) +{ + struct mvkpcie_softc *sc = (struct mvkpcie_softc *)self; + struct fdt_attach_args *faa = aux; + struct pcibus_attach_args pba; + bus_addr_t iobase, iolimit; + bus_addr_t membase, memlimit; + uint32_t bus_range[2]; + uint32_t *ranges; + int i, j, nranges, rangeslen; + pcireg_t bir, blr; + uint32_t reg; + int timo; + + if (faa->fa_nreg < 1) { + printf(": no registers\n"); + return; + } + + sc->sc_iot = faa->fa_iot; + sc->sc_dmat = faa->fa_dmat; + sc->sc_node = faa->fa_node; + + sc->sc_acells = OF_getpropint(sc->sc_node, "#address-cells", + faa->fa_acells); + sc->sc_scells = OF_getpropint(sc->sc_node, "#size-cells", + faa->fa_scells); + sc->sc_pacells = faa->fa_acells; + sc->sc_pscells = faa->fa_scells; + + rangeslen = OF_getproplen(sc->sc_node, "ranges"); + if (rangeslen <= 0 || (rangeslen % sizeof(uint32_t)) || + (rangeslen / sizeof(uint32_t)) % (sc->sc_acells + + sc->sc_pacells + sc->sc_scells)) { + printf(": invalid ranges property\n"); + return; + } + + sc->sc_msi_addr = mvkpcie_dmamem_alloc(sc, sizeof(uint16_t), + sizeof(uint64_t)); + if (sc->sc_msi_addr == NULL) { + printf(": cannot allocate MSI address\n"); + return; + } + + ranges = malloc(rangeslen, M_TEMP, M_WAITOK); + OF_getpropintarray(sc->sc_node, "ranges", ranges, + rangeslen); + + nranges = (rangeslen / sizeof(uint32_t)) / + (sc->sc_acells + sc->sc_pacells + sc->sc_scells); + sc->sc_ranges = mallocarray(nranges, + sizeof(struct mvkpcie_range), M_TEMP, M_WAITOK); + sc->sc_nranges = nranges; + + for (i = 0, j = 0; i < sc->sc_nranges; i++) { + sc->sc_ranges[i].flags = ranges[j++]; + sc->sc_ranges[i].pci_base = ranges[j++]; + if (sc->sc_acells - 1 == 2) { + sc->sc_ranges[i].pci_base <<= 32; + sc->sc_ranges[i].pci_base |= ranges[j++]; + } + sc->sc_ranges[i].phys_base = ranges[j++]; + if (sc->sc_pacells == 2) { + sc->sc_ranges[i].phys_base <<= 32; + sc->sc_ranges[i].phys_base |= ranges[j++]; + } + sc->sc_ranges[i].size = ranges[j++]; + if (sc->sc_scells == 2) { + sc->sc_ranges[i].size <<= 32; + sc->sc_ranges[i].size |= ranges[j++]; + } + } + + free(ranges, M_TEMP, rangeslen); + + if (bus_space_map(sc->sc_iot, faa->fa_reg[0].addr, + faa->fa_reg[0].size, 0, &sc->sc_ioh)) { + free(sc->sc_ranges, M_TEMP, sc->sc_nranges * + sizeof(struct mvkpcie_range)); + printf(": can't map ctrl registers\n"); + return; + } + + printf("\n"); + + pinctrl_byname(sc->sc_node, "default"); + + clock_set_assigned(sc->sc_node); + clock_enable_all(sc->sc_node); + + reg = HREAD4(sc, CTRL_CORE_CONFIG); + reg &= ~CTRL_CORE_CONFIG_MODE_MASK; + reg |= CTRL_CORE_CONFIG_MODE_DIRECT; + HWRITE4(sc, CTRL_CORE_CONFIG, reg); + + HSET4(sc, PCIE_CORE_CTRL0, PCIE_CORE_CTRL0_IS_RC); + + HWRITE4(sc, PCIE_ERR_CAPCTL, + PCIE_ERR_CAPCTL_ECRC_CHK_TX | + PCIE_ERR_CAPCTL_ECRC_CHK_TX_EN | + PCIE_ERR_CAPCTL_ECRC_CHCK | + PCIE_ERR_CAPCTL_ECRC_CHCK_RCV); + + HWRITE4(sc, PCIE_DEV_CTRL_STATS, + PCIE_DEV_CTRL_STATS_MAX_PAYLOAD_7 | + PCIE_DEV_CTRL_STATS_MAX_RD_REQ_SZ); + + HWRITE4(sc, PCIE_CORE_CTRL2, + PCIE_CORE_CTRL2_RESERVED | + PCIE_CORE_CTRL2_TD_ENABLE); + + reg = HREAD4(sc, PCIE_CORE_CTRL0); + reg &= ~PCIE_CORE_CTRL0_GEN_MASK; + reg |= PCIE_CORE_CTRL0_GEN_2; + HWRITE4(sc, PCIE_CORE_CTRL0, reg); + + reg = HREAD4(sc, PCIE_CORE_CTRL0); + reg &= ~PCIE_CORE_CTRL0_LANE_MASK; + reg |= PCIE_CORE_CTRL0_LANE_1; + HWRITE4(sc, PCIE_CORE_CTRL0, reg); + + HSET4(sc, PCIE_CORE_CTRL0, PCIE_CORE_CTRL0_LINK_TRAINING); + HSET4(sc, PCIE_CORE_CTRL2, PCIE_CORE_CTRL2_MSI_ENABLE); + + HWRITE4(sc, PCIE_CORE_ISR0_STATUS, PCIE_CORE_ISR0_MASK_ALL); + HWRITE4(sc, PCIE_CORE_ISR1_STATUS, PCIE_CORE_ISR1_MASK_ALL); + HWRITE4(sc, HOST_CTRL_INT_STATUS, HOST_CTRL_INT_MASK_ALL); + + HWRITE4(sc, PCIE_CORE_ISR0_MASK, PCIE_CORE_ISR0_MASK_ALL & + ~PCIE_CORE_ISR0_MASK_MSI_INT); + HWRITE4(sc, PCIE_CORE_ISR1_MASK, PCIE_CORE_ISR1_MASK_ALL); + HWRITE4(sc, PCIE_CORE_MSI_MASK, 0); + HWRITE4(sc, HOST_CTRL_INT_MASK, HOST_CTRL_INT_MASK_ALL & + ~HOST_CTRL_INT_MASK_CORE_INT); + + HSET4(sc, PCIE_CORE_CTRL2, PCIE_CORE_CTRL2_OB_WIN_ENABLE); + HSET4(sc, PIO_CTRL, PIO_CTRL_ADDR_WIN_DISABLE); + + delay(100 * 1000); + + HSET4(sc, PCIE_LINK_CTRL_STAT, PCIE_LINK_CTRL_STAT_LINK_TRAINING); + + for (timo = 40; timo > 0; timo--) { + if (mvkpcie_link_up(sc)) + break; + delay(1000); + } + if (timo == 0) { + printf("%s: timeout\n", sc->sc_dev.dv_xname); + return; + } + + HWRITE4(sc, PCIE_LINK_CTRL_STAT, + PCIE_LINK_CTRL_STAT_LINK_L0S_ENTRY | + PCIE_LINK_CTRL_STAT_LINK_WIDTH_1); + + HSET4(sc, PCIE_CMD, PCI_COMMAND_IO_ENABLE | + PCI_COMMAND_MEM_ENABLE | PCI_COMMAND_MASTER_ENABLE); + + HWRITE4(sc, PCIE_CORE_MSI_ADDR_LOW, + MVKPCIE_DMA_DVA(sc->sc_msi_addr) & 0xffffffff); + HWRITE4(sc, PCIE_CORE_MSI_ADDR_HIGH, + MVKPCIE_DMA_DVA(sc->sc_msi_addr) >> 32); + + /* Set up address translation for I/O space. */ + sc->sc_io_bus_addr = sc->sc_mem_bus_addr = -1; + for (i = 0; i < sc->sc_nranges; i++) { + if ((sc->sc_ranges[i].flags & 0x03000000) == 0x01000000 && + sc->sc_ranges[i].size > 0) { + sc->sc_io_base = sc->sc_ranges[i].phys_base; + sc->sc_io_bus_addr = sc->sc_ranges[i].pci_base; + sc->sc_io_size = sc->sc_ranges[i].size; + } + if ((sc->sc_ranges[i].flags & 0x03000000) == 0x02000000 && + sc->sc_ranges[i].size > 0) { + sc->sc_mem_base = sc->sc_ranges[i].phys_base; + sc->sc_mem_bus_addr = sc->sc_ranges[i].pci_base; + sc->sc_mem_size = sc->sc_ranges[i].size; + } + } + + /* Set up bus range. */ + if (OF_getpropintarray(sc->sc_node, "bus-range", bus_range, + sizeof(bus_range)) != sizeof(bus_range) || + bus_range[0] >= 256 || bus_range[1] >= 256) { + bus_range[0] = 0; + bus_range[1] = 255; + } + sc->sc_bus = bus_range[0]; + + /* Initialize bus range. */ + bir = bus_range[0]; + bir |= ((bus_range[0] + 1) << 8); + bir |= (bus_range[1] << 16); + sc->sc_bridge_businfo = bir; + + /* Initialize I/O window. */ + iobase = sc->sc_io_bus_addr; + iolimit = iobase + sc->sc_io_size - 1; + blr = iolimit & PPB_IO_MASK; + blr |= (iobase >> PPB_IO_SHIFT); + sc->sc_bridge_iostatus = blr; + blr = (iobase & 0xffff0000) >> 16; + blr |= iolimit & 0xffff0000; + sc->sc_bridge_io_hi = blr; + + /* Initialize memory mapped I/O window. */ + membase = sc->sc_mem_bus_addr; + memlimit = membase + sc->sc_mem_size - 1; + blr = memlimit & PPB_MEM_MASK; + blr |= (membase >> PPB_MEM_SHIFT); + sc->sc_bridge_mem = blr; + + memcpy(&sc->sc_bus_iot, sc->sc_iot, sizeof(sc->sc_bus_iot)); + sc->sc_bus_iot.bus_private = sc; + sc->sc_bus_iot._space_map = mvkpcie_bs_iomap; + memcpy(&sc->sc_bus_memt, sc->sc_iot, sizeof(sc->sc_bus_memt)); + sc->sc_bus_memt.bus_private = sc; + sc->sc_bus_memt._space_map = mvkpcie_bs_memmap; + + sc->sc_pc.pc_conf_v = sc; + sc->sc_pc.pc_attach_hook = mvkpcie_attach_hook; + sc->sc_pc.pc_bus_maxdevs = mvkpcie_bus_maxdevs; + sc->sc_pc.pc_make_tag = mvkpcie_make_tag; + sc->sc_pc.pc_decompose_tag = mvkpcie_decompose_tag; + sc->sc_pc.pc_conf_size = mvkpcie_conf_size; + sc->sc_pc.pc_conf_read = mvkpcie_conf_read; + sc->sc_pc.pc_conf_write = mvkpcie_conf_write; + + sc->sc_pc.pc_intr_v = sc; + sc->sc_pc.pc_intr_map = mvkpcie_intr_map; + sc->sc_pc.pc_intr_map_msi = _pci_intr_map_msi; + sc->sc_pc.pc_intr_map_msix = _pci_intr_map_msix; + sc->sc_pc.pc_intr_string = mvkpcie_intr_string; + sc->sc_pc.pc_intr_establish = mvkpcie_intr_establish; + sc->sc_pc.pc_intr_disestablish = mvkpcie_intr_disestablish; + + memset(&pba, 0, sizeof(pba)); + pba.pba_busname = "pci"; + pba.pba_iot = &sc->sc_bus_iot; + pba.pba_memt = &sc->sc_bus_memt; + pba.pba_dmat = faa->fa_dmat; + pba.pba_pc = &sc->sc_pc; + pba.pba_domain = pci_ndomains++; + pba.pba_bus = sc->sc_bus; + pba.pba_flags |= PCI_FLAGS_MSI_ENABLED; + + sc->sc_msi_ic.ic_node = faa->fa_node; + sc->sc_msi_ic.ic_cookie = self; + sc->sc_msi_ic.ic_establish_msi = mvkpcie_intc_intr_establish_msi; + sc->sc_msi_ic.ic_disestablish = mvkpcie_intc_intr_disestablish_msi; + arm_intr_register_fdt(&sc->sc_msi_ic); + + config_found(self, &pba, NULL); +} + +int +mvkpcie_link_up(struct mvkpcie_softc *sc) +{ + uint32_t reg; + + reg = HREAD4(sc, LMI_CFG); + return LMI_CFG_LTSSM_VAL(reg) >= LMI_CFG_LTSSM_L0; +} + +void +mvkpcie_attach_hook(struct device *parent, struct device *self, + struct pcibus_attach_args *pba) +{ +} + +int +mvkpcie_bus_maxdevs(void *v, int bus) +{ + struct mvkpcie_softc *sc = v; + + if (bus == sc->sc_bus || bus == sc->sc_bus + 1) + return 1; + return 32; +} + +pcitag_t +mvkpcie_make_tag(void *v, int bus, int device, int function) +{ + return ((bus << 20) | (device << 15) | (function << 12)); +} + +void +mvkpcie_decompose_tag(void *v, pcitag_t tag, int *bp, int *dp, int *fp) +{ + if (bp != NULL) + *bp = (tag >> 20) & 0xff; + if (dp != NULL) + *dp = (tag >> 15) & 0x1f; + if (fp != NULL) + *fp = (tag >> 12) & 0x7; +} + +int +mvkpcie_conf_size(void *v, pcitag_t tag) +{ + return PCIE_CONFIG_SPACE_SIZE; +} + +pcireg_t +mvkpcie_conf_read_bridge(struct mvkpcie_softc *sc, int reg) +{ + switch (reg) { + case PCI_ID_REG: + return PCI_VENDOR_MARVELL | + (HREAD4(sc, PCIE_DEV_ID) & 0xffff0000); + case PCI_COMMAND_STATUS_REG: + return sc->sc_bridge_command; + case PCI_CLASS_REG: + return PCI_CLASS_BRIDGE << PCI_CLASS_SHIFT | + PCI_SUBCLASS_BRIDGE_PCI << PCI_SUBCLASS_SHIFT | + (HREAD4(sc, PCIE_DEV_REV) & 0xff); + case PCI_BHLC_REG: + return 1 << PCI_HDRTYPE_SHIFT | + 0x10 << PCI_CACHELINE_SHIFT; + case PPB_REG_BASE0: + return sc->sc_bridge_bar0; + case PPB_REG_BASE1: + return sc->sc_bridge_bar1; + case PPB_REG_BUSINFO: + return sc->sc_bridge_businfo; + case PPB_REG_IOSTATUS: + return sc->sc_bridge_iostatus; + case PPB_REG_MEM: + return sc->sc_bridge_mem; + case PPB_REG_IO_HI: + return sc->sc_bridge_io_hi; + case PPB_REG_PREFMEM: + case PPB_REG_PREFBASE_HI32: + case PPB_REG_PREFLIM_HI32: + case PPB_REG_BRIDGECONTROL: + return 0; + default: + printf("%s: reg %x\n", __func__, reg); + break; + } + return 0; +} + +void +mvkpcie_conf_write_bridge(struct mvkpcie_softc *sc, int reg, pcireg_t data) +{ + switch (reg) { + case PCI_COMMAND_STATUS_REG: { + sc->sc_bridge_command = data & 0xffffff; + break; + } + case PPB_REG_BASE0: + sc->sc_bridge_bar0 = data; + break; + case PPB_REG_BASE1: + sc->sc_bridge_bar1 = data; + break; + case PPB_REG_BUSINFO: + sc->sc_bridge_businfo = data; + break; + case PPB_REG_IOSTATUS: + sc->sc_bridge_iostatus = data; + break; + case PPB_REG_MEM: + sc->sc_bridge_mem = data; + break; + case PPB_REG_IO_HI: + sc->sc_bridge_io_hi = data; + break; + case PPB_REG_PREFMEM: + case PPB_REG_PREFBASE_HI32: + case PPB_REG_PREFLIM_HI32: + break; + default: + printf("%s: reg %x data %x\n", __func__, reg, data); + break; + } +} + +pcireg_t +mvkpcie_conf_read(void *v, pcitag_t tag, int off) +{ + struct mvkpcie_softc *sc = v; + int bus, dev, fn; + uint32_t reg; + int i; + + mvkpcie_decompose_tag(sc, tag, &bus, &dev, &fn); + if (bus == sc->sc_bus) { + KASSERT(dev == 0); + return mvkpcie_conf_read_bridge(sc, off); + } + + HWRITE4(sc, PIO_START, PIO_START_STOP); + HWRITE4(sc, PIO_ISR, PIO_ISR_CLEAR); + reg = HREAD4(sc, PIO_CTRL); + reg &= ~PIO_CTRL_TYPE_MASK; + if (bus == sc->sc_bus + 1) + reg |= PIO_CTRL_TYPE_RD0; + else + reg |= PIO_CTRL_TYPE_RD1; + HWRITE4(sc, PIO_CTRL, reg); + HWRITE4(sc, PIO_ADDR_LS, tag | off); + HWRITE4(sc, PIO_ADDR_MS, 0); + HWRITE4(sc, PIO_WR_DATA_STRB, PIO_WR_DATA_STRB_VALUE); + HWRITE4(sc, PIO_START, PIO_START_START); + + for (i = 500; i > 0; i--) { + if (HREAD4(sc, PIO_START) == 0 && + HREAD4(sc, PIO_ISR) != 0) + break; + delay(2); + } + if (i == 0) { + printf("%s: timeout\n", sc->sc_dev.dv_xname); + return 0xffffffff; + } + + return HREAD4(sc, PIO_RD_DATA); +} + +void +mvkpcie_conf_write(void *v, pcitag_t tag, int off, pcireg_t data) +{ + struct mvkpcie_softc *sc = v; + int bus, dev, fn; + uint32_t reg; + int i; + + mvkpcie_decompose_tag(sc, tag, &bus, &dev, &fn); + if (bus == sc->sc_bus) { + KASSERT(dev == 0); + mvkpcie_conf_write_bridge(sc, off, data); + return; + } + + HWRITE4(sc, PIO_START, PIO_START_STOP); + HWRITE4(sc, PIO_ISR, PIO_ISR_CLEAR); + reg = HREAD4(sc, PIO_CTRL); + reg &= ~PIO_CTRL_TYPE_MASK; + if (bus == sc->sc_bus + 1) + reg |= PIO_CTRL_TYPE_WR0; + else + reg |= PIO_CTRL_TYPE_WR1; + HWRITE4(sc, PIO_CTRL, reg); + HWRITE4(sc, PIO_ADDR_LS, tag | off); + HWRITE4(sc, PIO_ADDR_MS, 0); + HWRITE4(sc, PIO_WR_DATA, data); + HWRITE4(sc, PIO_WR_DATA_STRB, PIO_WR_DATA_STRB_VALUE); + HWRITE4(sc, PIO_START, PIO_START_START); + + for (i = 500; i > 0; i--) { + if (HREAD4(sc, PIO_START) == 0 && + HREAD4(sc, PIO_ISR) != 0) + break; + delay(2); + } + if (i == 0) { + printf("%s: timeout\n", sc->sc_dev.dv_xname); + return; + } +} + +int +mvkpcie_intr_map(struct pci_attach_args *pa, pci_intr_handle_t *ihp) +{ + return -1; +} + +const char * +mvkpcie_intr_string(void *v, pci_intr_handle_t ih) +{ + switch (ih.ih_type) { + case PCI_MSI: + return "msi"; + case PCI_MSIX: + return "msix"; + } + + return "intx"; +} + +void * +mvkpcie_intr_establish(void *v, pci_intr_handle_t ih, int level, + int (*func)(void *), void *arg, char *name) +{ + struct mvkpcie_softc *sc = v; + void *cookie; + + KASSERT(ih.ih_type != PCI_NONE); + + if (ih.ih_type != PCI_INTX) { + uint64_t addr, data; + + /* Assume hardware passes Requester ID as sideband data. */ + data = pci_requester_id(ih.ih_pc, ih.ih_tag); + cookie = fdt_intr_establish_msi(sc->sc_node, &addr, + &data, level, func, arg, (void *)name); + if (cookie == NULL) + return NULL; + + /* TODO: translate address to the PCI device's view */ + + if (ih.ih_type == PCI_MSIX) { + pci_msix_enable(ih.ih_pc, ih.ih_tag, + &sc->sc_bus_memt, ih.ih_intrpin, addr, data); + } else + pci_msi_enable(ih.ih_pc, ih.ih_tag, addr, data); + } else { + int bus, dev, fn; + uint32_t reg[4]; + + mvkpcie_decompose_tag(sc, ih.ih_tag, &bus, &dev, &fn); + + reg[0] = bus << 16 | dev << 11 | fn << 8; + reg[1] = reg[2] = 0; + reg[3] = ih.ih_intrpin; + + cookie = fdt_intr_establish_imap(sc->sc_node, reg, + sizeof(reg), level, func, arg, name); + } + + return cookie; +} + +void +mvkpcie_intr_disestablish(void *v, void *cookie) +{ + panic("%s", __func__); +} + +int +mvkpcie_bs_iomap(bus_space_tag_t t, bus_addr_t addr, bus_size_t size, + int flags, bus_space_handle_t *bshp) +{ + struct mvkpcie_softc *sc = t->bus_private; + int i; + + for (i = 0; i < sc->sc_nranges; i++) { + uint64_t pci_start = sc->sc_ranges[i].pci_base; + uint64_t pci_end = pci_start + sc->sc_ranges[i].size; + uint64_t phys_start = sc->sc_ranges[i].phys_base; + + if ((sc->sc_ranges[i].flags & 0x03000000) == 0x01000000 && + addr >= pci_start && addr + size <= pci_end) { + return bus_space_map(sc->sc_iot, + addr - pci_start + phys_start, size, flags, bshp); + } + } + + return ENXIO; +} + +int +mvkpcie_bs_memmap(bus_space_tag_t t, bus_addr_t addr, bus_size_t size, + int flags, bus_space_handle_t *bshp) +{ + struct mvkpcie_softc *sc = t->bus_private; + int i; + + for (i = 0; i < sc->sc_nranges; i++) { + uint64_t pci_start = sc->sc_ranges[i].pci_base; + uint64_t pci_end = pci_start + sc->sc_ranges[i].size; + uint64_t phys_start = sc->sc_ranges[i].phys_base; + + if ((sc->sc_ranges[i].flags & 0x03000000) == 0x02000000 && + addr >= pci_start && addr + size <= pci_end) { + return bus_space_map(sc->sc_iot, + addr - pci_start + phys_start, size, flags, bshp); + } + } + + return ENXIO; +} + +int +mvkpcie_intc_intr(void *cookie) +{ + struct mvkpcie_softc *sc = (struct mvkpcie_softc *)cookie; + struct intrhand *ih; + uint32_t pending; + int i, s; + + if (!(HREAD4(sc, HOST_CTRL_INT_STATUS) & HOST_CTRL_INT_MASK_CORE_INT)) + return 0; + + if (!(HREAD4(sc, PCIE_CORE_ISR0_STATUS) & PCIE_CORE_ISR0_MASK_MSI_INT)) + return 0; + + pending = HREAD4(sc, PCIE_CORE_MSI_STATUS); + while (pending) { + i = ffs(pending) - 1; + HWRITE4(sc, PCIE_CORE_MSI_STATUS, (1 << i)); + pending &= ~(1 << i); + + i = HREAD4(sc, PCIE_CORE_MSI_PAYLOAD) & 0xff; + if ((ih = sc->sc_msi_handlers[i]) != NULL) { + s = splraise(ih->ih_ipl); + if (ih->ih_func(ih->ih_arg)) + ih->ih_count.ec_count++; + splx(s); + } + } + + HWRITE4(sc, PCIE_CORE_ISR0_STATUS, PCIE_CORE_ISR0_MASK_MSI_INT); + HWRITE4(sc, HOST_CTRL_INT_STATUS, HOST_CTRL_INT_MASK_CORE_INT); + return 1; +} + +void * +mvkpcie_intc_intr_establish_msi(void *cookie, uint64_t *addr, uint64_t *data, + int level, int (*func)(void *), void *arg, char *name) +{ + struct mvkpcie_softc *sc = (struct mvkpcie_softc *)cookie; + struct intrhand *ih; + int i, s; + + for (i = 0; i < nitems(sc->sc_msi_handlers); i++) { + if (sc->sc_msi_handlers[i] == NULL) + break; + } + + if (i == nitems(sc->sc_msi_handlers)) + return NULL; + + ih = malloc(sizeof(*ih), M_DEVBUF, M_WAITOK); + ih->ih_func = func; + ih->ih_arg = arg; + ih->ih_ipl = level & IPL_IRQMASK; + ih->ih_irq = i; + ih->ih_name = name; + ih->ih_sc = sc; + + s = splhigh(); + + sc->sc_msi_handlers[i] = ih; + + if (name != NULL) + evcount_attach(&ih->ih_count, name, &ih->ih_irq); + + mvkpcie_intc_recalc_ipl(sc); + + *addr = MVKPCIE_DMA_DVA(sc->sc_msi_addr); + *data = i; + + splx(s); + return (ih); +} + +void +mvkpcie_intc_intr_disestablish_msi(void *cookie) +{ + struct intrhand *ih = cookie; + struct mvkpcie_softc *sc = ih->ih_sc; + int s; + + s = splhigh(); + + sc->sc_msi_handlers[ih->ih_irq] = NULL; + if (ih->ih_name != NULL) + evcount_detach(&ih->ih_count); + free(ih, M_DEVBUF, sizeof(*ih)); + + mvkpcie_intc_recalc_ipl(sc); + + splx(s); +} + +void +mvkpcie_intc_recalc_ipl(struct mvkpcie_softc *sc) +{ + struct intrhand *ih; + int max = IPL_NONE; + int min = IPL_HIGH; + int irq; + + for (irq = 0; irq < nitems(sc->sc_msi_handlers); irq++) { + ih = sc->sc_msi_handlers[irq]; + if (ih == NULL) + continue; + + if (ih->ih_ipl > max) + max = ih->ih_ipl; + + if (ih->ih_ipl < min) + min = ih->ih_ipl; + } + + for (irq = 0; irq < nitems(sc->sc_msi_handlers); irq++) { + ih = sc->sc_msi_handlers[irq]; + if (ih == NULL) + continue; + + if (ih->ih_ipl > max) + max = ih->ih_ipl; + + if (ih->ih_ipl < min) + min = ih->ih_ipl; + } + + if (max == IPL_NONE) + min = IPL_NONE; + + if (sc->sc_ipl != max) { + sc->sc_ipl = max; + + if (sc->sc_ih != NULL) + fdt_intr_disestablish(sc->sc_ih); + + if (sc->sc_ipl != IPL_NONE) + sc->sc_ih = fdt_intr_establish(sc->sc_node, sc->sc_ipl, + mvkpcie_intc_intr, sc, sc->sc_dev.dv_xname); + } +} + +/* Only needed for the 16-bit MSI address */ +struct mvkpcie_dmamem * +mvkpcie_dmamem_alloc(struct mvkpcie_softc *sc, bus_size_t size, bus_size_t align) +{ + struct mvkpcie_dmamem *mdm; + int nsegs; + + mdm = malloc(sizeof(*mdm), M_DEVBUF, M_WAITOK | M_ZERO); + mdm->mdm_size = size; + + if (bus_dmamap_create(sc->sc_dmat, size, 1, size, 0, + BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW, &mdm->mdm_map) != 0) + goto mdmfree; + + if (bus_dmamem_alloc(sc->sc_dmat, size, align, 0, &mdm->mdm_seg, 1, + &nsegs, BUS_DMA_WAITOK) != 0) + goto destroy; + + if (bus_dmamem_map(sc->sc_dmat, &mdm->mdm_seg, nsegs, size, + &mdm->mdm_kva, BUS_DMA_WAITOK | BUS_DMA_COHERENT) != 0) + goto free; + + if (bus_dmamap_load(sc->sc_dmat, mdm->mdm_map, mdm->mdm_kva, size, + NULL, BUS_DMA_WAITOK) != 0) + goto unmap; + + bzero(mdm->mdm_kva, size); + + return (mdm); + +unmap: + bus_dmamem_unmap(sc->sc_dmat, mdm->mdm_kva, size); +free: + bus_dmamem_free(sc->sc_dmat, &mdm->mdm_seg, 1); +destroy: + bus_dmamap_destroy(sc->sc_dmat, mdm->mdm_map); +mdmfree: + free(mdm, M_DEVBUF, sizeof(*mdm)); + + return (NULL); +} + +void +mvkpcie_dmamem_free(struct mvkpcie_softc *sc, struct mvkpcie_dmamem *mdm) +{ + bus_dmamem_unmap(sc->sc_dmat, mdm->mdm_kva, mdm->mdm_size); + bus_dmamem_free(sc->sc_dmat, &mdm->mdm_seg, 1); + bus_dmamap_destroy(sc->sc_dmat, mdm->mdm_map); + free(mdm, M_DEVBUF, sizeof(*mdm)); +} |