summaryrefslogtreecommitdiffstats
path: root/sys
diff options
context:
space:
mode:
authorpatrick <patrick@openbsd.org>2020-05-22 08:48:49 +0000
committerpatrick <patrick@openbsd.org>2020-05-22 08:48:49 +0000
commita69c39e4dda6bcd002b2110ac97ebdb484b5624b (patch)
tree15c111bf59f7087b470c0785bdeda82e535a7234 /sys
parentAdd test covering revision 1.64 of c_sh, fix exit code of compound lists (diff)
downloadwireguard-openbsd-a69c39e4dda6bcd002b2110ac97ebdb484b5624b.tar.xz
wireguard-openbsd-a69c39e4dda6bcd002b2110ac97ebdb484b5624b.zip
Add mvkpcie(4), a driver for the Aardvark PCIe controller found on
the Armada 3700 SoC. So far only MSI is supported, since it was easier to implement. Both MSI and INTx share a single interrupt pin, so there's nothing to gain anyway, apart from legacy device support. With this I can push traffic through a bwfm(4) in my Turris Mox. ok kettenis@
Diffstat (limited to 'sys')
-rw-r--r--sys/dev/fdt/files.fdt6
-rw-r--r--sys/dev/fdt/mvkpcie.c1026
2 files changed, 1031 insertions, 1 deletions
diff --git a/sys/dev/fdt/files.fdt b/sys/dev/fdt/files.fdt
index 364e6a28e1c..46f1cd15449 100644
--- a/sys/dev/fdt/files.fdt
+++ b/sys/dev/fdt/files.fdt
@@ -1,4 +1,4 @@
-# $OpenBSD: files.fdt,v 1.133 2020/05/18 10:39:33 kettenis Exp $
+# $OpenBSD: files.fdt,v 1.134 2020/05/22 08:48:49 patrick Exp $
#
# Config file and device description for machine-independent FDT code.
# Included by ports that need it.
@@ -374,6 +374,10 @@ device mviic: i2cbus
attach mviic at fdt
file dev/fdt/mviic.c mviic
+device mvkpcie: pcibus
+attach mvkpcie at fdt
+file dev/fdt/mvkpcie.c mvkpcie
+
device mvpinctrl
attach mvpinctrl at fdt
file dev/fdt/mvpinctrl.c mvpinctrl
diff --git a/sys/dev/fdt/mvkpcie.c b/sys/dev/fdt/mvkpcie.c
new file mode 100644
index 00000000000..387d57cec01
--- /dev/null
+++ b/sys/dev/fdt/mvkpcie.c
@@ -0,0 +1,1026 @@
+/* $OpenBSD: mvkpcie.c,v 1.1 2020/05/22 08:48:49 patrick Exp $ */
+/*
+ * Copyright (c) 2018 Mark Kettenis <kettenis@openbsd.org>
+ * Copyright (c) 2020 Patrick Wildt <patrick@blueri.se>
+ *
+ * Permission to use, copy, modify, and distribute this software for any
+ * purpose with or without fee is hereby granted, provided that the above
+ * copyright notice and this permission notice appear in all copies.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
+ * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
+ * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
+ * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
+ * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
+ * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
+ * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
+ */
+
+#include <sys/param.h>
+#include <sys/systm.h>
+#include <sys/device.h>
+#include <sys/extent.h>
+#include <sys/malloc.h>
+#include <sys/evcount.h>
+
+#include <machine/intr.h>
+#include <machine/bus.h>
+#include <machine/fdt.h>
+
+#include <dev/pci/pcidevs.h>
+#include <dev/pci/pcireg.h>
+#include <dev/pci/pcivar.h>
+#include <dev/pci/ppbreg.h>
+
+#include <dev/ofw/openfirm.h>
+#include <dev/ofw/ofw_clock.h>
+#include <dev/ofw/ofw_gpio.h>
+#include <dev/ofw/ofw_misc.h>
+#include <dev/ofw/ofw_pinctrl.h>
+#include <dev/ofw/ofw_power.h>
+#include <dev/ofw/fdt.h>
+
+/* Registers */
+#define PCIE_DEV_ID 0x0000
+#define PCIE_CMD 0x0004
+#define PCIE_DEV_REV 0x0008
+#define PCIE_DEV_CTRL_STATS 0x00c8
+#define PCIE_DEV_CTRL_STATS_SNOOP (1 << 1)
+#define PCIE_DEV_CTRL_STATS_RELAX_ORDER (1 << 4)
+#define PCIE_DEV_CTRL_STATS_MAX_PAYLOAD_7 (0x7 << 5)
+#define PCIE_DEV_CTRL_STATS_MAX_RD_REQ_SZ (0x2 << 12)
+#define PCIE_LINK_CTRL_STAT 0x00d0
+#define PCIE_LINK_CTRL_STAT_LINK_L0S_ENTRY (1 << 0)
+#define PCIE_LINK_CTRL_STAT_LINK_TRAINING (1 << 5)
+#define PCIE_LINK_CTRL_STAT_LINK_WIDTH_1 (1 << 20)
+#define PCIE_ERR_CAPCTL 0x0118
+#define PCIE_ERR_CAPCTL_ECRC_CHK_TX (1 << 5)
+#define PCIE_ERR_CAPCTL_ECRC_CHK_TX_EN (1 << 6)
+#define PCIE_ERR_CAPCTL_ECRC_CHCK (1 << 7)
+#define PCIE_ERR_CAPCTL_ECRC_CHCK_RCV (1 << 8)
+#define PIO_CTRL 0x4000
+#define PIO_CTRL_TYPE_MASK (0xf << 0)
+#define PIO_CTRL_TYPE_RD0 (0x8 << 0)
+#define PIO_CTRL_TYPE_RD1 (0x9 << 0)
+#define PIO_CTRL_TYPE_WR0 (0xa << 0)
+#define PIO_CTRL_TYPE_WR1 (0xb << 0)
+#define PIO_CTRL_ADDR_WIN_DISABLE (1 << 24)
+#define PIO_STAT 0x4004
+#define PIO_STAT_COMP_STATUS (0x7 << 7)
+#define PIO_ADDR_LS 0x4008
+#define PIO_ADDR_MS 0x400c
+#define PIO_WR_DATA 0x4010
+#define PIO_WR_DATA_STRB 0x4014
+#define PIO_WR_DATA_STRB_VALUE 0xf
+#define PIO_RD_DATA 0x4018
+#define PIO_START 0x401c
+#define PIO_START_STOP (0 << 0)
+#define PIO_START_START (1 << 0)
+#define PIO_ISR 0x4020
+#define PIO_ISR_CLEAR (1 << 0)
+#define PIO_ISRM 0x4024
+#define PCIE_CORE_CTRL0 0x4800
+#define PCIE_CORE_CTRL0_GEN_1 (0 << 0)
+#define PCIE_CORE_CTRL0_GEN_2 (1 << 0)
+#define PCIE_CORE_CTRL0_GEN_3 (2 << 0)
+#define PCIE_CORE_CTRL0_GEN_MASK (0x3 << 0)
+#define PCIE_CORE_CTRL0_IS_RC (1 << 2)
+#define PCIE_CORE_CTRL0_LANE_1 (0 << 3)
+#define PCIE_CORE_CTRL0_LANE_2 (1 << 3)
+#define PCIE_CORE_CTRL0_LANE_4 (2 << 3)
+#define PCIE_CORE_CTRL0_LANE_8 (3 << 3)
+#define PCIE_CORE_CTRL0_LANE_MASK (0x3 << 3)
+#define PCIE_CORE_CTRL0_LINK_TRAINING (1 << 6)
+#define PCIE_CORE_CTRL2 0x4808
+#define PCIE_CORE_CTRL2_RESERVED (0x7 << 0)
+#define PCIE_CORE_CTRL2_TD_ENABLE (1 << 4)
+#define PCIE_CORE_CTRL2_STRICT_ORDER_ENABLE (1 << 5)
+#define PCIE_CORE_CTRL2_OB_WIN_ENABLE (1 << 6)
+#define PCIE_CORE_CTRL2_MSI_ENABLE (1 << 10)
+#define PCIE_CORE_ISR0_STATUS 0x4840
+#define PCIE_CORE_ISR0_MASK 0x4844
+#define PCIE_CORE_ISR0_MASK_MSI_INT (1 << 24)
+#define PCIE_CORE_ISR0_MASK_ALL 0x07ffffff
+#define PCIE_CORE_ISR1_STATUS 0x4848
+#define PCIE_CORE_ISR1_MASK 0x484c
+#define PCIE_CORE_ISR1_MASK_ALL 0x00000ff0
+#define PCIE_CORE_ISR1_MASK_INTX(x) (1 << (x + 8))
+#define PCIE_CORE_MSI_ADDR_LOW 0x4850
+#define PCIE_CORE_MSI_ADDR_HIGH 0x4854
+#define PCIE_CORE_MSI_STATUS 0x4858
+#define PCIE_CORE_MSI_MASK 0x485c
+#define PCIE_CORE_MSI_PAYLOAD 0x489c
+#define LMI_CFG 0x6000
+#define LMI_CFG_LTSSM_VAL(x) (((x) >> 24) & 0x3f)
+#define LMI_CFG_LTSSM_L0 0x10
+#define CTRL_CORE_CONFIG 0x18000
+#define CTRL_CORE_CONFIG_MODE_DIRECT (0 << 0)
+#define CTRL_CORE_CONFIG_MODE_COMMAND (1 << 0)
+#define CTRL_CORE_CONFIG_MODE_MASK (1 << 0)
+#define HOST_CTRL_INT_STATUS 0x1b000
+#define HOST_CTRL_INT_MASK 0x1b004
+#define HOST_CTRL_INT_MASK_CORE_INT (1 << 16)
+#define HOST_CTRL_INT_MASK_ALL 0xfff0fb
+
+#define HREAD4(sc, reg) \
+ (bus_space_read_4((sc)->sc_iot, (sc)->sc_ioh, (reg)))
+#define HWRITE4(sc, reg, val) \
+ bus_space_write_4((sc)->sc_iot, (sc)->sc_ioh, (reg), (val))
+#define HSET4(sc, reg, bits) \
+ HWRITE4((sc), (reg), HREAD4((sc), (reg)) | (bits))
+#define HCLR4(sc, reg, bits) \
+ HWRITE4((sc), (reg), HREAD4((sc), (reg)) & ~(bits))
+
+struct mvkpcie_dmamem {
+ bus_dmamap_t mdm_map;
+ bus_dma_segment_t mdm_seg;
+ size_t mdm_size;
+ caddr_t mdm_kva;
+};
+
+#define MVKPCIE_DMA_MAP(_mdm) ((_mdm)->mdm_map)
+#define MVKPCIE_DMA_LEN(_mdm) ((_mdm)->mdm_size)
+#define MVKPCIE_DMA_DVA(_mdm) ((uint64_t)(_mdm)->mdm_map->dm_segs[0].ds_addr)
+#define MVKPCIE_DMA_KVA(_mdm) ((void *)(_mdm)->mdm_kva)
+
+struct intrhand {
+ int (*ih_func)(void *); /* handler */
+ void *ih_arg; /* arg for handler */
+ int ih_ipl; /* IPL_* */
+ int ih_irq; /* IRQ number */
+ struct evcount ih_count;
+ char *ih_name;
+ void *ih_sc;
+};
+
+struct mvkpcie_range {
+ uint32_t flags;
+ uint64_t pci_base;
+ uint64_t phys_base;
+ uint64_t size;
+};
+
+struct mvkpcie_softc {
+ struct device sc_dev;
+ bus_space_tag_t sc_iot;
+ bus_space_handle_t sc_ioh;
+ bus_dma_tag_t sc_dmat;
+
+ bus_addr_t sc_io_base;
+ bus_addr_t sc_io_bus_addr;
+ bus_size_t sc_io_size;
+ bus_addr_t sc_mem_base;
+ bus_addr_t sc_mem_bus_addr;
+ bus_size_t sc_mem_size;
+
+ int sc_node;
+ int sc_acells;
+ int sc_scells;
+ int sc_pacells;
+ int sc_pscells;
+ struct mvkpcie_range *sc_ranges;
+ int sc_nranges;
+
+ struct bus_space sc_bus_iot;
+ struct bus_space sc_bus_memt;
+
+ struct arm64_pci_chipset sc_pc;
+ int sc_bus;
+
+ uint32_t sc_bridge_command;
+ uint32_t sc_bridge_bar0;
+ uint32_t sc_bridge_bar1;
+ uint32_t sc_bridge_businfo;
+ uint32_t sc_bridge_iostatus;
+ uint32_t sc_bridge_io_hi;
+ uint32_t sc_bridge_mem;
+
+ struct interrupt_controller sc_msi_ic;
+ struct intrhand *sc_msi_handlers[32];
+ struct mvkpcie_dmamem *sc_msi_addr;
+ void *sc_ih;
+ int sc_ipl;
+};
+
+int mvkpcie_match(struct device *, void *, void *);
+void mvkpcie_attach(struct device *, struct device *, void *);
+
+struct cfattach mvkpcie_ca = {
+ sizeof (struct mvkpcie_softc), mvkpcie_match, mvkpcie_attach
+};
+
+struct cfdriver mvkpcie_cd = {
+ NULL, "mvkpcie", DV_DULL
+};
+
+int
+mvkpcie_match(struct device *parent, void *match, void *aux)
+{
+ struct fdt_attach_args *faa = aux;
+
+ return OF_is_compatible(faa->fa_node, "marvell,armada-3700-pcie");
+}
+
+int mvkpcie_link_up(struct mvkpcie_softc *);
+
+void mvkpcie_attach_hook(struct device *, struct device *,
+ struct pcibus_attach_args *);
+int mvkpcie_bus_maxdevs(void *, int);
+pcitag_t mvkpcie_make_tag(void *, int, int, int);
+void mvkpcie_decompose_tag(void *, pcitag_t, int *, int *, int *);
+int mvkpcie_conf_size(void *, pcitag_t);
+pcireg_t mvkpcie_conf_read(void *, pcitag_t, int);
+void mvkpcie_conf_write(void *, pcitag_t, int, pcireg_t);
+
+int mvkpcie_intr_map(struct pci_attach_args *, pci_intr_handle_t *);
+const char *mvkpcie_intr_string(void *, pci_intr_handle_t);
+void *mvkpcie_intr_establish(void *, pci_intr_handle_t, int,
+ int (*)(void *), void *, char *);
+void mvkpcie_intr_disestablish(void *, void *);
+
+void *mvkpcie_intr_establish(void *, pci_intr_handle_t, int,
+ int (*)(void *), void *, char *);
+
+int mvkpcie_bs_iomap(bus_space_tag_t, bus_addr_t, bus_size_t, int,
+ bus_space_handle_t *);
+int mvkpcie_bs_memmap(bus_space_tag_t, bus_addr_t, bus_size_t, int,
+ bus_space_handle_t *);
+
+int mvkpcie_intc_intr(void *);
+void *mvkpcie_intc_intr_establish_msi(void *, uint64_t *, uint64_t *,
+ int , int (*)(void *), void *, char *);
+void mvkpcie_intc_intr_disestablish_msi(void *);
+void mvkpcie_intc_recalc_ipl(struct mvkpcie_softc *);
+
+struct mvkpcie_dmamem *mvkpcie_dmamem_alloc(struct mvkpcie_softc *, bus_size_t,
+ bus_size_t);
+void mvkpcie_dmamem_free(struct mvkpcie_softc *, struct mvkpcie_dmamem *);
+
+void
+mvkpcie_attach(struct device *parent, struct device *self, void *aux)
+{
+ struct mvkpcie_softc *sc = (struct mvkpcie_softc *)self;
+ struct fdt_attach_args *faa = aux;
+ struct pcibus_attach_args pba;
+ bus_addr_t iobase, iolimit;
+ bus_addr_t membase, memlimit;
+ uint32_t bus_range[2];
+ uint32_t *ranges;
+ int i, j, nranges, rangeslen;
+ pcireg_t bir, blr;
+ uint32_t reg;
+ int timo;
+
+ if (faa->fa_nreg < 1) {
+ printf(": no registers\n");
+ return;
+ }
+
+ sc->sc_iot = faa->fa_iot;
+ sc->sc_dmat = faa->fa_dmat;
+ sc->sc_node = faa->fa_node;
+
+ sc->sc_acells = OF_getpropint(sc->sc_node, "#address-cells",
+ faa->fa_acells);
+ sc->sc_scells = OF_getpropint(sc->sc_node, "#size-cells",
+ faa->fa_scells);
+ sc->sc_pacells = faa->fa_acells;
+ sc->sc_pscells = faa->fa_scells;
+
+ rangeslen = OF_getproplen(sc->sc_node, "ranges");
+ if (rangeslen <= 0 || (rangeslen % sizeof(uint32_t)) ||
+ (rangeslen / sizeof(uint32_t)) % (sc->sc_acells +
+ sc->sc_pacells + sc->sc_scells)) {
+ printf(": invalid ranges property\n");
+ return;
+ }
+
+ sc->sc_msi_addr = mvkpcie_dmamem_alloc(sc, sizeof(uint16_t),
+ sizeof(uint64_t));
+ if (sc->sc_msi_addr == NULL) {
+ printf(": cannot allocate MSI address\n");
+ return;
+ }
+
+ ranges = malloc(rangeslen, M_TEMP, M_WAITOK);
+ OF_getpropintarray(sc->sc_node, "ranges", ranges,
+ rangeslen);
+
+ nranges = (rangeslen / sizeof(uint32_t)) /
+ (sc->sc_acells + sc->sc_pacells + sc->sc_scells);
+ sc->sc_ranges = mallocarray(nranges,
+ sizeof(struct mvkpcie_range), M_TEMP, M_WAITOK);
+ sc->sc_nranges = nranges;
+
+ for (i = 0, j = 0; i < sc->sc_nranges; i++) {
+ sc->sc_ranges[i].flags = ranges[j++];
+ sc->sc_ranges[i].pci_base = ranges[j++];
+ if (sc->sc_acells - 1 == 2) {
+ sc->sc_ranges[i].pci_base <<= 32;
+ sc->sc_ranges[i].pci_base |= ranges[j++];
+ }
+ sc->sc_ranges[i].phys_base = ranges[j++];
+ if (sc->sc_pacells == 2) {
+ sc->sc_ranges[i].phys_base <<= 32;
+ sc->sc_ranges[i].phys_base |= ranges[j++];
+ }
+ sc->sc_ranges[i].size = ranges[j++];
+ if (sc->sc_scells == 2) {
+ sc->sc_ranges[i].size <<= 32;
+ sc->sc_ranges[i].size |= ranges[j++];
+ }
+ }
+
+ free(ranges, M_TEMP, rangeslen);
+
+ if (bus_space_map(sc->sc_iot, faa->fa_reg[0].addr,
+ faa->fa_reg[0].size, 0, &sc->sc_ioh)) {
+ free(sc->sc_ranges, M_TEMP, sc->sc_nranges *
+ sizeof(struct mvkpcie_range));
+ printf(": can't map ctrl registers\n");
+ return;
+ }
+
+ printf("\n");
+
+ pinctrl_byname(sc->sc_node, "default");
+
+ clock_set_assigned(sc->sc_node);
+ clock_enable_all(sc->sc_node);
+
+ reg = HREAD4(sc, CTRL_CORE_CONFIG);
+ reg &= ~CTRL_CORE_CONFIG_MODE_MASK;
+ reg |= CTRL_CORE_CONFIG_MODE_DIRECT;
+ HWRITE4(sc, CTRL_CORE_CONFIG, reg);
+
+ HSET4(sc, PCIE_CORE_CTRL0, PCIE_CORE_CTRL0_IS_RC);
+
+ HWRITE4(sc, PCIE_ERR_CAPCTL,
+ PCIE_ERR_CAPCTL_ECRC_CHK_TX |
+ PCIE_ERR_CAPCTL_ECRC_CHK_TX_EN |
+ PCIE_ERR_CAPCTL_ECRC_CHCK |
+ PCIE_ERR_CAPCTL_ECRC_CHCK_RCV);
+
+ HWRITE4(sc, PCIE_DEV_CTRL_STATS,
+ PCIE_DEV_CTRL_STATS_MAX_PAYLOAD_7 |
+ PCIE_DEV_CTRL_STATS_MAX_RD_REQ_SZ);
+
+ HWRITE4(sc, PCIE_CORE_CTRL2,
+ PCIE_CORE_CTRL2_RESERVED |
+ PCIE_CORE_CTRL2_TD_ENABLE);
+
+ reg = HREAD4(sc, PCIE_CORE_CTRL0);
+ reg &= ~PCIE_CORE_CTRL0_GEN_MASK;
+ reg |= PCIE_CORE_CTRL0_GEN_2;
+ HWRITE4(sc, PCIE_CORE_CTRL0, reg);
+
+ reg = HREAD4(sc, PCIE_CORE_CTRL0);
+ reg &= ~PCIE_CORE_CTRL0_LANE_MASK;
+ reg |= PCIE_CORE_CTRL0_LANE_1;
+ HWRITE4(sc, PCIE_CORE_CTRL0, reg);
+
+ HSET4(sc, PCIE_CORE_CTRL0, PCIE_CORE_CTRL0_LINK_TRAINING);
+ HSET4(sc, PCIE_CORE_CTRL2, PCIE_CORE_CTRL2_MSI_ENABLE);
+
+ HWRITE4(sc, PCIE_CORE_ISR0_STATUS, PCIE_CORE_ISR0_MASK_ALL);
+ HWRITE4(sc, PCIE_CORE_ISR1_STATUS, PCIE_CORE_ISR1_MASK_ALL);
+ HWRITE4(sc, HOST_CTRL_INT_STATUS, HOST_CTRL_INT_MASK_ALL);
+
+ HWRITE4(sc, PCIE_CORE_ISR0_MASK, PCIE_CORE_ISR0_MASK_ALL &
+ ~PCIE_CORE_ISR0_MASK_MSI_INT);
+ HWRITE4(sc, PCIE_CORE_ISR1_MASK, PCIE_CORE_ISR1_MASK_ALL);
+ HWRITE4(sc, PCIE_CORE_MSI_MASK, 0);
+ HWRITE4(sc, HOST_CTRL_INT_MASK, HOST_CTRL_INT_MASK_ALL &
+ ~HOST_CTRL_INT_MASK_CORE_INT);
+
+ HSET4(sc, PCIE_CORE_CTRL2, PCIE_CORE_CTRL2_OB_WIN_ENABLE);
+ HSET4(sc, PIO_CTRL, PIO_CTRL_ADDR_WIN_DISABLE);
+
+ delay(100 * 1000);
+
+ HSET4(sc, PCIE_LINK_CTRL_STAT, PCIE_LINK_CTRL_STAT_LINK_TRAINING);
+
+ for (timo = 40; timo > 0; timo--) {
+ if (mvkpcie_link_up(sc))
+ break;
+ delay(1000);
+ }
+ if (timo == 0) {
+ printf("%s: timeout\n", sc->sc_dev.dv_xname);
+ return;
+ }
+
+ HWRITE4(sc, PCIE_LINK_CTRL_STAT,
+ PCIE_LINK_CTRL_STAT_LINK_L0S_ENTRY |
+ PCIE_LINK_CTRL_STAT_LINK_WIDTH_1);
+
+ HSET4(sc, PCIE_CMD, PCI_COMMAND_IO_ENABLE |
+ PCI_COMMAND_MEM_ENABLE | PCI_COMMAND_MASTER_ENABLE);
+
+ HWRITE4(sc, PCIE_CORE_MSI_ADDR_LOW,
+ MVKPCIE_DMA_DVA(sc->sc_msi_addr) & 0xffffffff);
+ HWRITE4(sc, PCIE_CORE_MSI_ADDR_HIGH,
+ MVKPCIE_DMA_DVA(sc->sc_msi_addr) >> 32);
+
+ /* Set up address translation for I/O space. */
+ sc->sc_io_bus_addr = sc->sc_mem_bus_addr = -1;
+ for (i = 0; i < sc->sc_nranges; i++) {
+ if ((sc->sc_ranges[i].flags & 0x03000000) == 0x01000000 &&
+ sc->sc_ranges[i].size > 0) {
+ sc->sc_io_base = sc->sc_ranges[i].phys_base;
+ sc->sc_io_bus_addr = sc->sc_ranges[i].pci_base;
+ sc->sc_io_size = sc->sc_ranges[i].size;
+ }
+ if ((sc->sc_ranges[i].flags & 0x03000000) == 0x02000000 &&
+ sc->sc_ranges[i].size > 0) {
+ sc->sc_mem_base = sc->sc_ranges[i].phys_base;
+ sc->sc_mem_bus_addr = sc->sc_ranges[i].pci_base;
+ sc->sc_mem_size = sc->sc_ranges[i].size;
+ }
+ }
+
+ /* Set up bus range. */
+ if (OF_getpropintarray(sc->sc_node, "bus-range", bus_range,
+ sizeof(bus_range)) != sizeof(bus_range) ||
+ bus_range[0] >= 256 || bus_range[1] >= 256) {
+ bus_range[0] = 0;
+ bus_range[1] = 255;
+ }
+ sc->sc_bus = bus_range[0];
+
+ /* Initialize bus range. */
+ bir = bus_range[0];
+ bir |= ((bus_range[0] + 1) << 8);
+ bir |= (bus_range[1] << 16);
+ sc->sc_bridge_businfo = bir;
+
+ /* Initialize I/O window. */
+ iobase = sc->sc_io_bus_addr;
+ iolimit = iobase + sc->sc_io_size - 1;
+ blr = iolimit & PPB_IO_MASK;
+ blr |= (iobase >> PPB_IO_SHIFT);
+ sc->sc_bridge_iostatus = blr;
+ blr = (iobase & 0xffff0000) >> 16;
+ blr |= iolimit & 0xffff0000;
+ sc->sc_bridge_io_hi = blr;
+
+ /* Initialize memory mapped I/O window. */
+ membase = sc->sc_mem_bus_addr;
+ memlimit = membase + sc->sc_mem_size - 1;
+ blr = memlimit & PPB_MEM_MASK;
+ blr |= (membase >> PPB_MEM_SHIFT);
+ sc->sc_bridge_mem = blr;
+
+ memcpy(&sc->sc_bus_iot, sc->sc_iot, sizeof(sc->sc_bus_iot));
+ sc->sc_bus_iot.bus_private = sc;
+ sc->sc_bus_iot._space_map = mvkpcie_bs_iomap;
+ memcpy(&sc->sc_bus_memt, sc->sc_iot, sizeof(sc->sc_bus_memt));
+ sc->sc_bus_memt.bus_private = sc;
+ sc->sc_bus_memt._space_map = mvkpcie_bs_memmap;
+
+ sc->sc_pc.pc_conf_v = sc;
+ sc->sc_pc.pc_attach_hook = mvkpcie_attach_hook;
+ sc->sc_pc.pc_bus_maxdevs = mvkpcie_bus_maxdevs;
+ sc->sc_pc.pc_make_tag = mvkpcie_make_tag;
+ sc->sc_pc.pc_decompose_tag = mvkpcie_decompose_tag;
+ sc->sc_pc.pc_conf_size = mvkpcie_conf_size;
+ sc->sc_pc.pc_conf_read = mvkpcie_conf_read;
+ sc->sc_pc.pc_conf_write = mvkpcie_conf_write;
+
+ sc->sc_pc.pc_intr_v = sc;
+ sc->sc_pc.pc_intr_map = mvkpcie_intr_map;
+ sc->sc_pc.pc_intr_map_msi = _pci_intr_map_msi;
+ sc->sc_pc.pc_intr_map_msix = _pci_intr_map_msix;
+ sc->sc_pc.pc_intr_string = mvkpcie_intr_string;
+ sc->sc_pc.pc_intr_establish = mvkpcie_intr_establish;
+ sc->sc_pc.pc_intr_disestablish = mvkpcie_intr_disestablish;
+
+ memset(&pba, 0, sizeof(pba));
+ pba.pba_busname = "pci";
+ pba.pba_iot = &sc->sc_bus_iot;
+ pba.pba_memt = &sc->sc_bus_memt;
+ pba.pba_dmat = faa->fa_dmat;
+ pba.pba_pc = &sc->sc_pc;
+ pba.pba_domain = pci_ndomains++;
+ pba.pba_bus = sc->sc_bus;
+ pba.pba_flags |= PCI_FLAGS_MSI_ENABLED;
+
+ sc->sc_msi_ic.ic_node = faa->fa_node;
+ sc->sc_msi_ic.ic_cookie = self;
+ sc->sc_msi_ic.ic_establish_msi = mvkpcie_intc_intr_establish_msi;
+ sc->sc_msi_ic.ic_disestablish = mvkpcie_intc_intr_disestablish_msi;
+ arm_intr_register_fdt(&sc->sc_msi_ic);
+
+ config_found(self, &pba, NULL);
+}
+
+int
+mvkpcie_link_up(struct mvkpcie_softc *sc)
+{
+ uint32_t reg;
+
+ reg = HREAD4(sc, LMI_CFG);
+ return LMI_CFG_LTSSM_VAL(reg) >= LMI_CFG_LTSSM_L0;
+}
+
+void
+mvkpcie_attach_hook(struct device *parent, struct device *self,
+ struct pcibus_attach_args *pba)
+{
+}
+
+int
+mvkpcie_bus_maxdevs(void *v, int bus)
+{
+ struct mvkpcie_softc *sc = v;
+
+ if (bus == sc->sc_bus || bus == sc->sc_bus + 1)
+ return 1;
+ return 32;
+}
+
+pcitag_t
+mvkpcie_make_tag(void *v, int bus, int device, int function)
+{
+ return ((bus << 20) | (device << 15) | (function << 12));
+}
+
+void
+mvkpcie_decompose_tag(void *v, pcitag_t tag, int *bp, int *dp, int *fp)
+{
+ if (bp != NULL)
+ *bp = (tag >> 20) & 0xff;
+ if (dp != NULL)
+ *dp = (tag >> 15) & 0x1f;
+ if (fp != NULL)
+ *fp = (tag >> 12) & 0x7;
+}
+
+int
+mvkpcie_conf_size(void *v, pcitag_t tag)
+{
+ return PCIE_CONFIG_SPACE_SIZE;
+}
+
+pcireg_t
+mvkpcie_conf_read_bridge(struct mvkpcie_softc *sc, int reg)
+{
+ switch (reg) {
+ case PCI_ID_REG:
+ return PCI_VENDOR_MARVELL |
+ (HREAD4(sc, PCIE_DEV_ID) & 0xffff0000);
+ case PCI_COMMAND_STATUS_REG:
+ return sc->sc_bridge_command;
+ case PCI_CLASS_REG:
+ return PCI_CLASS_BRIDGE << PCI_CLASS_SHIFT |
+ PCI_SUBCLASS_BRIDGE_PCI << PCI_SUBCLASS_SHIFT |
+ (HREAD4(sc, PCIE_DEV_REV) & 0xff);
+ case PCI_BHLC_REG:
+ return 1 << PCI_HDRTYPE_SHIFT |
+ 0x10 << PCI_CACHELINE_SHIFT;
+ case PPB_REG_BASE0:
+ return sc->sc_bridge_bar0;
+ case PPB_REG_BASE1:
+ return sc->sc_bridge_bar1;
+ case PPB_REG_BUSINFO:
+ return sc->sc_bridge_businfo;
+ case PPB_REG_IOSTATUS:
+ return sc->sc_bridge_iostatus;
+ case PPB_REG_MEM:
+ return sc->sc_bridge_mem;
+ case PPB_REG_IO_HI:
+ return sc->sc_bridge_io_hi;
+ case PPB_REG_PREFMEM:
+ case PPB_REG_PREFBASE_HI32:
+ case PPB_REG_PREFLIM_HI32:
+ case PPB_REG_BRIDGECONTROL:
+ return 0;
+ default:
+ printf("%s: reg %x\n", __func__, reg);
+ break;
+ }
+ return 0;
+}
+
+void
+mvkpcie_conf_write_bridge(struct mvkpcie_softc *sc, int reg, pcireg_t data)
+{
+ switch (reg) {
+ case PCI_COMMAND_STATUS_REG: {
+ sc->sc_bridge_command = data & 0xffffff;
+ break;
+ }
+ case PPB_REG_BASE0:
+ sc->sc_bridge_bar0 = data;
+ break;
+ case PPB_REG_BASE1:
+ sc->sc_bridge_bar1 = data;
+ break;
+ case PPB_REG_BUSINFO:
+ sc->sc_bridge_businfo = data;
+ break;
+ case PPB_REG_IOSTATUS:
+ sc->sc_bridge_iostatus = data;
+ break;
+ case PPB_REG_MEM:
+ sc->sc_bridge_mem = data;
+ break;
+ case PPB_REG_IO_HI:
+ sc->sc_bridge_io_hi = data;
+ break;
+ case PPB_REG_PREFMEM:
+ case PPB_REG_PREFBASE_HI32:
+ case PPB_REG_PREFLIM_HI32:
+ break;
+ default:
+ printf("%s: reg %x data %x\n", __func__, reg, data);
+ break;
+ }
+}
+
+pcireg_t
+mvkpcie_conf_read(void *v, pcitag_t tag, int off)
+{
+ struct mvkpcie_softc *sc = v;
+ int bus, dev, fn;
+ uint32_t reg;
+ int i;
+
+ mvkpcie_decompose_tag(sc, tag, &bus, &dev, &fn);
+ if (bus == sc->sc_bus) {
+ KASSERT(dev == 0);
+ return mvkpcie_conf_read_bridge(sc, off);
+ }
+
+ HWRITE4(sc, PIO_START, PIO_START_STOP);
+ HWRITE4(sc, PIO_ISR, PIO_ISR_CLEAR);
+ reg = HREAD4(sc, PIO_CTRL);
+ reg &= ~PIO_CTRL_TYPE_MASK;
+ if (bus == sc->sc_bus + 1)
+ reg |= PIO_CTRL_TYPE_RD0;
+ else
+ reg |= PIO_CTRL_TYPE_RD1;
+ HWRITE4(sc, PIO_CTRL, reg);
+ HWRITE4(sc, PIO_ADDR_LS, tag | off);
+ HWRITE4(sc, PIO_ADDR_MS, 0);
+ HWRITE4(sc, PIO_WR_DATA_STRB, PIO_WR_DATA_STRB_VALUE);
+ HWRITE4(sc, PIO_START, PIO_START_START);
+
+ for (i = 500; i > 0; i--) {
+ if (HREAD4(sc, PIO_START) == 0 &&
+ HREAD4(sc, PIO_ISR) != 0)
+ break;
+ delay(2);
+ }
+ if (i == 0) {
+ printf("%s: timeout\n", sc->sc_dev.dv_xname);
+ return 0xffffffff;
+ }
+
+ return HREAD4(sc, PIO_RD_DATA);
+}
+
+void
+mvkpcie_conf_write(void *v, pcitag_t tag, int off, pcireg_t data)
+{
+ struct mvkpcie_softc *sc = v;
+ int bus, dev, fn;
+ uint32_t reg;
+ int i;
+
+ mvkpcie_decompose_tag(sc, tag, &bus, &dev, &fn);
+ if (bus == sc->sc_bus) {
+ KASSERT(dev == 0);
+ mvkpcie_conf_write_bridge(sc, off, data);
+ return;
+ }
+
+ HWRITE4(sc, PIO_START, PIO_START_STOP);
+ HWRITE4(sc, PIO_ISR, PIO_ISR_CLEAR);
+ reg = HREAD4(sc, PIO_CTRL);
+ reg &= ~PIO_CTRL_TYPE_MASK;
+ if (bus == sc->sc_bus + 1)
+ reg |= PIO_CTRL_TYPE_WR0;
+ else
+ reg |= PIO_CTRL_TYPE_WR1;
+ HWRITE4(sc, PIO_CTRL, reg);
+ HWRITE4(sc, PIO_ADDR_LS, tag | off);
+ HWRITE4(sc, PIO_ADDR_MS, 0);
+ HWRITE4(sc, PIO_WR_DATA, data);
+ HWRITE4(sc, PIO_WR_DATA_STRB, PIO_WR_DATA_STRB_VALUE);
+ HWRITE4(sc, PIO_START, PIO_START_START);
+
+ for (i = 500; i > 0; i--) {
+ if (HREAD4(sc, PIO_START) == 0 &&
+ HREAD4(sc, PIO_ISR) != 0)
+ break;
+ delay(2);
+ }
+ if (i == 0) {
+ printf("%s: timeout\n", sc->sc_dev.dv_xname);
+ return;
+ }
+}
+
+int
+mvkpcie_intr_map(struct pci_attach_args *pa, pci_intr_handle_t *ihp)
+{
+ return -1;
+}
+
+const char *
+mvkpcie_intr_string(void *v, pci_intr_handle_t ih)
+{
+ switch (ih.ih_type) {
+ case PCI_MSI:
+ return "msi";
+ case PCI_MSIX:
+ return "msix";
+ }
+
+ return "intx";
+}
+
+void *
+mvkpcie_intr_establish(void *v, pci_intr_handle_t ih, int level,
+ int (*func)(void *), void *arg, char *name)
+{
+ struct mvkpcie_softc *sc = v;
+ void *cookie;
+
+ KASSERT(ih.ih_type != PCI_NONE);
+
+ if (ih.ih_type != PCI_INTX) {
+ uint64_t addr, data;
+
+ /* Assume hardware passes Requester ID as sideband data. */
+ data = pci_requester_id(ih.ih_pc, ih.ih_tag);
+ cookie = fdt_intr_establish_msi(sc->sc_node, &addr,
+ &data, level, func, arg, (void *)name);
+ if (cookie == NULL)
+ return NULL;
+
+ /* TODO: translate address to the PCI device's view */
+
+ if (ih.ih_type == PCI_MSIX) {
+ pci_msix_enable(ih.ih_pc, ih.ih_tag,
+ &sc->sc_bus_memt, ih.ih_intrpin, addr, data);
+ } else
+ pci_msi_enable(ih.ih_pc, ih.ih_tag, addr, data);
+ } else {
+ int bus, dev, fn;
+ uint32_t reg[4];
+
+ mvkpcie_decompose_tag(sc, ih.ih_tag, &bus, &dev, &fn);
+
+ reg[0] = bus << 16 | dev << 11 | fn << 8;
+ reg[1] = reg[2] = 0;
+ reg[3] = ih.ih_intrpin;
+
+ cookie = fdt_intr_establish_imap(sc->sc_node, reg,
+ sizeof(reg), level, func, arg, name);
+ }
+
+ return cookie;
+}
+
+void
+mvkpcie_intr_disestablish(void *v, void *cookie)
+{
+ panic("%s", __func__);
+}
+
+int
+mvkpcie_bs_iomap(bus_space_tag_t t, bus_addr_t addr, bus_size_t size,
+ int flags, bus_space_handle_t *bshp)
+{
+ struct mvkpcie_softc *sc = t->bus_private;
+ int i;
+
+ for (i = 0; i < sc->sc_nranges; i++) {
+ uint64_t pci_start = sc->sc_ranges[i].pci_base;
+ uint64_t pci_end = pci_start + sc->sc_ranges[i].size;
+ uint64_t phys_start = sc->sc_ranges[i].phys_base;
+
+ if ((sc->sc_ranges[i].flags & 0x03000000) == 0x01000000 &&
+ addr >= pci_start && addr + size <= pci_end) {
+ return bus_space_map(sc->sc_iot,
+ addr - pci_start + phys_start, size, flags, bshp);
+ }
+ }
+
+ return ENXIO;
+}
+
+int
+mvkpcie_bs_memmap(bus_space_tag_t t, bus_addr_t addr, bus_size_t size,
+ int flags, bus_space_handle_t *bshp)
+{
+ struct mvkpcie_softc *sc = t->bus_private;
+ int i;
+
+ for (i = 0; i < sc->sc_nranges; i++) {
+ uint64_t pci_start = sc->sc_ranges[i].pci_base;
+ uint64_t pci_end = pci_start + sc->sc_ranges[i].size;
+ uint64_t phys_start = sc->sc_ranges[i].phys_base;
+
+ if ((sc->sc_ranges[i].flags & 0x03000000) == 0x02000000 &&
+ addr >= pci_start && addr + size <= pci_end) {
+ return bus_space_map(sc->sc_iot,
+ addr - pci_start + phys_start, size, flags, bshp);
+ }
+ }
+
+ return ENXIO;
+}
+
+int
+mvkpcie_intc_intr(void *cookie)
+{
+ struct mvkpcie_softc *sc = (struct mvkpcie_softc *)cookie;
+ struct intrhand *ih;
+ uint32_t pending;
+ int i, s;
+
+ if (!(HREAD4(sc, HOST_CTRL_INT_STATUS) & HOST_CTRL_INT_MASK_CORE_INT))
+ return 0;
+
+ if (!(HREAD4(sc, PCIE_CORE_ISR0_STATUS) & PCIE_CORE_ISR0_MASK_MSI_INT))
+ return 0;
+
+ pending = HREAD4(sc, PCIE_CORE_MSI_STATUS);
+ while (pending) {
+ i = ffs(pending) - 1;
+ HWRITE4(sc, PCIE_CORE_MSI_STATUS, (1 << i));
+ pending &= ~(1 << i);
+
+ i = HREAD4(sc, PCIE_CORE_MSI_PAYLOAD) & 0xff;
+ if ((ih = sc->sc_msi_handlers[i]) != NULL) {
+ s = splraise(ih->ih_ipl);
+ if (ih->ih_func(ih->ih_arg))
+ ih->ih_count.ec_count++;
+ splx(s);
+ }
+ }
+
+ HWRITE4(sc, PCIE_CORE_ISR0_STATUS, PCIE_CORE_ISR0_MASK_MSI_INT);
+ HWRITE4(sc, HOST_CTRL_INT_STATUS, HOST_CTRL_INT_MASK_CORE_INT);
+ return 1;
+}
+
+void *
+mvkpcie_intc_intr_establish_msi(void *cookie, uint64_t *addr, uint64_t *data,
+ int level, int (*func)(void *), void *arg, char *name)
+{
+ struct mvkpcie_softc *sc = (struct mvkpcie_softc *)cookie;
+ struct intrhand *ih;
+ int i, s;
+
+ for (i = 0; i < nitems(sc->sc_msi_handlers); i++) {
+ if (sc->sc_msi_handlers[i] == NULL)
+ break;
+ }
+
+ if (i == nitems(sc->sc_msi_handlers))
+ return NULL;
+
+ ih = malloc(sizeof(*ih), M_DEVBUF, M_WAITOK);
+ ih->ih_func = func;
+ ih->ih_arg = arg;
+ ih->ih_ipl = level & IPL_IRQMASK;
+ ih->ih_irq = i;
+ ih->ih_name = name;
+ ih->ih_sc = sc;
+
+ s = splhigh();
+
+ sc->sc_msi_handlers[i] = ih;
+
+ if (name != NULL)
+ evcount_attach(&ih->ih_count, name, &ih->ih_irq);
+
+ mvkpcie_intc_recalc_ipl(sc);
+
+ *addr = MVKPCIE_DMA_DVA(sc->sc_msi_addr);
+ *data = i;
+
+ splx(s);
+ return (ih);
+}
+
+void
+mvkpcie_intc_intr_disestablish_msi(void *cookie)
+{
+ struct intrhand *ih = cookie;
+ struct mvkpcie_softc *sc = ih->ih_sc;
+ int s;
+
+ s = splhigh();
+
+ sc->sc_msi_handlers[ih->ih_irq] = NULL;
+ if (ih->ih_name != NULL)
+ evcount_detach(&ih->ih_count);
+ free(ih, M_DEVBUF, sizeof(*ih));
+
+ mvkpcie_intc_recalc_ipl(sc);
+
+ splx(s);
+}
+
+void
+mvkpcie_intc_recalc_ipl(struct mvkpcie_softc *sc)
+{
+ struct intrhand *ih;
+ int max = IPL_NONE;
+ int min = IPL_HIGH;
+ int irq;
+
+ for (irq = 0; irq < nitems(sc->sc_msi_handlers); irq++) {
+ ih = sc->sc_msi_handlers[irq];
+ if (ih == NULL)
+ continue;
+
+ if (ih->ih_ipl > max)
+ max = ih->ih_ipl;
+
+ if (ih->ih_ipl < min)
+ min = ih->ih_ipl;
+ }
+
+ for (irq = 0; irq < nitems(sc->sc_msi_handlers); irq++) {
+ ih = sc->sc_msi_handlers[irq];
+ if (ih == NULL)
+ continue;
+
+ if (ih->ih_ipl > max)
+ max = ih->ih_ipl;
+
+ if (ih->ih_ipl < min)
+ min = ih->ih_ipl;
+ }
+
+ if (max == IPL_NONE)
+ min = IPL_NONE;
+
+ if (sc->sc_ipl != max) {
+ sc->sc_ipl = max;
+
+ if (sc->sc_ih != NULL)
+ fdt_intr_disestablish(sc->sc_ih);
+
+ if (sc->sc_ipl != IPL_NONE)
+ sc->sc_ih = fdt_intr_establish(sc->sc_node, sc->sc_ipl,
+ mvkpcie_intc_intr, sc, sc->sc_dev.dv_xname);
+ }
+}
+
+/* Only needed for the 16-bit MSI address */
+struct mvkpcie_dmamem *
+mvkpcie_dmamem_alloc(struct mvkpcie_softc *sc, bus_size_t size, bus_size_t align)
+{
+ struct mvkpcie_dmamem *mdm;
+ int nsegs;
+
+ mdm = malloc(sizeof(*mdm), M_DEVBUF, M_WAITOK | M_ZERO);
+ mdm->mdm_size = size;
+
+ if (bus_dmamap_create(sc->sc_dmat, size, 1, size, 0,
+ BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW, &mdm->mdm_map) != 0)
+ goto mdmfree;
+
+ if (bus_dmamem_alloc(sc->sc_dmat, size, align, 0, &mdm->mdm_seg, 1,
+ &nsegs, BUS_DMA_WAITOK) != 0)
+ goto destroy;
+
+ if (bus_dmamem_map(sc->sc_dmat, &mdm->mdm_seg, nsegs, size,
+ &mdm->mdm_kva, BUS_DMA_WAITOK | BUS_DMA_COHERENT) != 0)
+ goto free;
+
+ if (bus_dmamap_load(sc->sc_dmat, mdm->mdm_map, mdm->mdm_kva, size,
+ NULL, BUS_DMA_WAITOK) != 0)
+ goto unmap;
+
+ bzero(mdm->mdm_kva, size);
+
+ return (mdm);
+
+unmap:
+ bus_dmamem_unmap(sc->sc_dmat, mdm->mdm_kva, size);
+free:
+ bus_dmamem_free(sc->sc_dmat, &mdm->mdm_seg, 1);
+destroy:
+ bus_dmamap_destroy(sc->sc_dmat, mdm->mdm_map);
+mdmfree:
+ free(mdm, M_DEVBUF, sizeof(*mdm));
+
+ return (NULL);
+}
+
+void
+mvkpcie_dmamem_free(struct mvkpcie_softc *sc, struct mvkpcie_dmamem *mdm)
+{
+ bus_dmamem_unmap(sc->sc_dmat, mdm->mdm_kva, mdm->mdm_size);
+ bus_dmamem_free(sc->sc_dmat, &mdm->mdm_seg, 1);
+ bus_dmamap_destroy(sc->sc_dmat, mdm->mdm_map);
+ free(mdm, M_DEVBUF, sizeof(*mdm));
+}