/* * linux/arch/arm/mach-mmp/irq.c * * Generic IRQ handling, GPIO IRQ demultiplexing, etc. * Copyright (C) 2008 - 2012 Marvell Technology Group Ltd. * * Author: Bin Yang * Haojian Zhuang * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. */ #include #include #include #include #include #include #include #include #include #ifdef CONFIG_CPU_MMP2 #include #endif #ifdef CONFIG_CPU_PXA910 #include #endif #include "common.h" #define MAX_ICU_NR 16 struct icu_chip_data { int nr_irqs; unsigned int virq_base; unsigned int cascade_irq; void __iomem *reg_status; void __iomem *reg_mask; unsigned int conf_enable; unsigned int conf_disable; unsigned int conf_mask; unsigned int clr_mfp_irq_base; unsigned int clr_mfp_hwirq; struct irq_domain *domain; }; struct mmp_intc_conf { unsigned int conf_enable; unsigned int conf_disable; unsigned int conf_mask; }; void __iomem *mmp_icu_base; static struct icu_chip_data icu_data[MAX_ICU_NR]; static int max_icu_nr; extern void mmp2_clear_pmic_int(void); static void icu_mask_ack_irq(struct irq_data *d) { struct irq_domain *domain = d->domain; struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data; int hwirq; u32 r; hwirq = d->irq - data->virq_base; if (data == &icu_data[0]) { r = readl_relaxed(mmp_icu_base + (hwirq << 2)); r &= ~data->conf_mask; r |= data->conf_disable; writel_relaxed(r, mmp_icu_base + (hwirq << 2)); } else { #ifdef CONFIG_CPU_MMP2 if ((data->virq_base == data->clr_mfp_irq_base) && (hwirq == data->clr_mfp_hwirq)) mmp2_clear_pmic_int(); #endif r = readl_relaxed(data->reg_mask) | (1 << hwirq); writel_relaxed(r, data->reg_mask); } } static void icu_mask_irq(struct irq_data *d) { struct irq_domain *domain = d->domain; struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data; int hwirq; u32 r; hwirq = d->irq - data->virq_base; if (data == &icu_data[0]) { r = readl_relaxed(mmp_icu_base + (hwirq << 2)); r &= ~data->conf_mask; r |= data->conf_disable; writel_relaxed(r, mmp_icu_base + (hwirq << 2)); } else { r = readl_relaxed(data->reg_mask) | (1 << hwirq); writel_relaxed(r, data->reg_mask); } } static void icu_unmask_irq(struct irq_data *d) { struct irq_domain *domain = d->domain; struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data; int hwirq; u32 r; hwirq = d->irq - data->virq_base; if (data == &icu_data[0]) { r = readl_relaxed(mmp_icu_base + (hwirq << 2)); r &= ~data->conf_mask; r |= data->conf_enable; writel_relaxed(r, mmp_icu_base + (hwirq << 2)); } else { r = readl_relaxed(data->reg_mask) & ~(1 << hwirq); writel_relaxed(r, data->reg_mask); } } static struct irq_chip icu_irq_chip = { .name = "icu_irq", .irq_mask = icu_mask_irq, .irq_mask_ack = icu_mask_ack_irq, .irq_unmask = icu_unmask_irq, }; static void icu_mux_irq_demux(unsigned int irq, struct irq_desc *desc) { struct irq_domain *domain; struct icu_chip_data *data; int i; unsigned long mask, status, n; for (i = 1; i < max_icu_nr; i++) { if (irq == icu_data[i].cascade_irq) { domain = icu_data[i].domain; data = (struct icu_chip_data *)domain->host_data; break; } } if (i >= max_icu_nr) { pr_err("Spurious irq %d in MMP INTC\n", irq); return; } mask = readl_relaxed(data->reg_mask); while (1) { status = readl_relaxed(data->reg_status) & ~mask; if (status == 0) break; n = find_first_bit(&status, BITS_PER_LONG); while (n < BITS_PER_LONG) { generic_handle_irq(icu_data[i].virq_base + n); n = find_next_bit(&status, BITS_PER_LONG, n + 1); } } } static int mmp_irq_domain_map(struct irq_domain *d, unsigned int irq, irq_hw_number_t hw) { irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq); set_irq_flags(irq, IRQF_VALID); return 0; } static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node, const u32 *intspec, unsigned int intsize, unsigned long *out_hwirq, unsigned int *out_type) { *out_hwirq = intspec[0]; return 0; } const struct irq_domain_ops mmp_irq_domain_ops = { .map = mmp_irq_domain_map, .xlate = mmp_irq_domain_xlate, }; static struct mmp_intc_conf mmp_conf = { .conf_enable = 0x51, .conf_disable = 0x0, .conf_mask = 0x7f, }; static struct mmp_intc_conf mmp2_conf = { .conf_enable = 0x20, .conf_disable = 0x0, .conf_mask = 0x7f, }; /* MMP (ARMv5) */ void __init icu_init_irq(void) { int irq; max_icu_nr = 1; mmp_icu_base = ioremap(0xd4282000, 0x1000); icu_data[0].conf_enable = mmp_conf.conf_enable; icu_data[0].conf_disable = mmp_conf.conf_disable; icu_data[0].conf_mask = mmp_conf.conf_mask; icu_data[0].nr_irqs = 64; icu_data[0].virq_base = 0; icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0, &irq_domain_simple_ops, &icu_data[0]); for (irq = 0; irq < 64; irq++) { icu_mask_irq(irq_get_irq_data(irq)); irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq); set_irq_flags(irq, IRQF_VALID); } irq_set_default_host(icu_data[0].domain); #ifdef CONFIG_CPU_PXA910 icu_irq_chip.irq_set_wake = pxa910_set_wake; #endif } /* MMP2 (ARMv7) */ void __init mmp2_init_icu(void) { int irq; max_icu_nr = 8; mmp_icu_base = ioremap(0xd4282000, 0x1000); icu_data[0].conf_enable = mmp2_conf.conf_enable; icu_data[0].conf_disable = mmp2_conf.conf_disable; icu_data[0].conf_mask = mmp2_conf.conf_mask; icu_data[0].nr_irqs = 64; icu_data[0].virq_base = 0; icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0, &irq_domain_simple_ops, &icu_data[0]); icu_data[1].reg_status = mmp_icu_base + 0x150; icu_data[1].reg_mask = mmp_icu_base + 0x168; icu_data[1].clr_mfp_irq_base = IRQ_MMP2_PMIC_BASE; icu_data[1].clr_mfp_hwirq = IRQ_MMP2_PMIC - IRQ_MMP2_PMIC_BASE; icu_data[1].nr_irqs = 2; icu_data[1].cascade_irq = 4; icu_data[1].virq_base = IRQ_MMP2_PMIC_BASE; icu_data[1].domain = irq_domain_add_legacy(NULL, icu_data[1].nr_irqs, icu_data[1].virq_base, 0, &irq_domain_simple_ops, &icu_data[1]); icu_data[2].reg_status = mmp_icu_base + 0x154; icu_data[2].reg_mask = mmp_icu_base + 0x16c; icu_data[2].nr_irqs = 2; icu_data[2].cascade_irq = 5; icu_data[2].virq_base = IRQ_MMP2_RTC_BASE; icu_data[2].domain = irq_domain_add_legacy(NULL, icu_data[2].nr_irqs, icu_data[2].virq_base, 0, &irq_domain_simple_ops, &icu_data[2]); icu_data[3].reg_status = mmp_icu_base + 0x180; icu_data[3].reg_mask = mmp_icu_base + 0x17c; icu_data[3].nr_irqs = 3; icu_data[3].cascade_irq = 9; icu_data[3].virq_base = IRQ_MMP2_KEYPAD_BASE; icu_data[3].domain = irq_domain_add_legacy(NULL, icu_data[3].nr_irqs, icu_data[3].virq_base, 0, &irq_domain_simple_ops, &icu_data[3]); icu_data[4].reg_status = mmp_icu_base + 0x158; icu_data[4].reg_mask = mmp_icu_base + 0x170; icu_data[4].nr_irqs = 5; icu_data[4].cascade_irq = 17; icu_data[4].virq_base = IRQ_MMP2_TWSI_BASE; icu_data[4].domain = irq_domain_add_legacy(NULL, icu_data[4].nr_irqs, icu_data[4].virq_base, 0, &irq_domain_simple_ops, &icu_data[4]); icu_data[5].reg_status = mmp_icu_base + 0x15c; icu_data[5].reg_mask = mmp_icu_base + 0x174; icu_data[5].nr_irqs = 15; icu_data[5].cascade_irq = 35; icu_data[5].virq_base = IRQ_MMP2_MISC_BASE; icu_data[5].domain = irq_domain_add_legacy(NULL, icu_data[5].nr_irqs, icu_data[5].virq_base, 0, &irq_domain_simple_ops, &icu_data[5]); icu_data[6].reg_status = mmp_icu_base + 0x160; icu_data[6].reg_mask = mmp_icu_base + 0x178; icu_data[6].nr_irqs = 2; icu_data[6].cascade_irq = 51; icu_data[6].virq_base = IRQ_MMP2_MIPI_HSI1_BASE; icu_data[6].domain = irq_domain_add_legacy(NULL, icu_data[6].nr_irqs, icu_data[6].virq_base, 0, &irq_domain_simple_ops, &icu_data[6]); icu_data[7].reg_status = mmp_icu_base + 0x188; icu_data[7].reg_mask = mmp_icu_base + 0x184; icu_data[7].nr_irqs = 2; icu_data[7].cascade_irq = 55; icu_data[7].virq_base = IRQ_MMP2_MIPI_HSI0_BASE; icu_data[7].domain = irq_domain_add_legacy(NULL, icu_data[7].nr_irqs, icu_data[7].virq_base, 0, &irq_domain_simple_ops, &icu_data[7]); for (irq = 0; irq < IRQ_MMP2_MUX_END; irq++) { icu_mask_irq(irq_get_irq_data(irq)); switch (irq) { case IRQ_MMP2_PMIC_MUX: case IRQ_MMP2_RTC_MUX: case IRQ_MMP2_KEYPAD_MUX: case IRQ_MMP2_TWSI_MUX: case IRQ_MMP2_MISC_MUX: case IRQ_MMP2_MIPI_HSI1_MUX: case IRQ_MMP2_MIPI_HSI0_MUX: irq_set_chip(irq, &icu_irq_chip); irq_set_chained_handler(irq, icu_mux_irq_demux); break; default: irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq); break; } set_irq_flags(irq, IRQF_VALID); } irq_set_default_host(icu_data[0].domain); #ifdef CONFIG_CPU_MMP2 icu_irq_chip.irq_set_wake = mmp2_set_wake; #endif } #ifdef CONFIG_OF static const struct of_device_id intc_ids[] __initconst = { { .compatible = "mrvl,mmp-intc", .data = &mmp_conf }, { .compatible = "mrvl,mmp2-intc", .data = &mmp2_conf }, {} }; static const struct of_device_id mmp_mux_irq_match[] __initconst = { { .compatible = "mrvl,mmp2-mux-intc" }, {} }; int __init mmp2_mux_init(struct device_node *parent) { struct device_node *node; const struct of_device_id *of_id; struct resource res; int i, irq_base, ret, irq; u32 nr_irqs, mfp_irq; node = parent; max_icu_nr = 1; for (i = 1; i < MAX_ICU_NR; i++) { node = of_find_matching_node(node, mmp_mux_irq_match); if (!node) break; of_id = of_match_node(&mmp_mux_irq_match[0], node); ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs); if (ret) { pr_err("Not found mrvl,intc-nr-irqs property\n"); ret = -EINVAL; goto err; } ret = of_address_to_resource(node, 0, &res); if (ret < 0) { pr_err("Not found reg property\n"); ret = -EINVAL; goto err; } icu_data[i].reg_status = mmp_icu_base + res.start; ret = of_address_to_resource(node, 1, &res); if (ret < 0) { pr_err("Not found reg property\n"); ret = -EINVAL; goto err; } icu_data[i].reg_mask = mmp_icu_base + res.start; icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0); if (!icu_data[i].cascade_irq) { ret = -EINVAL; goto err; } irq_base = irq_alloc_descs(-1, 0, nr_irqs, 0); if (irq_base < 0) { pr_err("Failed to allocate IRQ numbers for mux intc\n"); ret = irq_base; goto err; } if (!of_property_read_u32(node, "mrvl,clr-mfp-irq", &mfp_irq)) { icu_data[i].clr_mfp_irq_base = irq_base; icu_data[i].clr_mfp_hwirq = mfp_irq; } irq_set_chained_handler(icu_data[i].cascade_irq, icu_mux_irq_demux); icu_data[i].nr_irqs = nr_irqs; icu_data[i].virq_base = irq_base; icu_data[i].domain = irq_domain_add_legacy(node, nr_irqs, irq_base, 0, &mmp_irq_domain_ops, &icu_data[i]); for (irq = irq_base; irq < irq_base + nr_irqs; irq++) icu_mask_irq(irq_get_irq_data(irq)); } max_icu_nr = i; return 0; err: of_node_put(node); max_icu_nr = i; return ret; } void __init mmp_dt_irq_init(void) { struct device_node *node; const struct of_device_id *of_id; struct mmp_intc_conf *conf; int nr_irqs, irq_base, ret, irq; node = of_find_matching_node(NULL, intc_ids); if (!node) { pr_err("Failed to find interrupt controller in arch-mmp\n"); return; } of_id = of_match_node(intc_ids, node); conf = of_id->data; ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs); if (ret) { pr_err("Not found mrvl,intc-nr-irqs property\n"); return; } mmp_icu_base = of_iomap(node, 0); if (!mmp_icu_base) { pr_err("Failed to get interrupt controller register\n"); return; } irq_base = irq_alloc_descs(-1, 0, nr_irqs - NR_IRQS_LEGACY, 0); if (irq_base < 0) { pr_err("Failed to allocate IRQ numbers\n"); goto err; } else if (irq_base != NR_IRQS_LEGACY) { pr_err("ICU's irqbase should be started from 0\n"); goto err; } icu_data[0].conf_enable = conf->conf_enable; icu_data[0].conf_disable = conf->conf_disable; icu_data[0].conf_mask = conf->conf_mask; icu_data[0].nr_irqs = nr_irqs; icu_data[0].virq_base = 0; icu_data[0].domain = irq_domain_add_legacy(node, nr_irqs, 0, 0, &mmp_irq_domain_ops, &icu_data[0]); irq_set_default_host(icu_data[0].domain); for (irq = 0; irq < nr_irqs; irq++) icu_mask_irq(irq_get_irq_data(irq)); mmp2_mux_init(node); return; err: iounmap(mmp_icu_base); } #endif