mirror of
https://github.com/openwrt/openwrt.git
synced 2024-12-27 01:11:14 +00:00
394 lines
11 KiB
Diff
394 lines
11 KiB
Diff
|
--- a/drivers/irqchip/irq-realtek-rtl.c
|
||
|
+++ b/drivers/irqchip/irq-realtek-rtl.c
|
||
|
@@ -21,21 +21,63 @@
|
||
|
#define RTL_ICTL_IRR2 0x10
|
||
|
#define RTL_ICTL_IRR3 0x14
|
||
|
|
||
|
-#define REG(x) (realtek_ictl_base + x)
|
||
|
+#define RTL_ICTL_NUM_INPUTS 32
|
||
|
+#define RTL_ICTL_NUM_OUTPUTS 15
|
||
|
|
||
|
static DEFINE_RAW_SPINLOCK(irq_lock);
|
||
|
-static void __iomem *realtek_ictl_base;
|
||
|
+
|
||
|
+#define REG(offset, cpu) (realtek_ictl_base[cpu] + offset)
|
||
|
+
|
||
|
+static void __iomem *realtek_ictl_base[NR_CPUS];
|
||
|
+static cpumask_t realtek_ictl_cpu_configurable;
|
||
|
+
|
||
|
+struct realtek_ictl_output {
|
||
|
+ /* IRQ controller data */
|
||
|
+ struct fwnode_handle *fwnode;
|
||
|
+ /* Output specific data */
|
||
|
+ unsigned int output_index;
|
||
|
+ struct irq_domain *domain;
|
||
|
+ u32 child_mask;
|
||
|
+};
|
||
|
+
|
||
|
+/*
|
||
|
+ * IRR0-IRR3 store 4 bits per interrupt, but Realtek uses inverted numbering,
|
||
|
+ * placing IRQ 31 in the first four bits. A routing value of '0' means the
|
||
|
+ * interrupt is left disconnected. Routing values {1..15} connect to output
|
||
|
+ * lines {0..14}.
|
||
|
+ */
|
||
|
+#define IRR_OFFSET(idx) (4 * (3 - (idx * 4) / 32))
|
||
|
+#define IRR_SHIFT(idx) ((idx * 4) % 32)
|
||
|
+
|
||
|
+static inline u32 read_irr(void __iomem *irr0, int idx)
|
||
|
+{
|
||
|
+ return (readl(irr0 + IRR_OFFSET(idx)) >> IRR_SHIFT(idx)) & 0xf;
|
||
|
+}
|
||
|
+
|
||
|
+static inline void write_irr(void __iomem *irr0, int idx, u32 value)
|
||
|
+{
|
||
|
+ unsigned int offset = IRR_OFFSET(idx);
|
||
|
+ unsigned int shift = IRR_SHIFT(idx);
|
||
|
+ u32 irr;
|
||
|
+
|
||
|
+ irr = readl(irr0 + offset) & ~(0xf << shift);
|
||
|
+ irr |= (value & 0xf) << shift;
|
||
|
+ writel(irr, irr0 + offset);
|
||
|
+}
|
||
|
|
||
|
static void realtek_ictl_unmask_irq(struct irq_data *i)
|
||
|
{
|
||
|
unsigned long flags;
|
||
|
u32 value;
|
||
|
+ int cpu;
|
||
|
|
||
|
raw_spin_lock_irqsave(&irq_lock, flags);
|
||
|
|
||
|
- value = readl(REG(RTL_ICTL_GIMR));
|
||
|
- value |= BIT(i->hwirq);
|
||
|
- writel(value, REG(RTL_ICTL_GIMR));
|
||
|
+ for_each_cpu(cpu, &realtek_ictl_cpu_configurable) {
|
||
|
+ value = readl(REG(RTL_ICTL_GIMR, cpu));
|
||
|
+ value |= BIT(i->hwirq);
|
||
|
+ writel(value, REG(RTL_ICTL_GIMR, cpu));
|
||
|
+ }
|
||
|
|
||
|
raw_spin_unlock_irqrestore(&irq_lock, flags);
|
||
|
}
|
||
|
@@ -44,137 +86,247 @@ static void realtek_ictl_mask_irq(struct
|
||
|
{
|
||
|
unsigned long flags;
|
||
|
u32 value;
|
||
|
+ int cpu;
|
||
|
|
||
|
raw_spin_lock_irqsave(&irq_lock, flags);
|
||
|
|
||
|
- value = readl(REG(RTL_ICTL_GIMR));
|
||
|
- value &= ~BIT(i->hwirq);
|
||
|
- writel(value, REG(RTL_ICTL_GIMR));
|
||
|
+ for_each_cpu(cpu, &realtek_ictl_cpu_configurable) {
|
||
|
+ value = readl(REG(RTL_ICTL_GIMR, cpu));
|
||
|
+ value &= ~BIT(i->hwirq);
|
||
|
+ writel(value, REG(RTL_ICTL_GIMR, cpu));
|
||
|
+ }
|
||
|
|
||
|
raw_spin_unlock_irqrestore(&irq_lock, flags);
|
||
|
}
|
||
|
|
||
|
+static int __maybe_unused realtek_ictl_irq_affinity(struct irq_data *i,
|
||
|
+ const struct cpumask *dest, bool force)
|
||
|
+{
|
||
|
+ struct realtek_ictl_output *output = i->domain->host_data;
|
||
|
+ cpumask_t cpu_configure;
|
||
|
+ cpumask_t cpu_disable;
|
||
|
+ cpumask_t cpu_enable;
|
||
|
+ unsigned long flags;
|
||
|
+ int cpu;
|
||
|
+
|
||
|
+ raw_spin_lock_irqsave(&irq_lock, flags);
|
||
|
+
|
||
|
+ cpumask_and(&cpu_configure, cpu_present_mask, &realtek_ictl_cpu_configurable);
|
||
|
+
|
||
|
+ cpumask_and(&cpu_enable, &cpu_configure, dest);
|
||
|
+ cpumask_andnot(&cpu_disable, &cpu_configure, dest);
|
||
|
+
|
||
|
+ for_each_cpu(cpu, &cpu_disable)
|
||
|
+ write_irr(REG(RTL_ICTL_IRR0, cpu), i->hwirq, 0);
|
||
|
+
|
||
|
+ for_each_cpu(cpu, &cpu_enable)
|
||
|
+ write_irr(REG(RTL_ICTL_IRR0, cpu), i->hwirq, output->output_index + 1);
|
||
|
+
|
||
|
+ irq_data_update_effective_affinity(i, &cpu_enable);
|
||
|
+
|
||
|
+ raw_spin_unlock_irqrestore(&irq_lock, flags);
|
||
|
+
|
||
|
+ return IRQ_SET_MASK_OK;
|
||
|
+}
|
||
|
+
|
||
|
static struct irq_chip realtek_ictl_irq = {
|
||
|
.name = "realtek-rtl-intc",
|
||
|
.irq_mask = realtek_ictl_mask_irq,
|
||
|
.irq_unmask = realtek_ictl_unmask_irq,
|
||
|
+#ifdef CONFIG_SMP
|
||
|
+ .irq_set_affinity = realtek_ictl_irq_affinity,
|
||
|
+#endif
|
||
|
};
|
||
|
|
||
|
static int intc_map(struct irq_domain *d, unsigned int irq, irq_hw_number_t hw)
|
||
|
{
|
||
|
- irq_set_chip_and_handler(hw, &realtek_ictl_irq, handle_level_irq);
|
||
|
+ struct realtek_ictl_output *output = d->host_data;
|
||
|
+ unsigned long flags;
|
||
|
+
|
||
|
+ irq_set_chip_and_handler(irq, &realtek_ictl_irq, handle_level_irq);
|
||
|
+
|
||
|
+ raw_spin_lock_irqsave(&irq_lock, flags);
|
||
|
+
|
||
|
+ output->child_mask |= BIT(hw);
|
||
|
+ write_irr(REG(RTL_ICTL_IRR0, 0), hw, output->output_index + 1);
|
||
|
+
|
||
|
+ raw_spin_unlock_irqrestore(&irq_lock, flags);
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
+static int intc_select(struct irq_domain *d, struct irq_fwspec *fwspec,
|
||
|
+ enum irq_domain_bus_token bus_token)
|
||
|
+{
|
||
|
+ struct realtek_ictl_output *output = d->host_data;
|
||
|
+ bool routed_elsewhere;
|
||
|
+ unsigned long flags;
|
||
|
+ u32 routing_old;
|
||
|
+ int cpu;
|
||
|
+
|
||
|
+ if (fwspec->fwnode != output->fwnode)
|
||
|
+ return false;
|
||
|
+
|
||
|
+ /* Original specifiers had only one parameter */
|
||
|
+ if (fwspec->param_count < 2)
|
||
|
+ return true;
|
||
|
+
|
||
|
+ raw_spin_lock_irqsave(&irq_lock, flags);
|
||
|
+
|
||
|
+ /*
|
||
|
+ * Inputs can only be routed to one output, so they shouldn't be
|
||
|
+ * allowed to end up in multiple domains.
|
||
|
+ */
|
||
|
+ for_each_cpu(cpu, &realtek_ictl_cpu_configurable) {
|
||
|
+ routing_old = read_irr(REG(RTL_ICTL_IRR0, cpu), fwspec->param[0]);
|
||
|
+ routed_elsewhere = routing_old && fwspec->param[1] != routing_old - 1;
|
||
|
+ if (routed_elsewhere) {
|
||
|
+ pr_warn("soc int %d already routed to output %d\n",
|
||
|
+ fwspec->param[0], routing_old - 1);
|
||
|
+ break;
|
||
|
+ }
|
||
|
+ }
|
||
|
+
|
||
|
+ raw_spin_unlock_irqrestore(&irq_lock, flags);
|
||
|
+
|
||
|
+ return !routed_elsewhere && fwspec->param[1] == output->output_index;
|
||
|
+}
|
||
|
+
|
||
|
static const struct irq_domain_ops irq_domain_ops = {
|
||
|
.map = intc_map,
|
||
|
+ .select = intc_select,
|
||
|
.xlate = irq_domain_xlate_onecell,
|
||
|
};
|
||
|
|
||
|
static void realtek_irq_dispatch(struct irq_desc *desc)
|
||
|
{
|
||
|
+ struct realtek_ictl_output *output = irq_desc_get_handler_data(desc);
|
||
|
struct irq_chip *chip = irq_desc_get_chip(desc);
|
||
|
- struct irq_domain *domain;
|
||
|
- unsigned int pending;
|
||
|
+ int cpu = smp_processor_id();
|
||
|
+ unsigned long pending;
|
||
|
+ unsigned int soc_int;
|
||
|
|
||
|
chained_irq_enter(chip, desc);
|
||
|
- pending = readl(REG(RTL_ICTL_GIMR)) & readl(REG(RTL_ICTL_GISR));
|
||
|
+ pending = readl(REG(RTL_ICTL_GIMR, cpu)) & readl(REG(RTL_ICTL_GISR, cpu))
|
||
|
+ & output->child_mask;
|
||
|
+
|
||
|
if (unlikely(!pending)) {
|
||
|
spurious_interrupt();
|
||
|
goto out;
|
||
|
}
|
||
|
- domain = irq_desc_get_handler_data(desc);
|
||
|
- generic_handle_irq(irq_find_mapping(domain, __ffs(pending)));
|
||
|
+
|
||
|
+ for_each_set_bit(soc_int, &pending, RTL_ICTL_NUM_INPUTS)
|
||
|
+ generic_handle_irq(irq_find_mapping(output->domain, soc_int));
|
||
|
+// generic_handle_domain_irq(output->domain, soc_int);
|
||
|
|
||
|
out:
|
||
|
chained_irq_exit(chip, desc);
|
||
|
}
|
||
|
|
||
|
-/*
|
||
|
- * SoC interrupts are cascaded to MIPS CPU interrupts according to the
|
||
|
- * interrupt-map in the device tree. Each SoC interrupt gets 4 bits for
|
||
|
- * the CPU interrupt in an Interrupt Routing Register. Max 32 SoC interrupts
|
||
|
- * thus go into 4 IRRs.
|
||
|
- */
|
||
|
-static int __init map_interrupts(struct device_node *node, struct irq_domain *domain)
|
||
|
+static int __init setup_parent_interrupts(struct device_node *node, int *parents,
|
||
|
+ unsigned int num_parents)
|
||
|
{
|
||
|
- struct device_node *cpu_ictl;
|
||
|
- const __be32 *imap;
|
||
|
- u32 imaplen, soc_int, cpu_int, tmp, regs[4];
|
||
|
- int ret, i, irr_regs[] = {
|
||
|
- RTL_ICTL_IRR3,
|
||
|
- RTL_ICTL_IRR2,
|
||
|
- RTL_ICTL_IRR1,
|
||
|
- RTL_ICTL_IRR0,
|
||
|
- };
|
||
|
- u8 mips_irqs_set;
|
||
|
+ struct realtek_ictl_output *outputs;
|
||
|
+ struct realtek_ictl_output *output;
|
||
|
+ struct irq_domain *domain;
|
||
|
+ unsigned int p;
|
||
|
|
||
|
- ret = of_property_read_u32(node, "#address-cells", &tmp);
|
||
|
- if (ret || tmp)
|
||
|
- return -EINVAL;
|
||
|
+ outputs = kcalloc(num_parents, sizeof(*outputs), GFP_KERNEL);
|
||
|
+ if (!outputs)
|
||
|
+ return -ENOMEM;
|
||
|
|
||
|
- imap = of_get_property(node, "interrupt-map", &imaplen);
|
||
|
- if (!imap || imaplen % 3)
|
||
|
- return -EINVAL;
|
||
|
+ for (p = 0; p < num_parents; p++) {
|
||
|
+ output = outputs + p;
|
||
|
|
||
|
- mips_irqs_set = 0;
|
||
|
- memset(regs, 0, sizeof(regs));
|
||
|
- for (i = 0; i < imaplen; i += 3 * sizeof(u32)) {
|
||
|
- soc_int = be32_to_cpup(imap);
|
||
|
- if (soc_int > 31)
|
||
|
- return -EINVAL;
|
||
|
-
|
||
|
- cpu_ictl = of_find_node_by_phandle(be32_to_cpup(imap + 1));
|
||
|
- if (!cpu_ictl)
|
||
|
- return -EINVAL;
|
||
|
- ret = of_property_read_u32(cpu_ictl, "#interrupt-cells", &tmp);
|
||
|
- if (ret || tmp != 1)
|
||
|
- return -EINVAL;
|
||
|
- of_node_put(cpu_ictl);
|
||
|
-
|
||
|
- cpu_int = be32_to_cpup(imap + 2);
|
||
|
- if (cpu_int > 7)
|
||
|
- return -EINVAL;
|
||
|
-
|
||
|
- if (!(mips_irqs_set & BIT(cpu_int))) {
|
||
|
- irq_set_chained_handler_and_data(cpu_int, realtek_irq_dispatch,
|
||
|
- domain);
|
||
|
- mips_irqs_set |= BIT(cpu_int);
|
||
|
- }
|
||
|
+ domain = irq_domain_add_linear(node, RTL_ICTL_NUM_INPUTS, &irq_domain_ops, output);
|
||
|
+ if (!domain)
|
||
|
+ goto domain_err;
|
||
|
|
||
|
- regs[(soc_int * 4) / 32] |= cpu_int << (soc_int * 4) % 32;
|
||
|
- imap += 3;
|
||
|
- }
|
||
|
+ output->fwnode = of_node_to_fwnode(node);
|
||
|
+ output->output_index = p;
|
||
|
+ output->domain = domain;
|
||
|
|
||
|
- for (i = 0; i < 4; i++)
|
||
|
- writel(regs[i], REG(irr_regs[i]));
|
||
|
+ irq_set_chained_handler_and_data(parents[p], realtek_irq_dispatch, output);
|
||
|
+ }
|
||
|
|
||
|
return 0;
|
||
|
+
|
||
|
+domain_err:
|
||
|
+ while (p--) {
|
||
|
+ irq_set_chained_handler_and_data(parents[p], NULL, NULL);
|
||
|
+ irq_domain_remove(outputs[p].domain);
|
||
|
+ }
|
||
|
+
|
||
|
+ kfree(outputs);
|
||
|
+
|
||
|
+ return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
static int __init realtek_rtl_of_init(struct device_node *node, struct device_node *parent)
|
||
|
{
|
||
|
- struct irq_domain *domain;
|
||
|
- int ret;
|
||
|
+ int parent_irqs[RTL_ICTL_NUM_OUTPUTS];
|
||
|
+ struct of_phandle_args oirq;
|
||
|
+ unsigned int num_parents;
|
||
|
+ unsigned int soc_irq;
|
||
|
+ unsigned int p;
|
||
|
+ int cpu;
|
||
|
+
|
||
|
+ cpumask_clear(&realtek_ictl_cpu_configurable);
|
||
|
+
|
||
|
+ for (cpu = 0; cpu < NR_CPUS; cpu++) {
|
||
|
+ realtek_ictl_base[cpu] = of_iomap(node, cpu);
|
||
|
+ if (realtek_ictl_base[cpu]) {
|
||
|
+ cpumask_set_cpu(cpu, &realtek_ictl_cpu_configurable);
|
||
|
+
|
||
|
+ /* Disable all cascaded interrupts and clear routing */
|
||
|
+ writel(0, REG(RTL_ICTL_GIMR, cpu));
|
||
|
+ for (soc_irq = 0; soc_irq < RTL_ICTL_NUM_INPUTS; soc_irq++)
|
||
|
+ write_irr(REG(RTL_ICTL_IRR0, cpu), soc_irq, 0);
|
||
|
+ }
|
||
|
+ }
|
||
|
|
||
|
- realtek_ictl_base = of_iomap(node, 0);
|
||
|
- if (!realtek_ictl_base)
|
||
|
+ if (cpumask_empty(&realtek_ictl_cpu_configurable))
|
||
|
return -ENXIO;
|
||
|
|
||
|
- /* Disable all cascaded interrupts */
|
||
|
- writel(0, REG(RTL_ICTL_GIMR));
|
||
|
+ num_parents = of_irq_count(node);
|
||
|
+ if (num_parents > RTL_ICTL_NUM_OUTPUTS) {
|
||
|
+ pr_err("too many parent interrupts\n");
|
||
|
+ return -EINVAL;
|
||
|
+ }
|
||
|
|
||
|
- domain = irq_domain_add_simple(node, 32, 0,
|
||
|
- &irq_domain_ops, NULL);
|
||
|
+ for (p = 0; p < num_parents; p++)
|
||
|
+ parent_irqs[p] = of_irq_get(node, p);
|
||
|
|
||
|
- ret = map_interrupts(node, domain);
|
||
|
- if (ret) {
|
||
|
- pr_err("invalid interrupt map\n");
|
||
|
- return ret;
|
||
|
+ if (WARN_ON(!num_parents)) {
|
||
|
+ /*
|
||
|
+ * If DT contains no parent interrupts, assume MIPS CPU IRQ 2
|
||
|
+ * (HW0) is connected to the first output. This is the case for
|
||
|
+ * all known hardware anyway. "interrupt-map" is deprecated, so
|
||
|
+ * don't bother trying to parse that.
|
||
|
+ * Since this is to account for old devicetrees with one-cell
|
||
|
+ * interrupt specifiers, only one output domain is needed.
|
||
|
+ */
|
||
|
+ oirq.np = of_find_compatible_node(NULL, NULL, "mti,cpu-interrupt-controller");
|
||
|
+ if (oirq.np) {
|
||
|
+ oirq.args_count = 1;
|
||
|
+ oirq.args[0] = 2;
|
||
|
+
|
||
|
+ parent_irqs[0] = irq_create_of_mapping(&oirq);
|
||
|
+ num_parents = 1;
|
||
|
+ }
|
||
|
+
|
||
|
+ of_node_put(oirq.np);
|
||
|
}
|
||
|
|
||
|
- return 0;
|
||
|
+ /* Ensure we haven't collected any errors before proceeding */
|
||
|
+ for (p = 0; p < num_parents; p++) {
|
||
|
+ if (parent_irqs[p] < 0)
|
||
|
+ return parent_irqs[p];
|
||
|
+ if (!parent_irqs[p])
|
||
|
+ return -ENODEV;
|
||
|
+ }
|
||
|
+
|
||
|
+ return setup_parent_interrupts(node, &parent_irqs[0], num_parents);
|
||
|
}
|
||
|
|
||
|
IRQCHIP_DECLARE(realtek_rtl_intc, "realtek,rtl-intc", realtek_rtl_of_init);
|