1 // SPDX-License-Identifier: GPL-2.0-only
3 * Freescale MU used as MSI controller
10 * Based on drivers/mailbox/imx-mailbox.c
13 #include <linux/clk.h>
14 #include <linux/irq.h>
15 #include <linux/irqchip.h>
16 #include <linux/irqchip/chained_irq.h>
17 #include <linux/irqdomain.h>
18 #include <linux/kernel.h>
19 #include <linux/module.h>
20 #include <linux/msi.h>
21 #include <linux/of_irq.h>
22 #include <linux/of_platform.h>
23 #include <linux/pm_runtime.h>
24 #include <linux/pm_domain.h>
25 #include <linux/spinlock.h>
27 #define IMX_MU_CHANS 4
49 /* Receive Interrupt Enable */
50 #define IMX_MU_xCR_RIEn(data, x) ((data->cfg->type) & IMX_MU_V2 ? BIT(x) : BIT(24 + (3 - (x))))
51 #define IMX_MU_xSR_RFn(data, x) ((data->cfg->type) & IMX_MU_V2 ? BIT(x) : BIT(24 + (3 - (x))))
54 enum imx_mu_type type;
55 u32 xTR; /* Transmit Register0 */
56 u32 xRR; /* Receive Register0 */
57 u32 xSR[IMX_MU_xSR_MAX]; /* Status Registers */
58 u32 xCR[IMX_MU_xCR_MAX]; /* Control Registers */
63 struct irq_domain *msi_domain;
65 phys_addr_t msiir_addr;
66 const struct imx_mu_dcfg *cfg;
71 static void imx_mu_write(struct imx_mu_msi *msi_data, u32 val, u32 offs)
73 iowrite32(val, msi_data->regs + offs);
76 static u32 imx_mu_read(struct imx_mu_msi *msi_data, u32 offs)
78 return ioread32(msi_data->regs + offs);
81 static u32 imx_mu_xcr_rmw(struct imx_mu_msi *msi_data, enum imx_mu_xcr type, u32 set, u32 clr)
86 raw_spin_lock_irqsave(&msi_data->lock, flags);
87 val = imx_mu_read(msi_data, msi_data->cfg->xCR[type]);
90 imx_mu_write(msi_data, val, msi_data->cfg->xCR[type]);
91 raw_spin_unlock_irqrestore(&msi_data->lock, flags);
96 static void imx_mu_msi_parent_mask_irq(struct irq_data *data)
98 struct imx_mu_msi *msi_data = irq_data_get_irq_chip_data(data);
100 imx_mu_xcr_rmw(msi_data, IMX_MU_RCR, 0, IMX_MU_xCR_RIEn(msi_data, data->hwirq));
103 static void imx_mu_msi_parent_unmask_irq(struct irq_data *data)
105 struct imx_mu_msi *msi_data = irq_data_get_irq_chip_data(data);
107 imx_mu_xcr_rmw(msi_data, IMX_MU_RCR, IMX_MU_xCR_RIEn(msi_data, data->hwirq), 0);
110 static void imx_mu_msi_parent_ack_irq(struct irq_data *data)
112 struct imx_mu_msi *msi_data = irq_data_get_irq_chip_data(data);
114 imx_mu_read(msi_data, msi_data->cfg->xRR + data->hwirq * 4);
117 static struct irq_chip imx_mu_msi_irq_chip = {
119 .irq_ack = irq_chip_ack_parent,
122 static struct msi_domain_ops imx_mu_msi_irq_ops = {
125 static struct msi_domain_info imx_mu_msi_domain_info = {
126 .flags = (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS),
127 .ops = &imx_mu_msi_irq_ops,
128 .chip = &imx_mu_msi_irq_chip,
131 static void imx_mu_msi_parent_compose_msg(struct irq_data *data,
134 struct imx_mu_msi *msi_data = irq_data_get_irq_chip_data(data);
135 u64 addr = msi_data->msiir_addr + 4 * data->hwirq;
137 msg->address_hi = upper_32_bits(addr);
138 msg->address_lo = lower_32_bits(addr);
139 msg->data = data->hwirq;
142 static int imx_mu_msi_parent_set_affinity(struct irq_data *irq_data,
143 const struct cpumask *mask, bool force)
148 static struct irq_chip imx_mu_msi_parent_chip = {
150 .irq_mask = imx_mu_msi_parent_mask_irq,
151 .irq_unmask = imx_mu_msi_parent_unmask_irq,
152 .irq_ack = imx_mu_msi_parent_ack_irq,
153 .irq_compose_msi_msg = imx_mu_msi_parent_compose_msg,
154 .irq_set_affinity = imx_mu_msi_parent_set_affinity,
157 static int imx_mu_msi_domain_irq_alloc(struct irq_domain *domain,
159 unsigned int nr_irqs,
162 struct imx_mu_msi *msi_data = domain->host_data;
166 WARN_ON(nr_irqs != 1);
168 raw_spin_lock_irqsave(&msi_data->lock, flags);
169 pos = find_first_zero_bit(&msi_data->used, IMX_MU_CHANS);
170 if (pos < IMX_MU_CHANS)
171 __set_bit(pos, &msi_data->used);
174 raw_spin_unlock_irqrestore(&msi_data->lock, flags);
179 irq_domain_set_info(domain, virq, pos,
180 &imx_mu_msi_parent_chip, msi_data,
181 handle_edge_irq, NULL, NULL);
185 static void imx_mu_msi_domain_irq_free(struct irq_domain *domain,
186 unsigned int virq, unsigned int nr_irqs)
188 struct irq_data *d = irq_domain_get_irq_data(domain, virq);
189 struct imx_mu_msi *msi_data = irq_data_get_irq_chip_data(d);
192 raw_spin_lock_irqsave(&msi_data->lock, flags);
193 __clear_bit(d->hwirq, &msi_data->used);
194 raw_spin_unlock_irqrestore(&msi_data->lock, flags);
197 static const struct irq_domain_ops imx_mu_msi_domain_ops = {
198 .alloc = imx_mu_msi_domain_irq_alloc,
199 .free = imx_mu_msi_domain_irq_free,
202 static void imx_mu_msi_irq_handler(struct irq_desc *desc)
204 struct imx_mu_msi *msi_data = irq_desc_get_handler_data(desc);
205 struct irq_chip *chip = irq_desc_get_chip(desc);
209 status = imx_mu_read(msi_data, msi_data->cfg->xSR[IMX_MU_RSR]);
211 chained_irq_enter(chip, desc);
212 for (i = 0; i < IMX_MU_CHANS; i++) {
213 if (status & IMX_MU_xSR_RFn(msi_data, i))
214 generic_handle_domain_irq(msi_data->msi_domain, i);
216 chained_irq_exit(chip, desc);
219 static int imx_mu_msi_domains_init(struct imx_mu_msi *msi_data, struct device *dev)
221 struct fwnode_handle *fwnodes = dev_fwnode(dev);
222 struct irq_domain *parent;
224 /* Initialize MSI domain parent */
225 parent = irq_domain_create_linear(fwnodes,
227 &imx_mu_msi_domain_ops,
230 dev_err(dev, "failed to create IRQ domain\n");
234 irq_domain_update_bus_token(parent, DOMAIN_BUS_NEXUS);
236 msi_data->msi_domain = platform_msi_create_irq_domain(fwnodes,
237 &imx_mu_msi_domain_info,
240 if (!msi_data->msi_domain) {
241 dev_err(dev, "failed to create MSI domain\n");
242 irq_domain_remove(parent);
246 irq_domain_set_pm_device(msi_data->msi_domain, dev);
251 /* Register offset of different version MU IP */
252 static const struct imx_mu_dcfg imx_mu_cfg_imx6sx = {
263 [IMX_MU_GIER] = 0x24,
270 static const struct imx_mu_dcfg imx_mu_cfg_imx7ulp = {
281 [IMX_MU_GIER] = 0x64,
288 static const struct imx_mu_dcfg imx_mu_cfg_imx8ulp = {
294 [IMX_MU_GSR] = 0x118,
295 [IMX_MU_TSR] = 0x124,
296 [IMX_MU_RSR] = 0x12C,
299 [IMX_MU_GIER] = 0x110,
300 [IMX_MU_GCR] = 0x114,
301 [IMX_MU_TCR] = 0x120,
306 static int __init imx_mu_of_init(struct device_node *dn,
307 struct device_node *parent,
308 const struct imx_mu_dcfg *cfg)
310 struct platform_device *pdev = of_find_device_by_node(dn);
311 struct device_link *pd_link_a;
312 struct device_link *pd_link_b;
313 struct imx_mu_msi *msi_data;
314 struct resource *res;
323 msi_data = devm_kzalloc(&pdev->dev, sizeof(*msi_data), GFP_KERNEL);
329 msi_data->regs = devm_platform_ioremap_resource_byname(pdev, "processor-a-side");
330 if (IS_ERR(msi_data->regs)) {
331 dev_err(&pdev->dev, "failed to initialize 'regs'\n");
332 return PTR_ERR(msi_data->regs);
335 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "processor-b-side");
339 msi_data->msiir_addr = res->start + msi_data->cfg->xTR;
341 irq = platform_get_irq(pdev, 0);
345 platform_set_drvdata(pdev, msi_data);
347 msi_data->clk = devm_clk_get(dev, NULL);
348 if (IS_ERR(msi_data->clk))
349 return PTR_ERR(msi_data->clk);
351 pd_a = dev_pm_domain_attach_by_name(dev, "processor-a-side");
353 return PTR_ERR(pd_a);
355 pd_b = dev_pm_domain_attach_by_name(dev, "processor-b-side");
357 return PTR_ERR(pd_b);
359 pd_link_a = device_link_add(dev, pd_a,
365 dev_err(dev, "Failed to add device_link to mu a.\n");
369 pd_link_b = device_link_add(dev, pd_b,
376 dev_err(dev, "Failed to add device_link to mu a.\n");
380 ret = imx_mu_msi_domains_init(msi_data, dev);
384 pm_runtime_enable(dev);
386 irq_set_chained_handler_and_data(irq,
387 imx_mu_msi_irq_handler,
393 device_link_remove(dev, pd_b);
395 device_link_remove(dev, pd_a);
400 static int __maybe_unused imx_mu_runtime_suspend(struct device *dev)
402 struct imx_mu_msi *priv = dev_get_drvdata(dev);
404 clk_disable_unprepare(priv->clk);
409 static int __maybe_unused imx_mu_runtime_resume(struct device *dev)
411 struct imx_mu_msi *priv = dev_get_drvdata(dev);
414 ret = clk_prepare_enable(priv->clk);
416 dev_err(dev, "failed to enable clock\n");
421 static const struct dev_pm_ops imx_mu_pm_ops = {
422 SET_RUNTIME_PM_OPS(imx_mu_runtime_suspend,
423 imx_mu_runtime_resume, NULL)
426 static int __init imx_mu_imx7ulp_of_init(struct device_node *dn,
427 struct device_node *parent)
429 return imx_mu_of_init(dn, parent, &imx_mu_cfg_imx7ulp);
432 static int __init imx_mu_imx6sx_of_init(struct device_node *dn,
433 struct device_node *parent)
435 return imx_mu_of_init(dn, parent, &imx_mu_cfg_imx6sx);
438 static int __init imx_mu_imx8ulp_of_init(struct device_node *dn,
439 struct device_node *parent)
441 return imx_mu_of_init(dn, parent, &imx_mu_cfg_imx8ulp);
444 IRQCHIP_PLATFORM_DRIVER_BEGIN(imx_mu_msi)
445 IRQCHIP_MATCH("fsl,imx7ulp-mu-msi", imx_mu_imx7ulp_of_init)
446 IRQCHIP_MATCH("fsl,imx6sx-mu-msi", imx_mu_imx6sx_of_init)
447 IRQCHIP_MATCH("fsl,imx8ulp-mu-msi", imx_mu_imx8ulp_of_init)
448 IRQCHIP_PLATFORM_DRIVER_END(imx_mu_msi, .pm = &imx_mu_pm_ops)
452 MODULE_DESCRIPTION("Freescale MU MSI controller driver");
453 MODULE_LICENSE("GPL");