Merge branch 'for-chris-4.12' of git://git.kernel.org/pub/scm/linux/kernel/git/fdmana...
[sfrench/cifs-2.6.git] / arch / powerpc / platforms / powernv / opal-irqchip.c
1 /*
2  * This file implements an irqchip for OPAL events. Whenever there is
3  * an interrupt that is handled by OPAL we get passed a list of events
4  * that Linux needs to do something about. These basically look like
5  * interrupts to Linux so we implement an irqchip to handle them.
6  *
7  * Copyright Alistair Popple, IBM Corporation 2014.
8  *
9  * This program is free software; you can redistribute  it and/or modify it
10  * under  the terms of  the GNU General  Public License as published by the
11  * Free Software Foundation;  either version 2 of the  License, or (at your
12  * option) any later version.
13  */
14 #include <linux/bitops.h>
15 #include <linux/irq.h>
16 #include <linux/irqchip.h>
17 #include <linux/irqdomain.h>
18 #include <linux/interrupt.h>
19 #include <linux/module.h>
20 #include <linux/of.h>
21 #include <linux/platform_device.h>
22 #include <linux/kthread.h>
23 #include <linux/delay.h>
24 #include <linux/slab.h>
25 #include <linux/irq_work.h>
26
27 #include <asm/machdep.h>
28 #include <asm/opal.h>
29
30 #include "powernv.h"
31
32 /* Maximum number of events supported by OPAL firmware */
33 #define MAX_NUM_EVENTS 64
34
35 struct opal_event_irqchip {
36         struct irq_chip irqchip;
37         struct irq_domain *domain;
38         unsigned long mask;
39 };
40 static struct opal_event_irqchip opal_event_irqchip;
41
42 static unsigned int opal_irq_count;
43 static unsigned int *opal_irqs;
44
45 static void opal_handle_irq_work(struct irq_work *work);
46 static u64 last_outstanding_events;
47 static struct irq_work opal_event_irq_work = {
48         .func = opal_handle_irq_work,
49 };
50
51 void opal_handle_events(uint64_t events)
52 {
53         int virq, hwirq = 0;
54         u64 mask = opal_event_irqchip.mask;
55
56         if (!in_irq() && (events & mask)) {
57                 last_outstanding_events = events;
58                 irq_work_queue(&opal_event_irq_work);
59                 return;
60         }
61
62         while (events & mask) {
63                 hwirq = fls64(events) - 1;
64                 if (BIT_ULL(hwirq) & mask) {
65                         virq = irq_find_mapping(opal_event_irqchip.domain,
66                                                 hwirq);
67                         if (virq)
68                                 generic_handle_irq(virq);
69                 }
70                 events &= ~BIT_ULL(hwirq);
71         }
72 }
73
74 static void opal_event_mask(struct irq_data *d)
75 {
76         clear_bit(d->hwirq, &opal_event_irqchip.mask);
77 }
78
79 static void opal_event_unmask(struct irq_data *d)
80 {
81         __be64 events;
82
83         set_bit(d->hwirq, &opal_event_irqchip.mask);
84
85         opal_poll_events(&events);
86         last_outstanding_events = be64_to_cpu(events);
87
88         /*
89          * We can't just handle the events now with opal_handle_events().
90          * If we did we would deadlock when opal_event_unmask() is called from
91          * handle_level_irq() with the irq descriptor lock held, because
92          * calling opal_handle_events() would call generic_handle_irq() and
93          * then handle_level_irq() which would try to take the descriptor lock
94          * again. Instead queue the events for later.
95          */
96         if (last_outstanding_events & opal_event_irqchip.mask)
97                 /* Need to retrigger the interrupt */
98                 irq_work_queue(&opal_event_irq_work);
99 }
100
101 static int opal_event_set_type(struct irq_data *d, unsigned int flow_type)
102 {
103         /*
104          * For now we only support level triggered events. The irq
105          * handler will be called continuously until the event has
106          * been cleared in OPAL.
107          */
108         if (flow_type != IRQ_TYPE_LEVEL_HIGH)
109                 return -EINVAL;
110
111         return 0;
112 }
113
114 static struct opal_event_irqchip opal_event_irqchip = {
115         .irqchip = {
116                 .name = "OPAL EVT",
117                 .irq_mask = opal_event_mask,
118                 .irq_unmask = opal_event_unmask,
119                 .irq_set_type = opal_event_set_type,
120         },
121         .mask = 0,
122 };
123
124 static int opal_event_map(struct irq_domain *d, unsigned int irq,
125                         irq_hw_number_t hwirq)
126 {
127         irq_set_chip_data(irq, &opal_event_irqchip);
128         irq_set_chip_and_handler(irq, &opal_event_irqchip.irqchip,
129                                 handle_level_irq);
130
131         return 0;
132 }
133
134 static irqreturn_t opal_interrupt(int irq, void *data)
135 {
136         __be64 events;
137
138         opal_handle_interrupt(virq_to_hw(irq), &events);
139         opal_handle_events(be64_to_cpu(events));
140
141         return IRQ_HANDLED;
142 }
143
144 static void opal_handle_irq_work(struct irq_work *work)
145 {
146         opal_handle_events(last_outstanding_events);
147 }
148
149 static int opal_event_match(struct irq_domain *h, struct device_node *node,
150                             enum irq_domain_bus_token bus_token)
151 {
152         return irq_domain_get_of_node(h) == node;
153 }
154
155 static int opal_event_xlate(struct irq_domain *h, struct device_node *np,
156                            const u32 *intspec, unsigned int intsize,
157                            irq_hw_number_t *out_hwirq, unsigned int *out_flags)
158 {
159         *out_hwirq = intspec[0];
160         *out_flags = IRQ_TYPE_LEVEL_HIGH;
161
162         return 0;
163 }
164
165 static const struct irq_domain_ops opal_event_domain_ops = {
166         .match  = opal_event_match,
167         .map    = opal_event_map,
168         .xlate  = opal_event_xlate,
169 };
170
171 void opal_event_shutdown(void)
172 {
173         unsigned int i;
174
175         /* First free interrupts, which will also mask them */
176         for (i = 0; i < opal_irq_count; i++) {
177                 if (opal_irqs[i])
178                         free_irq(opal_irqs[i], NULL);
179                 opal_irqs[i] = 0;
180         }
181 }
182
183 int __init opal_event_init(void)
184 {
185         struct device_node *dn, *opal_node;
186         const char **names;
187         u32 *irqs;
188         int i, rc;
189
190         opal_node = of_find_node_by_path("/ibm,opal");
191         if (!opal_node) {
192                 pr_warn("opal: Node not found\n");
193                 return -ENODEV;
194         }
195
196         /* If dn is NULL it means the domain won't be linked to a DT
197          * node so therefore irq_of_parse_and_map(...) wont work. But
198          * that shouldn't be problem because if we're running a
199          * version of skiboot that doesn't have the dn then the
200          * devices won't have the correct properties and will have to
201          * fall back to the legacy method (opal_event_request(...))
202          * anyway. */
203         dn = of_find_compatible_node(NULL, NULL, "ibm,opal-event");
204         opal_event_irqchip.domain = irq_domain_add_linear(dn, MAX_NUM_EVENTS,
205                                 &opal_event_domain_ops, &opal_event_irqchip);
206         of_node_put(dn);
207         if (!opal_event_irqchip.domain) {
208                 pr_warn("opal: Unable to create irq domain\n");
209                 rc = -ENOMEM;
210                 goto out;
211         }
212
213         /* Get opal-interrupts property and names if present */
214         rc = of_property_count_u32_elems(opal_node, "opal-interrupts");
215         if (rc < 0)
216                 goto out;
217
218         opal_irq_count = rc;
219         pr_debug("Found %d interrupts reserved for OPAL\n", opal_irq_count);
220
221         irqs = kcalloc(opal_irq_count, sizeof(*irqs), GFP_KERNEL);
222         names = kcalloc(opal_irq_count, sizeof(*names), GFP_KERNEL);
223         opal_irqs = kcalloc(opal_irq_count, sizeof(*opal_irqs), GFP_KERNEL);
224
225         if (WARN_ON(!irqs || !names || !opal_irqs))
226                 goto out_free;
227
228         rc = of_property_read_u32_array(opal_node, "opal-interrupts",
229                                         irqs, opal_irq_count);
230         if (rc < 0) {
231                 pr_err("Error %d reading opal-interrupts array\n", rc);
232                 goto out_free;
233         }
234
235         /* It's not an error for the names to be missing */
236         of_property_read_string_array(opal_node, "opal-interrupts-names",
237                                       names, opal_irq_count);
238
239         /* Install interrupt handlers */
240         for (i = 0; i < opal_irq_count; i++) {
241                 unsigned int virq;
242                 char *name;
243
244                 /* Get hardware and virtual IRQ */
245                 virq = irq_create_mapping(NULL, irqs[i]);
246                 if (!virq) {
247                         pr_warn("Failed to map irq 0x%x\n", irqs[i]);
248                         continue;
249                 }
250
251                 if (names[i] && strlen(names[i]))
252                         name = kasprintf(GFP_KERNEL, "opal-%s", names[i]);
253                 else
254                         name = kasprintf(GFP_KERNEL, "opal");
255
256                 /* Install interrupt handler */
257                 rc = request_irq(virq, opal_interrupt, IRQF_TRIGGER_LOW,
258                                  name, NULL);
259                 if (rc) {
260                         irq_dispose_mapping(virq);
261                         pr_warn("Error %d requesting irq %d (0x%x)\n",
262                                  rc, virq, irqs[i]);
263                         continue;
264                 }
265
266                 /* Cache IRQ */
267                 opal_irqs[i] = virq;
268         }
269
270 out_free:
271         kfree(irqs);
272         kfree(names);
273 out:
274         of_node_put(opal_node);
275         return rc;
276 }
277 machine_arch_initcall(powernv, opal_event_init);
278
279 /**
280  * opal_event_request(unsigned int opal_event_nr) - Request an event
281  * @opal_event_nr: the opal event number to request
282  *
283  * This routine can be used to find the linux virq number which can
284  * then be passed to request_irq to assign a handler for a particular
285  * opal event. This should only be used by legacy devices which don't
286  * have proper device tree bindings. Most devices should use
287  * irq_of_parse_and_map() instead.
288  */
289 int opal_event_request(unsigned int opal_event_nr)
290 {
291         if (WARN_ON_ONCE(!opal_event_irqchip.domain))
292                 return 0;
293
294         return irq_create_mapping(opal_event_irqchip.domain, opal_event_nr);
295 }
296 EXPORT_SYMBOL(opal_event_request);