Loading...
1// SPDX-License-Identifier: GPL-2.0
2// Copyright (c) 2014-2018 MediaTek Inc.
3
4/*
5 * Library for MediaTek External Interrupt Support
6 *
7 * Author: Maoguang Meng <maoguang.meng@mediatek.com>
8 * Sean Wang <sean.wang@mediatek.com>
9 *
10 */
11
12#include <linux/delay.h>
13#include <linux/err.h>
14#include <linux/gpio/driver.h>
15#include <linux/io.h>
16#include <linux/irqchip/chained_irq.h>
17#include <linux/irqdomain.h>
18#include <linux/module.h>
19#include <linux/of_irq.h>
20#include <linux/platform_device.h>
21
22#include "mtk-eint.h"
23
24#define MTK_EINT_EDGE_SENSITIVE 0
25#define MTK_EINT_LEVEL_SENSITIVE 1
26#define MTK_EINT_DBNC_SET_DBNC_BITS 4
27#define MTK_EINT_DBNC_MAX 16
28#define MTK_EINT_DBNC_RST_BIT (0x1 << 1)
29#define MTK_EINT_DBNC_SET_EN (0x1 << 0)
30
31static const struct mtk_eint_regs mtk_generic_eint_regs = {
32 .stat = 0x000,
33 .ack = 0x040,
34 .mask = 0x080,
35 .mask_set = 0x0c0,
36 .mask_clr = 0x100,
37 .sens = 0x140,
38 .sens_set = 0x180,
39 .sens_clr = 0x1c0,
40 .soft = 0x200,
41 .soft_set = 0x240,
42 .soft_clr = 0x280,
43 .pol = 0x300,
44 .pol_set = 0x340,
45 .pol_clr = 0x380,
46 .dom_en = 0x400,
47 .dbnc_ctrl = 0x500,
48 .dbnc_set = 0x600,
49 .dbnc_clr = 0x700,
50};
51
52const unsigned int debounce_time_mt2701[] = {
53 500, 1000, 16000, 32000, 64000, 128000, 256000, 0
54};
55EXPORT_SYMBOL_GPL(debounce_time_mt2701);
56
57const unsigned int debounce_time_mt6765[] = {
58 125, 250, 500, 1000, 16000, 32000, 64000, 128000, 256000, 512000, 0
59};
60EXPORT_SYMBOL_GPL(debounce_time_mt6765);
61
62const unsigned int debounce_time_mt6795[] = {
63 500, 1000, 16000, 32000, 64000, 128000, 256000, 512000, 0
64};
65EXPORT_SYMBOL_GPL(debounce_time_mt6795);
66
67static void __iomem *mtk_eint_get_offset(struct mtk_eint *eint,
68 unsigned int eint_num,
69 unsigned int offset)
70{
71 unsigned int eint_base = 0;
72 void __iomem *reg;
73
74 if (eint_num >= eint->hw->ap_num)
75 eint_base = eint->hw->ap_num;
76
77 reg = eint->base + offset + ((eint_num - eint_base) / 32) * 4;
78
79 return reg;
80}
81
82static unsigned int mtk_eint_can_en_debounce(struct mtk_eint *eint,
83 unsigned int eint_num)
84{
85 unsigned int sens;
86 unsigned int bit = BIT(eint_num % 32);
87 void __iomem *reg = mtk_eint_get_offset(eint, eint_num,
88 eint->regs->sens);
89
90 if (readl(reg) & bit)
91 sens = MTK_EINT_LEVEL_SENSITIVE;
92 else
93 sens = MTK_EINT_EDGE_SENSITIVE;
94
95 if (eint_num < eint->hw->db_cnt && sens != MTK_EINT_EDGE_SENSITIVE)
96 return 1;
97 else
98 return 0;
99}
100
101static int mtk_eint_flip_edge(struct mtk_eint *eint, int hwirq)
102{
103 int start_level, curr_level;
104 unsigned int reg_offset;
105 u32 mask = BIT(hwirq & 0x1f);
106 u32 port = (hwirq >> 5) & eint->hw->port_mask;
107 void __iomem *reg = eint->base + (port << 2);
108
109 curr_level = eint->gpio_xlate->get_gpio_state(eint->pctl, hwirq);
110
111 do {
112 start_level = curr_level;
113 if (start_level)
114 reg_offset = eint->regs->pol_clr;
115 else
116 reg_offset = eint->regs->pol_set;
117 writel(mask, reg + reg_offset);
118
119 curr_level = eint->gpio_xlate->get_gpio_state(eint->pctl,
120 hwirq);
121 } while (start_level != curr_level);
122
123 return start_level;
124}
125
126static void mtk_eint_mask(struct irq_data *d)
127{
128 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
129 u32 mask = BIT(d->hwirq & 0x1f);
130 void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
131 eint->regs->mask_set);
132
133 eint->cur_mask[d->hwirq >> 5] &= ~mask;
134
135 writel(mask, reg);
136}
137
138static void mtk_eint_unmask(struct irq_data *d)
139{
140 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
141 u32 mask = BIT(d->hwirq & 0x1f);
142 void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
143 eint->regs->mask_clr);
144
145 eint->cur_mask[d->hwirq >> 5] |= mask;
146
147 writel(mask, reg);
148
149 if (eint->dual_edge[d->hwirq])
150 mtk_eint_flip_edge(eint, d->hwirq);
151}
152
153static unsigned int mtk_eint_get_mask(struct mtk_eint *eint,
154 unsigned int eint_num)
155{
156 unsigned int bit = BIT(eint_num % 32);
157 void __iomem *reg = mtk_eint_get_offset(eint, eint_num,
158 eint->regs->mask);
159
160 return !!(readl(reg) & bit);
161}
162
163static void mtk_eint_ack(struct irq_data *d)
164{
165 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
166 u32 mask = BIT(d->hwirq & 0x1f);
167 void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
168 eint->regs->ack);
169
170 writel(mask, reg);
171}
172
173static int mtk_eint_set_type(struct irq_data *d, unsigned int type)
174{
175 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
176 bool masked;
177 u32 mask = BIT(d->hwirq & 0x1f);
178 void __iomem *reg;
179
180 if (((type & IRQ_TYPE_EDGE_BOTH) && (type & IRQ_TYPE_LEVEL_MASK)) ||
181 ((type & IRQ_TYPE_LEVEL_MASK) == IRQ_TYPE_LEVEL_MASK)) {
182 dev_err(eint->dev,
183 "Can't configure IRQ%d (EINT%lu) for type 0x%X\n",
184 d->irq, d->hwirq, type);
185 return -EINVAL;
186 }
187
188 if ((type & IRQ_TYPE_EDGE_BOTH) == IRQ_TYPE_EDGE_BOTH)
189 eint->dual_edge[d->hwirq] = 1;
190 else
191 eint->dual_edge[d->hwirq] = 0;
192
193 if (!mtk_eint_get_mask(eint, d->hwirq)) {
194 mtk_eint_mask(d);
195 masked = false;
196 } else {
197 masked = true;
198 }
199
200 if (type & (IRQ_TYPE_LEVEL_LOW | IRQ_TYPE_EDGE_FALLING)) {
201 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->pol_clr);
202 writel(mask, reg);
203 } else {
204 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->pol_set);
205 writel(mask, reg);
206 }
207
208 if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
209 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->sens_clr);
210 writel(mask, reg);
211 } else {
212 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->sens_set);
213 writel(mask, reg);
214 }
215
216 mtk_eint_ack(d);
217 if (!masked)
218 mtk_eint_unmask(d);
219
220 return 0;
221}
222
223static int mtk_eint_irq_set_wake(struct irq_data *d, unsigned int on)
224{
225 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
226 int shift = d->hwirq & 0x1f;
227 int reg = d->hwirq >> 5;
228
229 if (on)
230 eint->wake_mask[reg] |= BIT(shift);
231 else
232 eint->wake_mask[reg] &= ~BIT(shift);
233
234 return 0;
235}
236
237static void mtk_eint_chip_write_mask(const struct mtk_eint *eint,
238 void __iomem *base, u32 *buf)
239{
240 int port;
241 void __iomem *reg;
242
243 for (port = 0; port < eint->hw->ports; port++) {
244 reg = base + (port << 2);
245 writel_relaxed(~buf[port], reg + eint->regs->mask_set);
246 writel_relaxed(buf[port], reg + eint->regs->mask_clr);
247 }
248}
249
250static int mtk_eint_irq_request_resources(struct irq_data *d)
251{
252 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
253 struct gpio_chip *gpio_c;
254 unsigned int gpio_n;
255 int err;
256
257 err = eint->gpio_xlate->get_gpio_n(eint->pctl, d->hwirq,
258 &gpio_n, &gpio_c);
259 if (err < 0) {
260 dev_err(eint->dev, "Can not find pin\n");
261 return err;
262 }
263
264 err = gpiochip_lock_as_irq(gpio_c, gpio_n);
265 if (err < 0) {
266 dev_err(eint->dev, "unable to lock HW IRQ %lu for IRQ\n",
267 irqd_to_hwirq(d));
268 return err;
269 }
270
271 err = eint->gpio_xlate->set_gpio_as_eint(eint->pctl, d->hwirq);
272 if (err < 0) {
273 dev_err(eint->dev, "Can not eint mode\n");
274 return err;
275 }
276
277 return 0;
278}
279
280static void mtk_eint_irq_release_resources(struct irq_data *d)
281{
282 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
283 struct gpio_chip *gpio_c;
284 unsigned int gpio_n;
285
286 eint->gpio_xlate->get_gpio_n(eint->pctl, d->hwirq, &gpio_n,
287 &gpio_c);
288
289 gpiochip_unlock_as_irq(gpio_c, gpio_n);
290}
291
292static struct irq_chip mtk_eint_irq_chip = {
293 .name = "mt-eint",
294 .irq_disable = mtk_eint_mask,
295 .irq_mask = mtk_eint_mask,
296 .irq_unmask = mtk_eint_unmask,
297 .irq_ack = mtk_eint_ack,
298 .irq_set_type = mtk_eint_set_type,
299 .irq_set_wake = mtk_eint_irq_set_wake,
300 .irq_request_resources = mtk_eint_irq_request_resources,
301 .irq_release_resources = mtk_eint_irq_release_resources,
302};
303
304static unsigned int mtk_eint_hw_init(struct mtk_eint *eint)
305{
306 void __iomem *dom_en = eint->base + eint->regs->dom_en;
307 void __iomem *mask_set = eint->base + eint->regs->mask_set;
308 unsigned int i;
309
310 for (i = 0; i < eint->hw->ap_num; i += 32) {
311 writel(0xffffffff, dom_en);
312 writel(0xffffffff, mask_set);
313 dom_en += 4;
314 mask_set += 4;
315 }
316
317 return 0;
318}
319
320static inline void
321mtk_eint_debounce_process(struct mtk_eint *eint, int index)
322{
323 unsigned int rst, ctrl_offset;
324 unsigned int bit, dbnc;
325
326 ctrl_offset = (index / 4) * 4 + eint->regs->dbnc_ctrl;
327 dbnc = readl(eint->base + ctrl_offset);
328 bit = MTK_EINT_DBNC_SET_EN << ((index % 4) * 8);
329 if ((bit & dbnc) > 0) {
330 ctrl_offset = (index / 4) * 4 + eint->regs->dbnc_set;
331 rst = MTK_EINT_DBNC_RST_BIT << ((index % 4) * 8);
332 writel(rst, eint->base + ctrl_offset);
333 }
334}
335
336static void mtk_eint_irq_handler(struct irq_desc *desc)
337{
338 struct irq_chip *chip = irq_desc_get_chip(desc);
339 struct mtk_eint *eint = irq_desc_get_handler_data(desc);
340 unsigned int status, eint_num;
341 int offset, mask_offset, index;
342 void __iomem *reg = mtk_eint_get_offset(eint, 0, eint->regs->stat);
343 int dual_edge, start_level, curr_level;
344
345 chained_irq_enter(chip, desc);
346 for (eint_num = 0; eint_num < eint->hw->ap_num; eint_num += 32,
347 reg += 4) {
348 status = readl(reg);
349 while (status) {
350 offset = __ffs(status);
351 mask_offset = eint_num >> 5;
352 index = eint_num + offset;
353 status &= ~BIT(offset);
354
355 /*
356 * If we get an interrupt on pin that was only required
357 * for wake (but no real interrupt requested), mask the
358 * interrupt (as would mtk_eint_resume do anyway later
359 * in the resume sequence).
360 */
361 if (eint->wake_mask[mask_offset] & BIT(offset) &&
362 !(eint->cur_mask[mask_offset] & BIT(offset))) {
363 writel_relaxed(BIT(offset), reg -
364 eint->regs->stat +
365 eint->regs->mask_set);
366 }
367
368 dual_edge = eint->dual_edge[index];
369 if (dual_edge) {
370 /*
371 * Clear soft-irq in case we raised it last
372 * time.
373 */
374 writel(BIT(offset), reg - eint->regs->stat +
375 eint->regs->soft_clr);
376
377 start_level =
378 eint->gpio_xlate->get_gpio_state(eint->pctl,
379 index);
380 }
381
382 generic_handle_domain_irq(eint->domain, index);
383
384 if (dual_edge) {
385 curr_level = mtk_eint_flip_edge(eint, index);
386
387 /*
388 * If level changed, we might lost one edge
389 * interrupt, raised it through soft-irq.
390 */
391 if (start_level != curr_level)
392 writel(BIT(offset), reg -
393 eint->regs->stat +
394 eint->regs->soft_set);
395 }
396
397 if (index < eint->hw->db_cnt)
398 mtk_eint_debounce_process(eint, index);
399 }
400 }
401 chained_irq_exit(chip, desc);
402}
403
404int mtk_eint_do_suspend(struct mtk_eint *eint)
405{
406 mtk_eint_chip_write_mask(eint, eint->base, eint->wake_mask);
407
408 return 0;
409}
410EXPORT_SYMBOL_GPL(mtk_eint_do_suspend);
411
412int mtk_eint_do_resume(struct mtk_eint *eint)
413{
414 mtk_eint_chip_write_mask(eint, eint->base, eint->cur_mask);
415
416 return 0;
417}
418EXPORT_SYMBOL_GPL(mtk_eint_do_resume);
419
420int mtk_eint_set_debounce(struct mtk_eint *eint, unsigned long eint_num,
421 unsigned int debounce)
422{
423 int virq, eint_offset;
424 unsigned int set_offset, bit, clr_bit, clr_offset, rst, i, unmask,
425 dbnc;
426 struct irq_data *d;
427
428 if (!eint->hw->db_time)
429 return -EOPNOTSUPP;
430
431 virq = irq_find_mapping(eint->domain, eint_num);
432 eint_offset = (eint_num % 4) * 8;
433 d = irq_get_irq_data(virq);
434
435 set_offset = (eint_num / 4) * 4 + eint->regs->dbnc_set;
436 clr_offset = (eint_num / 4) * 4 + eint->regs->dbnc_clr;
437
438 if (!mtk_eint_can_en_debounce(eint, eint_num))
439 return -EINVAL;
440
441 dbnc = eint->num_db_time;
442 for (i = 0; i < eint->num_db_time; i++) {
443 if (debounce <= eint->hw->db_time[i]) {
444 dbnc = i;
445 break;
446 }
447 }
448
449 if (!mtk_eint_get_mask(eint, eint_num)) {
450 mtk_eint_mask(d);
451 unmask = 1;
452 } else {
453 unmask = 0;
454 }
455
456 clr_bit = 0xff << eint_offset;
457 writel(clr_bit, eint->base + clr_offset);
458
459 bit = ((dbnc << MTK_EINT_DBNC_SET_DBNC_BITS) | MTK_EINT_DBNC_SET_EN) <<
460 eint_offset;
461 rst = MTK_EINT_DBNC_RST_BIT << eint_offset;
462 writel(rst | bit, eint->base + set_offset);
463
464 /*
465 * Delay a while (more than 2T) to wait for hw debounce counter reset
466 * work correctly.
467 */
468 udelay(1);
469 if (unmask == 1)
470 mtk_eint_unmask(d);
471
472 return 0;
473}
474EXPORT_SYMBOL_GPL(mtk_eint_set_debounce);
475
476int mtk_eint_find_irq(struct mtk_eint *eint, unsigned long eint_n)
477{
478 int irq;
479
480 irq = irq_find_mapping(eint->domain, eint_n);
481 if (!irq)
482 return -EINVAL;
483
484 return irq;
485}
486EXPORT_SYMBOL_GPL(mtk_eint_find_irq);
487
488int mtk_eint_do_init(struct mtk_eint *eint)
489{
490 int i;
491
492 /* If clients don't assign a specific regs, let's use generic one */
493 if (!eint->regs)
494 eint->regs = &mtk_generic_eint_regs;
495
496 eint->wake_mask = devm_kcalloc(eint->dev, eint->hw->ports,
497 sizeof(*eint->wake_mask), GFP_KERNEL);
498 if (!eint->wake_mask)
499 return -ENOMEM;
500
501 eint->cur_mask = devm_kcalloc(eint->dev, eint->hw->ports,
502 sizeof(*eint->cur_mask), GFP_KERNEL);
503 if (!eint->cur_mask)
504 return -ENOMEM;
505
506 eint->dual_edge = devm_kcalloc(eint->dev, eint->hw->ap_num,
507 sizeof(int), GFP_KERNEL);
508 if (!eint->dual_edge)
509 return -ENOMEM;
510
511 eint->domain = irq_domain_add_linear(eint->dev->of_node,
512 eint->hw->ap_num,
513 &irq_domain_simple_ops, NULL);
514 if (!eint->domain)
515 return -ENOMEM;
516
517 if (eint->hw->db_time) {
518 for (i = 0; i < MTK_EINT_DBNC_MAX; i++)
519 if (eint->hw->db_time[i] == 0)
520 break;
521 eint->num_db_time = i;
522 }
523
524 mtk_eint_hw_init(eint);
525 for (i = 0; i < eint->hw->ap_num; i++) {
526 int virq = irq_create_mapping(eint->domain, i);
527
528 irq_set_chip_and_handler(virq, &mtk_eint_irq_chip,
529 handle_level_irq);
530 irq_set_chip_data(virq, eint);
531 }
532
533 irq_set_chained_handler_and_data(eint->irq, mtk_eint_irq_handler,
534 eint);
535
536 return 0;
537}
538EXPORT_SYMBOL_GPL(mtk_eint_do_init);
539
540MODULE_LICENSE("GPL v2");
541MODULE_DESCRIPTION("MediaTek EINT Driver");
1// SPDX-License-Identifier: GPL-2.0
2// Copyright (c) 2014-2018 MediaTek Inc.
3
4/*
5 * Library for MediaTek External Interrupt Support
6 *
7 * Author: Maoguang Meng <maoguang.meng@mediatek.com>
8 * Sean Wang <sean.wang@mediatek.com>
9 *
10 */
11
12#include <linux/delay.h>
13#include <linux/err.h>
14#include <linux/gpio/driver.h>
15#include <linux/io.h>
16#include <linux/irqchip/chained_irq.h>
17#include <linux/irqdomain.h>
18#include <linux/module.h>
19#include <linux/of_irq.h>
20#include <linux/platform_device.h>
21
22#include "mtk-eint.h"
23
24#define MTK_EINT_EDGE_SENSITIVE 0
25#define MTK_EINT_LEVEL_SENSITIVE 1
26#define MTK_EINT_DBNC_SET_DBNC_BITS 4
27#define MTK_EINT_DBNC_RST_BIT (0x1 << 1)
28#define MTK_EINT_DBNC_SET_EN (0x1 << 0)
29
30static const struct mtk_eint_regs mtk_generic_eint_regs = {
31 .stat = 0x000,
32 .ack = 0x040,
33 .mask = 0x080,
34 .mask_set = 0x0c0,
35 .mask_clr = 0x100,
36 .sens = 0x140,
37 .sens_set = 0x180,
38 .sens_clr = 0x1c0,
39 .soft = 0x200,
40 .soft_set = 0x240,
41 .soft_clr = 0x280,
42 .pol = 0x300,
43 .pol_set = 0x340,
44 .pol_clr = 0x380,
45 .dom_en = 0x400,
46 .dbnc_ctrl = 0x500,
47 .dbnc_set = 0x600,
48 .dbnc_clr = 0x700,
49};
50
51static void __iomem *mtk_eint_get_offset(struct mtk_eint *eint,
52 unsigned int eint_num,
53 unsigned int offset)
54{
55 unsigned int eint_base = 0;
56 void __iomem *reg;
57
58 if (eint_num >= eint->hw->ap_num)
59 eint_base = eint->hw->ap_num;
60
61 reg = eint->base + offset + ((eint_num - eint_base) / 32) * 4;
62
63 return reg;
64}
65
66static unsigned int mtk_eint_can_en_debounce(struct mtk_eint *eint,
67 unsigned int eint_num)
68{
69 unsigned int sens;
70 unsigned int bit = BIT(eint_num % 32);
71 void __iomem *reg = mtk_eint_get_offset(eint, eint_num,
72 eint->regs->sens);
73
74 if (readl(reg) & bit)
75 sens = MTK_EINT_LEVEL_SENSITIVE;
76 else
77 sens = MTK_EINT_EDGE_SENSITIVE;
78
79 if (eint_num < eint->hw->db_cnt && sens != MTK_EINT_EDGE_SENSITIVE)
80 return 1;
81 else
82 return 0;
83}
84
85static int mtk_eint_flip_edge(struct mtk_eint *eint, int hwirq)
86{
87 int start_level, curr_level;
88 unsigned int reg_offset;
89 u32 mask = BIT(hwirq & 0x1f);
90 u32 port = (hwirq >> 5) & eint->hw->port_mask;
91 void __iomem *reg = eint->base + (port << 2);
92
93 curr_level = eint->gpio_xlate->get_gpio_state(eint->pctl, hwirq);
94
95 do {
96 start_level = curr_level;
97 if (start_level)
98 reg_offset = eint->regs->pol_clr;
99 else
100 reg_offset = eint->regs->pol_set;
101 writel(mask, reg + reg_offset);
102
103 curr_level = eint->gpio_xlate->get_gpio_state(eint->pctl,
104 hwirq);
105 } while (start_level != curr_level);
106
107 return start_level;
108}
109
110static void mtk_eint_mask(struct irq_data *d)
111{
112 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
113 u32 mask = BIT(d->hwirq & 0x1f);
114 void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
115 eint->regs->mask_set);
116
117 eint->cur_mask[d->hwirq >> 5] &= ~mask;
118
119 writel(mask, reg);
120}
121
122static void mtk_eint_unmask(struct irq_data *d)
123{
124 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
125 u32 mask = BIT(d->hwirq & 0x1f);
126 void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
127 eint->regs->mask_clr);
128
129 eint->cur_mask[d->hwirq >> 5] |= mask;
130
131 writel(mask, reg);
132
133 if (eint->dual_edge[d->hwirq])
134 mtk_eint_flip_edge(eint, d->hwirq);
135}
136
137static unsigned int mtk_eint_get_mask(struct mtk_eint *eint,
138 unsigned int eint_num)
139{
140 unsigned int bit = BIT(eint_num % 32);
141 void __iomem *reg = mtk_eint_get_offset(eint, eint_num,
142 eint->regs->mask);
143
144 return !!(readl(reg) & bit);
145}
146
147static void mtk_eint_ack(struct irq_data *d)
148{
149 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
150 u32 mask = BIT(d->hwirq & 0x1f);
151 void __iomem *reg = mtk_eint_get_offset(eint, d->hwirq,
152 eint->regs->ack);
153
154 writel(mask, reg);
155}
156
157static int mtk_eint_set_type(struct irq_data *d, unsigned int type)
158{
159 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
160 u32 mask = BIT(d->hwirq & 0x1f);
161 void __iomem *reg;
162
163 if (((type & IRQ_TYPE_EDGE_BOTH) && (type & IRQ_TYPE_LEVEL_MASK)) ||
164 ((type & IRQ_TYPE_LEVEL_MASK) == IRQ_TYPE_LEVEL_MASK)) {
165 dev_err(eint->dev,
166 "Can't configure IRQ%d (EINT%lu) for type 0x%X\n",
167 d->irq, d->hwirq, type);
168 return -EINVAL;
169 }
170
171 if ((type & IRQ_TYPE_EDGE_BOTH) == IRQ_TYPE_EDGE_BOTH)
172 eint->dual_edge[d->hwirq] = 1;
173 else
174 eint->dual_edge[d->hwirq] = 0;
175
176 if (type & (IRQ_TYPE_LEVEL_LOW | IRQ_TYPE_EDGE_FALLING)) {
177 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->pol_clr);
178 writel(mask, reg);
179 } else {
180 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->pol_set);
181 writel(mask, reg);
182 }
183
184 if (type & (IRQ_TYPE_EDGE_RISING | IRQ_TYPE_EDGE_FALLING)) {
185 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->sens_clr);
186 writel(mask, reg);
187 } else {
188 reg = mtk_eint_get_offset(eint, d->hwirq, eint->regs->sens_set);
189 writel(mask, reg);
190 }
191
192 if (eint->dual_edge[d->hwirq])
193 mtk_eint_flip_edge(eint, d->hwirq);
194
195 return 0;
196}
197
198static int mtk_eint_irq_set_wake(struct irq_data *d, unsigned int on)
199{
200 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
201 int shift = d->hwirq & 0x1f;
202 int reg = d->hwirq >> 5;
203
204 if (on)
205 eint->wake_mask[reg] |= BIT(shift);
206 else
207 eint->wake_mask[reg] &= ~BIT(shift);
208
209 return 0;
210}
211
212static void mtk_eint_chip_write_mask(const struct mtk_eint *eint,
213 void __iomem *base, u32 *buf)
214{
215 int port;
216 void __iomem *reg;
217
218 for (port = 0; port < eint->hw->ports; port++) {
219 reg = base + (port << 2);
220 writel_relaxed(~buf[port], reg + eint->regs->mask_set);
221 writel_relaxed(buf[port], reg + eint->regs->mask_clr);
222 }
223}
224
225static int mtk_eint_irq_request_resources(struct irq_data *d)
226{
227 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
228 struct gpio_chip *gpio_c;
229 unsigned int gpio_n;
230 int err;
231
232 err = eint->gpio_xlate->get_gpio_n(eint->pctl, d->hwirq,
233 &gpio_n, &gpio_c);
234 if (err < 0) {
235 dev_err(eint->dev, "Can not find pin\n");
236 return err;
237 }
238
239 err = gpiochip_lock_as_irq(gpio_c, gpio_n);
240 if (err < 0) {
241 dev_err(eint->dev, "unable to lock HW IRQ %lu for IRQ\n",
242 irqd_to_hwirq(d));
243 return err;
244 }
245
246 err = eint->gpio_xlate->set_gpio_as_eint(eint->pctl, d->hwirq);
247 if (err < 0) {
248 dev_err(eint->dev, "Can not eint mode\n");
249 return err;
250 }
251
252 return 0;
253}
254
255static void mtk_eint_irq_release_resources(struct irq_data *d)
256{
257 struct mtk_eint *eint = irq_data_get_irq_chip_data(d);
258 struct gpio_chip *gpio_c;
259 unsigned int gpio_n;
260
261 eint->gpio_xlate->get_gpio_n(eint->pctl, d->hwirq, &gpio_n,
262 &gpio_c);
263
264 gpiochip_unlock_as_irq(gpio_c, gpio_n);
265}
266
267static struct irq_chip mtk_eint_irq_chip = {
268 .name = "mt-eint",
269 .irq_disable = mtk_eint_mask,
270 .irq_mask = mtk_eint_mask,
271 .irq_unmask = mtk_eint_unmask,
272 .irq_ack = mtk_eint_ack,
273 .irq_set_type = mtk_eint_set_type,
274 .irq_set_wake = mtk_eint_irq_set_wake,
275 .irq_request_resources = mtk_eint_irq_request_resources,
276 .irq_release_resources = mtk_eint_irq_release_resources,
277};
278
279static unsigned int mtk_eint_hw_init(struct mtk_eint *eint)
280{
281 void __iomem *reg = eint->base + eint->regs->dom_en;
282 unsigned int i;
283
284 for (i = 0; i < eint->hw->ap_num; i += 32) {
285 writel(0xffffffff, reg);
286 reg += 4;
287 }
288
289 return 0;
290}
291
292static inline void
293mtk_eint_debounce_process(struct mtk_eint *eint, int index)
294{
295 unsigned int rst, ctrl_offset;
296 unsigned int bit, dbnc;
297
298 ctrl_offset = (index / 4) * 4 + eint->regs->dbnc_ctrl;
299 dbnc = readl(eint->base + ctrl_offset);
300 bit = MTK_EINT_DBNC_SET_EN << ((index % 4) * 8);
301 if ((bit & dbnc) > 0) {
302 ctrl_offset = (index / 4) * 4 + eint->regs->dbnc_set;
303 rst = MTK_EINT_DBNC_RST_BIT << ((index % 4) * 8);
304 writel(rst, eint->base + ctrl_offset);
305 }
306}
307
308static void mtk_eint_irq_handler(struct irq_desc *desc)
309{
310 struct irq_chip *chip = irq_desc_get_chip(desc);
311 struct mtk_eint *eint = irq_desc_get_handler_data(desc);
312 unsigned int status, eint_num;
313 int offset, mask_offset, index, virq;
314 void __iomem *reg = mtk_eint_get_offset(eint, 0, eint->regs->stat);
315 int dual_edge, start_level, curr_level;
316
317 chained_irq_enter(chip, desc);
318 for (eint_num = 0; eint_num < eint->hw->ap_num; eint_num += 32,
319 reg += 4) {
320 status = readl(reg);
321 while (status) {
322 offset = __ffs(status);
323 mask_offset = eint_num >> 5;
324 index = eint_num + offset;
325 virq = irq_find_mapping(eint->domain, index);
326 status &= ~BIT(offset);
327
328 /*
329 * If we get an interrupt on pin that was only required
330 * for wake (but no real interrupt requested), mask the
331 * interrupt (as would mtk_eint_resume do anyway later
332 * in the resume sequence).
333 */
334 if (eint->wake_mask[mask_offset] & BIT(offset) &&
335 !(eint->cur_mask[mask_offset] & BIT(offset))) {
336 writel_relaxed(BIT(offset), reg -
337 eint->regs->stat +
338 eint->regs->mask_set);
339 }
340
341 dual_edge = eint->dual_edge[index];
342 if (dual_edge) {
343 /*
344 * Clear soft-irq in case we raised it last
345 * time.
346 */
347 writel(BIT(offset), reg - eint->regs->stat +
348 eint->regs->soft_clr);
349
350 start_level =
351 eint->gpio_xlate->get_gpio_state(eint->pctl,
352 index);
353 }
354
355 generic_handle_irq(virq);
356
357 if (dual_edge) {
358 curr_level = mtk_eint_flip_edge(eint, index);
359
360 /*
361 * If level changed, we might lost one edge
362 * interrupt, raised it through soft-irq.
363 */
364 if (start_level != curr_level)
365 writel(BIT(offset), reg -
366 eint->regs->stat +
367 eint->regs->soft_set);
368 }
369
370 if (index < eint->hw->db_cnt)
371 mtk_eint_debounce_process(eint, index);
372 }
373 }
374 chained_irq_exit(chip, desc);
375}
376
377int mtk_eint_do_suspend(struct mtk_eint *eint)
378{
379 mtk_eint_chip_write_mask(eint, eint->base, eint->wake_mask);
380
381 return 0;
382}
383EXPORT_SYMBOL_GPL(mtk_eint_do_suspend);
384
385int mtk_eint_do_resume(struct mtk_eint *eint)
386{
387 mtk_eint_chip_write_mask(eint, eint->base, eint->cur_mask);
388
389 return 0;
390}
391EXPORT_SYMBOL_GPL(mtk_eint_do_resume);
392
393int mtk_eint_set_debounce(struct mtk_eint *eint, unsigned long eint_num,
394 unsigned int debounce)
395{
396 int virq, eint_offset;
397 unsigned int set_offset, bit, clr_bit, clr_offset, rst, i, unmask,
398 dbnc;
399 static const unsigned int debounce_time[] = {500, 1000, 16000, 32000,
400 64000, 128000, 256000};
401 struct irq_data *d;
402
403 virq = irq_find_mapping(eint->domain, eint_num);
404 eint_offset = (eint_num % 4) * 8;
405 d = irq_get_irq_data(virq);
406
407 set_offset = (eint_num / 4) * 4 + eint->regs->dbnc_set;
408 clr_offset = (eint_num / 4) * 4 + eint->regs->dbnc_clr;
409
410 if (!mtk_eint_can_en_debounce(eint, eint_num))
411 return -EINVAL;
412
413 dbnc = ARRAY_SIZE(debounce_time);
414 for (i = 0; i < ARRAY_SIZE(debounce_time); i++) {
415 if (debounce <= debounce_time[i]) {
416 dbnc = i;
417 break;
418 }
419 }
420
421 if (!mtk_eint_get_mask(eint, eint_num)) {
422 mtk_eint_mask(d);
423 unmask = 1;
424 } else {
425 unmask = 0;
426 }
427
428 clr_bit = 0xff << eint_offset;
429 writel(clr_bit, eint->base + clr_offset);
430
431 bit = ((dbnc << MTK_EINT_DBNC_SET_DBNC_BITS) | MTK_EINT_DBNC_SET_EN) <<
432 eint_offset;
433 rst = MTK_EINT_DBNC_RST_BIT << eint_offset;
434 writel(rst | bit, eint->base + set_offset);
435
436 /*
437 * Delay a while (more than 2T) to wait for hw debounce counter reset
438 * work correctly.
439 */
440 udelay(1);
441 if (unmask == 1)
442 mtk_eint_unmask(d);
443
444 return 0;
445}
446EXPORT_SYMBOL_GPL(mtk_eint_set_debounce);
447
448int mtk_eint_find_irq(struct mtk_eint *eint, unsigned long eint_n)
449{
450 int irq;
451
452 irq = irq_find_mapping(eint->domain, eint_n);
453 if (!irq)
454 return -EINVAL;
455
456 return irq;
457}
458EXPORT_SYMBOL_GPL(mtk_eint_find_irq);
459
460int mtk_eint_do_init(struct mtk_eint *eint)
461{
462 int i;
463
464 /* If clients don't assign a specific regs, let's use generic one */
465 if (!eint->regs)
466 eint->regs = &mtk_generic_eint_regs;
467
468 eint->wake_mask = devm_kcalloc(eint->dev, eint->hw->ports,
469 sizeof(*eint->wake_mask), GFP_KERNEL);
470 if (!eint->wake_mask)
471 return -ENOMEM;
472
473 eint->cur_mask = devm_kcalloc(eint->dev, eint->hw->ports,
474 sizeof(*eint->cur_mask), GFP_KERNEL);
475 if (!eint->cur_mask)
476 return -ENOMEM;
477
478 eint->dual_edge = devm_kcalloc(eint->dev, eint->hw->ap_num,
479 sizeof(int), GFP_KERNEL);
480 if (!eint->dual_edge)
481 return -ENOMEM;
482
483 eint->domain = irq_domain_add_linear(eint->dev->of_node,
484 eint->hw->ap_num,
485 &irq_domain_simple_ops, NULL);
486 if (!eint->domain)
487 return -ENOMEM;
488
489 mtk_eint_hw_init(eint);
490 for (i = 0; i < eint->hw->ap_num; i++) {
491 int virq = irq_create_mapping(eint->domain, i);
492
493 irq_set_chip_and_handler(virq, &mtk_eint_irq_chip,
494 handle_level_irq);
495 irq_set_chip_data(virq, eint);
496 }
497
498 irq_set_chained_handler_and_data(eint->irq, mtk_eint_irq_handler,
499 eint);
500
501 return 0;
502}
503EXPORT_SYMBOL_GPL(mtk_eint_do_init);
504
505MODULE_LICENSE("GPL v2");
506MODULE_DESCRIPTION("MediaTek EINT Driver");