Linux Audio

Check our new training course

Loading...
v4.17
 
  1/*
  2 * This program is free software; you can redistribute it and/or modify
  3 * it under the terms of the GNU General Public License as published by
  4 * the Free Software Foundation; either version 2 of the License, or
  5 * (at your option) any later version.
  6 *
  7 * This program is distributed in the hope that it will be useful,
  8 * but WITHOUT ANY WARRANTY; without even the implied warranty of
  9 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 10 * GNU General Public License for more details.
 11 *
 12 * You should have received a copy of the GNU General Public License
 13 * along with this program; if not, write to the Free Software
 14 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA
 15 *
 16 * Copyright (c) 2005 Linas Vepstas <linas@linas.org>
 17 */
 18
 19#include <linux/delay.h>
 20#include <linux/list.h>
 21#include <linux/sched.h>
 22#include <linux/semaphore.h>
 23#include <linux/pci.h>
 24#include <linux/slab.h>
 25#include <linux/kthread.h>
 26#include <asm/eeh_event.h>
 27#include <asm/ppc-pci.h>
 28
 29/** Overview:
 30 *  EEH error states may be detected within exception handlers;
 31 *  however, the recovery processing needs to occur asynchronously
 32 *  in a normal kernel context and not an interrupt context.
 33 *  This pair of routines creates an event and queues it onto a
 34 *  work-queue, where a worker thread can drive recovery.
 35 */
 36
 37static DEFINE_SPINLOCK(eeh_eventlist_lock);
 38static struct semaphore eeh_eventlist_sem;
 39static LIST_HEAD(eeh_eventlist);
 40
 41/**
 42 * eeh_event_handler - Dispatch EEH events.
 43 * @dummy - unused
 44 *
 45 * The detection of a frozen slot can occur inside an interrupt,
 46 * where it can be hard to do anything about it.  The goal of this
 47 * routine is to pull these detection events out of the context
 48 * of the interrupt handler, and re-dispatch them for processing
 49 * at a later time in a normal context.
 50 */
 51static int eeh_event_handler(void * dummy)
 52{
 53	unsigned long flags;
 54	struct eeh_event *event;
 55	struct eeh_pe *pe;
 56
 57	while (!kthread_should_stop()) {
 58		if (down_interruptible(&eeh_eventlist_sem))
 59			break;
 60
 61		/* Fetch EEH event from the queue */
 62		spin_lock_irqsave(&eeh_eventlist_lock, flags);
 63		event = NULL;
 64		if (!list_empty(&eeh_eventlist)) {
 65			event = list_entry(eeh_eventlist.next,
 66					   struct eeh_event, list);
 67			list_del(&event->list);
 68		}
 69		spin_unlock_irqrestore(&eeh_eventlist_lock, flags);
 70		if (!event)
 71			continue;
 72
 73		/* We might have event without binding PE */
 74		pe = event->pe;
 75		if (pe) {
 76			if (pe->type & EEH_PE_PHB)
 77				pr_info("EEH: Detected error on PHB#%x\n",
 78					 pe->phb->global_number);
 79			else
 80				pr_info("EEH: Detected PCI bus error on "
 81					"PHB#%x-PE#%x\n",
 82					pe->phb->global_number, pe->addr);
 83			eeh_handle_normal_event(pe);
 84		} else {
 85			eeh_handle_special_event();
 86		}
 87
 88		kfree(event);
 89	}
 90
 91	return 0;
 92}
 93
 94/**
 95 * eeh_event_init - Start kernel thread to handle EEH events
 96 *
 97 * This routine is called to start the kernel thread for processing
 98 * EEH event.
 99 */
100int eeh_event_init(void)
101{
102	struct task_struct *t;
103	int ret = 0;
104
105	/* Initialize semaphore */
106	sema_init(&eeh_eventlist_sem, 0);
107
108	t = kthread_run(eeh_event_handler, NULL, "eehd");
109	if (IS_ERR(t)) {
110		ret = PTR_ERR(t);
111		pr_err("%s: Failed to start EEH daemon (%d)\n",
112			__func__, ret);
113		return ret;
114	}
115
116	return 0;
117}
118
119/**
120 * eeh_send_failure_event - Generate a PCI error event
121 * @pe: EEH PE
122 *
123 * This routine can be called within an interrupt context;
124 * the actual event will be delivered in a normal context
125 * (from a workqueue).
126 */
127int eeh_send_failure_event(struct eeh_pe *pe)
128{
129	unsigned long flags;
130	struct eeh_event *event;
131
132	event = kzalloc(sizeof(*event), GFP_ATOMIC);
133	if (!event) {
134		pr_err("EEH: out of memory, event not handled\n");
135		return -ENOMEM;
136	}
137	event->pe = pe;
138
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
139	/* We may or may not be called in an interrupt context */
140	spin_lock_irqsave(&eeh_eventlist_lock, flags);
141	list_add(&event->list, &eeh_eventlist);
142	spin_unlock_irqrestore(&eeh_eventlist_lock, flags);
143
144	/* For EEH deamon to knick in */
145	up(&eeh_eventlist_sem);
146
147	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
148}
149
150/**
151 * eeh_remove_event - Remove EEH event from the queue
152 * @pe: Event binding to the PE
153 * @force: Event will be removed unconditionally
154 *
155 * On PowerNV platform, we might have subsequent coming events
156 * is part of the former one. For that case, those subsequent
157 * coming events are totally duplicated and unnecessary, thus
158 * they should be removed.
159 */
160void eeh_remove_event(struct eeh_pe *pe, bool force)
161{
162	unsigned long flags;
163	struct eeh_event *event, *tmp;
164
165	/*
166	 * If we have NULL PE passed in, we have dead IOC
167	 * or we're sure we can report all existing errors
168	 * by the caller.
169	 *
170	 * With "force", the event with associated PE that
171	 * have been isolated, the event won't be removed
172	 * to avoid event lost.
173	 */
174	spin_lock_irqsave(&eeh_eventlist_lock, flags);
175	list_for_each_entry_safe(event, tmp, &eeh_eventlist, list) {
176		if (!force && event->pe &&
177		    (event->pe->state & EEH_PE_ISOLATED))
178			continue;
179
180		if (!pe) {
181			list_del(&event->list);
182			kfree(event);
183		} else if (pe->type & EEH_PE_PHB) {
184			if (event->pe && event->pe->phb == pe->phb) {
185				list_del(&event->list);
186				kfree(event);
187			}
188		} else if (event->pe == pe) {
189			list_del(&event->list);
190			kfree(event);
191		}
192	}
193	spin_unlock_irqrestore(&eeh_eventlist_lock, flags);
194}
v6.8
  1// SPDX-License-Identifier: GPL-2.0-or-later
  2/*
 
 
 
 
 
 
 
 
 
 
 
 
 
  3 *
  4 * Copyright (c) 2005 Linas Vepstas <linas@linas.org>
  5 */
  6
  7#include <linux/delay.h>
  8#include <linux/list.h>
  9#include <linux/sched.h>
 10#include <linux/semaphore.h>
 11#include <linux/pci.h>
 12#include <linux/slab.h>
 13#include <linux/kthread.h>
 14#include <asm/eeh_event.h>
 15#include <asm/ppc-pci.h>
 16
 17/** Overview:
 18 *  EEH error states may be detected within exception handlers;
 19 *  however, the recovery processing needs to occur asynchronously
 20 *  in a normal kernel context and not an interrupt context.
 21 *  This pair of routines creates an event and queues it onto a
 22 *  work-queue, where a worker thread can drive recovery.
 23 */
 24
 25static DEFINE_SPINLOCK(eeh_eventlist_lock);
 26static DECLARE_COMPLETION(eeh_eventlist_event);
 27static LIST_HEAD(eeh_eventlist);
 28
 29/**
 30 * eeh_event_handler - Dispatch EEH events.
 31 * @dummy - unused
 32 *
 33 * The detection of a frozen slot can occur inside an interrupt,
 34 * where it can be hard to do anything about it.  The goal of this
 35 * routine is to pull these detection events out of the context
 36 * of the interrupt handler, and re-dispatch them for processing
 37 * at a later time in a normal context.
 38 */
 39static int eeh_event_handler(void * dummy)
 40{
 41	unsigned long flags;
 42	struct eeh_event *event;
 
 43
 44	while (!kthread_should_stop()) {
 45		if (wait_for_completion_interruptible(&eeh_eventlist_event))
 46			break;
 47
 48		/* Fetch EEH event from the queue */
 49		spin_lock_irqsave(&eeh_eventlist_lock, flags);
 50		event = NULL;
 51		if (!list_empty(&eeh_eventlist)) {
 52			event = list_entry(eeh_eventlist.next,
 53					   struct eeh_event, list);
 54			list_del(&event->list);
 55		}
 56		spin_unlock_irqrestore(&eeh_eventlist_lock, flags);
 57		if (!event)
 58			continue;
 59
 60		/* We might have event without binding PE */
 61		if (event->pe)
 62			eeh_handle_normal_event(event->pe);
 63		else
 
 
 
 
 
 
 
 
 64			eeh_handle_special_event();
 
 65
 66		kfree(event);
 67	}
 68
 69	return 0;
 70}
 71
 72/**
 73 * eeh_event_init - Start kernel thread to handle EEH events
 74 *
 75 * This routine is called to start the kernel thread for processing
 76 * EEH event.
 77 */
 78int eeh_event_init(void)
 79{
 80	struct task_struct *t;
 81	int ret = 0;
 82
 
 
 
 83	t = kthread_run(eeh_event_handler, NULL, "eehd");
 84	if (IS_ERR(t)) {
 85		ret = PTR_ERR(t);
 86		pr_err("%s: Failed to start EEH daemon (%d)\n",
 87			__func__, ret);
 88		return ret;
 89	}
 90
 91	return 0;
 92}
 93
 94/**
 95 * eeh_send_failure_event - Generate a PCI error event
 96 * @pe: EEH PE
 97 *
 98 * This routine can be called within an interrupt context;
 99 * the actual event will be delivered in a normal context
100 * (from a workqueue).
101 */
102int __eeh_send_failure_event(struct eeh_pe *pe)
103{
104	unsigned long flags;
105	struct eeh_event *event;
106
107	event = kzalloc(sizeof(*event), GFP_ATOMIC);
108	if (!event) {
109		pr_err("EEH: out of memory, event not handled\n");
110		return -ENOMEM;
111	}
112	event->pe = pe;
113
114	/*
115	 * Mark the PE as recovering before inserting it in the queue.
116	 * This prevents the PE from being free()ed by a hotplug driver
117	 * while the PE is sitting in the event queue.
118	 */
119	if (pe) {
120#ifdef CONFIG_STACKTRACE
121		/*
122		 * Save the current stack trace so we can dump it from the
123		 * event handler thread.
124		 */
125		pe->trace_entries = stack_trace_save(pe->stack_trace,
126					 ARRAY_SIZE(pe->stack_trace), 0);
127#endif /* CONFIG_STACKTRACE */
128
129		eeh_pe_state_mark(pe, EEH_PE_RECOVERING);
130	}
131
132	/* We may or may not be called in an interrupt context */
133	spin_lock_irqsave(&eeh_eventlist_lock, flags);
134	list_add(&event->list, &eeh_eventlist);
135	spin_unlock_irqrestore(&eeh_eventlist_lock, flags);
136
137	/* For EEH deamon to knick in */
138	complete(&eeh_eventlist_event);
139
140	return 0;
141}
142
143int eeh_send_failure_event(struct eeh_pe *pe)
144{
145	/*
146	 * If we've manually suppressed recovery events via debugfs
147	 * then just drop it on the floor.
148	 */
149	if (eeh_debugfs_no_recover) {
150		pr_err("EEH: Event dropped due to no_recover setting\n");
151		return 0;
152	}
153
154	return __eeh_send_failure_event(pe);
155}
156
157/**
158 * eeh_remove_event - Remove EEH event from the queue
159 * @pe: Event binding to the PE
160 * @force: Event will be removed unconditionally
161 *
162 * On PowerNV platform, we might have subsequent coming events
163 * is part of the former one. For that case, those subsequent
164 * coming events are totally duplicated and unnecessary, thus
165 * they should be removed.
166 */
167void eeh_remove_event(struct eeh_pe *pe, bool force)
168{
169	unsigned long flags;
170	struct eeh_event *event, *tmp;
171
172	/*
173	 * If we have NULL PE passed in, we have dead IOC
174	 * or we're sure we can report all existing errors
175	 * by the caller.
176	 *
177	 * With "force", the event with associated PE that
178	 * have been isolated, the event won't be removed
179	 * to avoid event lost.
180	 */
181	spin_lock_irqsave(&eeh_eventlist_lock, flags);
182	list_for_each_entry_safe(event, tmp, &eeh_eventlist, list) {
183		if (!force && event->pe &&
184		    (event->pe->state & EEH_PE_ISOLATED))
185			continue;
186
187		if (!pe) {
188			list_del(&event->list);
189			kfree(event);
190		} else if (pe->type & EEH_PE_PHB) {
191			if (event->pe && event->pe->phb == pe->phb) {
192				list_del(&event->list);
193				kfree(event);
194			}
195		} else if (event->pe == pe) {
196			list_del(&event->list);
197			kfree(event);
198		}
199	}
200	spin_unlock_irqrestore(&eeh_eventlist_lock, flags);
201}