Linux Audio

Check our new training course

Linux debugging, profiling, tracing and performance analysis training

Apr 14-17, 2025
Register
Loading...
v3.5.6
  1/*
  2 * APEI Error INJection support
  3 *
  4 * EINJ provides a hardware error injection mechanism, this is useful
  5 * for debugging and testing of other APEI and RAS features.
  6 *
  7 * For more information about EINJ, please refer to ACPI Specification
  8 * version 4.0, section 17.5.
  9 *
 10 * Copyright 2009-2010 Intel Corp.
 11 *   Author: Huang Ying <ying.huang@intel.com>
 12 *
 13 * This program is free software; you can redistribute it and/or
 14 * modify it under the terms of the GNU General Public License version
 15 * 2 as published by the Free Software Foundation.
 16 *
 17 * This program is distributed in the hope that it will be useful,
 18 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 20 * GNU General Public License for more details.
 21 *
 22 * You should have received a copy of the GNU General Public License
 23 * along with this program; if not, write to the Free Software
 24 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
 25 */
 26
 27#include <linux/kernel.h>
 28#include <linux/module.h>
 29#include <linux/init.h>
 30#include <linux/io.h>
 31#include <linux/debugfs.h>
 32#include <linux/seq_file.h>
 33#include <linux/nmi.h>
 34#include <linux/delay.h>
 35#include <acpi/acpi.h>
 
 36
 37#include "apei-internal.h"
 38
 39#define EINJ_PFX "EINJ: "
 40
 41#define SPIN_UNIT		100			/* 100ns */
 42/* Firmware should respond within 1 milliseconds */
 43#define FIRMWARE_TIMEOUT	(1 * NSEC_PER_MSEC)
 
 
 
 
 44
 45/*
 46 * ACPI version 5 provides a SET_ERROR_TYPE_WITH_ADDRESS action.
 47 */
 48static int acpi5;
 49
 50struct set_error_type_with_address {
 51	u32	type;
 52	u32	vendor_extension;
 53	u32	flags;
 54	u32	apicid;
 55	u64	memory_address;
 56	u64	memory_address_range;
 57	u32	pcie_sbdf;
 58};
 59enum {
 60	SETWA_FLAGS_APICID = 1,
 61	SETWA_FLAGS_MEM = 2,
 62	SETWA_FLAGS_PCIE_SBDF = 4,
 63};
 64
 65/*
 66 * Vendor extensions for platform specific operations
 67 */
 68struct vendor_error_type_extension {
 69	u32	length;
 70	u32	pcie_sbdf;
 71	u16	vendor_id;
 72	u16	device_id;
 73	u8	rev_id;
 74	u8	reserved[3];
 75};
 76
 77static u32 notrigger;
 78
 79static u32 vendor_flags;
 80static struct debugfs_blob_wrapper vendor_blob;
 81static char vendor_dev[64];
 82
 83/*
 84 * Some BIOSes allow parameters to the SET_ERROR_TYPE entries in the
 85 * EINJ table through an unpublished extension. Use with caution as
 86 * most will ignore the parameter and make their own choice of address
 87 * for error injection.  This extension is used only if
 88 * param_extension module parameter is specified.
 89 */
 90struct einj_parameter {
 91	u64 type;
 92	u64 reserved1;
 93	u64 reserved2;
 94	u64 param1;
 95	u64 param2;
 96};
 97
 98#define EINJ_OP_BUSY			0x1
 99#define EINJ_STATUS_SUCCESS		0x0
100#define EINJ_STATUS_FAIL		0x1
101#define EINJ_STATUS_INVAL		0x2
102
103#define EINJ_TAB_ENTRY(tab)						\
104	((struct acpi_whea_header *)((char *)(tab) +			\
105				    sizeof(struct acpi_table_einj)))
106
107static bool param_extension;
108module_param(param_extension, bool, 0);
109
110static struct acpi_table_einj *einj_tab;
111
112static struct apei_resources einj_resources;
113
114static struct apei_exec_ins_type einj_ins_type[] = {
115	[ACPI_EINJ_READ_REGISTER] = {
116		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
117		.run   = apei_exec_read_register,
118	},
119	[ACPI_EINJ_READ_REGISTER_VALUE] = {
120		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
121		.run   = apei_exec_read_register_value,
122	},
123	[ACPI_EINJ_WRITE_REGISTER] = {
124		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
125		.run   = apei_exec_write_register,
126	},
127	[ACPI_EINJ_WRITE_REGISTER_VALUE] = {
128		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
129		.run   = apei_exec_write_register_value,
130	},
131	[ACPI_EINJ_NOOP] = {
132		.flags = 0,
133		.run   = apei_exec_noop,
134	},
135};
136
137/*
138 * Prevent EINJ interpreter to run simultaneously, because the
139 * corresponding firmware implementation may not work properly when
140 * invoked simultaneously.
141 */
142static DEFINE_MUTEX(einj_mutex);
143
144static void *einj_param;
145
146static void einj_exec_ctx_init(struct apei_exec_context *ctx)
147{
148	apei_exec_ctx_init(ctx, einj_ins_type, ARRAY_SIZE(einj_ins_type),
149			   EINJ_TAB_ENTRY(einj_tab), einj_tab->entries);
150}
151
152static int __einj_get_available_error_type(u32 *type)
153{
154	struct apei_exec_context ctx;
155	int rc;
156
157	einj_exec_ctx_init(&ctx);
158	rc = apei_exec_run(&ctx, ACPI_EINJ_GET_ERROR_TYPE);
159	if (rc)
160		return rc;
161	*type = apei_exec_ctx_get_output(&ctx);
162
163	return 0;
164}
165
166/* Get error injection capabilities of the platform */
167static int einj_get_available_error_type(u32 *type)
168{
169	int rc;
170
171	mutex_lock(&einj_mutex);
172	rc = __einj_get_available_error_type(type);
173	mutex_unlock(&einj_mutex);
174
175	return rc;
176}
177
178static int einj_timedout(u64 *t)
179{
180	if ((s64)*t < SPIN_UNIT) {
181		pr_warning(FW_WARN EINJ_PFX
182			   "Firmware does not respond in time\n");
183		return 1;
184	}
185	*t -= SPIN_UNIT;
186	ndelay(SPIN_UNIT);
187	touch_nmi_watchdog();
188	return 0;
189}
190
191static void check_vendor_extension(u64 paddr,
192				   struct set_error_type_with_address *v5param)
193{
194	int	offset = v5param->vendor_extension;
195	struct	vendor_error_type_extension *v;
196	u32	sbdf;
197
198	if (!offset)
199		return;
200	v = acpi_os_map_memory(paddr + offset, sizeof(*v));
201	if (!v)
202		return;
203	sbdf = v->pcie_sbdf;
204	sprintf(vendor_dev, "%x:%x:%x.%x vendor_id=%x device_id=%x rev_id=%x\n",
205		sbdf >> 24, (sbdf >> 16) & 0xff,
206		(sbdf >> 11) & 0x1f, (sbdf >> 8) & 0x7,
207		 v->vendor_id, v->device_id, v->rev_id);
208	acpi_os_unmap_memory(v, sizeof(*v));
209}
210
211static void *einj_get_parameter_address(void)
212{
213	int i;
214	u64 paddrv4 = 0, paddrv5 = 0;
215	struct acpi_whea_header *entry;
216
217	entry = EINJ_TAB_ENTRY(einj_tab);
218	for (i = 0; i < einj_tab->entries; i++) {
219		if (entry->action == ACPI_EINJ_SET_ERROR_TYPE &&
220		    entry->instruction == ACPI_EINJ_WRITE_REGISTER &&
221		    entry->register_region.space_id ==
222		    ACPI_ADR_SPACE_SYSTEM_MEMORY)
223			memcpy(&paddrv4, &entry->register_region.address,
224			       sizeof(paddrv4));
225		if (entry->action == ACPI_EINJ_SET_ERROR_TYPE_WITH_ADDRESS &&
226		    entry->instruction == ACPI_EINJ_WRITE_REGISTER &&
227		    entry->register_region.space_id ==
228		    ACPI_ADR_SPACE_SYSTEM_MEMORY)
229			memcpy(&paddrv5, &entry->register_region.address,
230			       sizeof(paddrv5));
231		entry++;
232	}
233	if (paddrv5) {
234		struct set_error_type_with_address *v5param;
235
236		v5param = acpi_os_map_memory(paddrv5, sizeof(*v5param));
237		if (v5param) {
238			acpi5 = 1;
239			check_vendor_extension(paddrv5, v5param);
240			return v5param;
241		}
242	}
243	if (param_extension && paddrv4) {
244		struct einj_parameter *v4param;
245
246		v4param = acpi_os_map_memory(paddrv4, sizeof(*v4param));
247		if (!v4param)
248			return NULL;
249		if (v4param->reserved1 || v4param->reserved2) {
250			acpi_os_unmap_memory(v4param, sizeof(*v4param));
251			return NULL;
252		}
253		return v4param;
254	}
255
256	return NULL;
257}
258
259/* do sanity check to trigger table */
260static int einj_check_trigger_header(struct acpi_einj_trigger *trigger_tab)
261{
262	if (trigger_tab->header_size != sizeof(struct acpi_einj_trigger))
263		return -EINVAL;
264	if (trigger_tab->table_size > PAGE_SIZE ||
265	    trigger_tab->table_size < trigger_tab->header_size)
266		return -EINVAL;
267	if (trigger_tab->entry_count !=
268	    (trigger_tab->table_size - trigger_tab->header_size) /
269	    sizeof(struct acpi_einj_entry))
270		return -EINVAL;
271
272	return 0;
273}
274
275static struct acpi_generic_address *einj_get_trigger_parameter_region(
276	struct acpi_einj_trigger *trigger_tab, u64 param1, u64 param2)
277{
278	int i;
279	struct acpi_whea_header *entry;
280
281	entry = (struct acpi_whea_header *)
282		((char *)trigger_tab + sizeof(struct acpi_einj_trigger));
283	for (i = 0; i < trigger_tab->entry_count; i++) {
284		if (entry->action == ACPI_EINJ_TRIGGER_ERROR &&
285		entry->instruction == ACPI_EINJ_WRITE_REGISTER_VALUE &&
286		entry->register_region.space_id ==
287			ACPI_ADR_SPACE_SYSTEM_MEMORY &&
288		(entry->register_region.address & param2) == (param1 & param2))
289			return &entry->register_region;
290		entry++;
291	}
292
293	return NULL;
294}
295/* Execute instructions in trigger error action table */
296static int __einj_error_trigger(u64 trigger_paddr, u32 type,
297				u64 param1, u64 param2)
298{
299	struct acpi_einj_trigger *trigger_tab = NULL;
300	struct apei_exec_context trigger_ctx;
301	struct apei_resources trigger_resources;
302	struct acpi_whea_header *trigger_entry;
303	struct resource *r;
304	u32 table_size;
305	int rc = -EIO;
306	struct acpi_generic_address *trigger_param_region = NULL;
307
308	r = request_mem_region(trigger_paddr, sizeof(*trigger_tab),
309			       "APEI EINJ Trigger Table");
310	if (!r) {
311		pr_err(EINJ_PFX
312	"Can not request [mem %#010llx-%#010llx] for Trigger table\n",
313		       (unsigned long long)trigger_paddr,
314		       (unsigned long long)trigger_paddr +
315			    sizeof(*trigger_tab) - 1);
316		goto out;
317	}
318	trigger_tab = ioremap_cache(trigger_paddr, sizeof(*trigger_tab));
319	if (!trigger_tab) {
320		pr_err(EINJ_PFX "Failed to map trigger table!\n");
321		goto out_rel_header;
322	}
323	rc = einj_check_trigger_header(trigger_tab);
324	if (rc) {
325		pr_warning(FW_BUG EINJ_PFX
326			   "The trigger error action table is invalid\n");
327		goto out_rel_header;
328	}
329
330	/* No action structures in the TRIGGER_ERROR table, nothing to do */
331	if (!trigger_tab->entry_count)
332		goto out_rel_header;
333
334	rc = -EIO;
335	table_size = trigger_tab->table_size;
336	r = request_mem_region(trigger_paddr + sizeof(*trigger_tab),
337			       table_size - sizeof(*trigger_tab),
338			       "APEI EINJ Trigger Table");
339	if (!r) {
340		pr_err(EINJ_PFX
341"Can not request [mem %#010llx-%#010llx] for Trigger Table Entry\n",
342		       (unsigned long long)trigger_paddr + sizeof(*trigger_tab),
343		       (unsigned long long)trigger_paddr + table_size - 1);
344		goto out_rel_header;
345	}
346	iounmap(trigger_tab);
347	trigger_tab = ioremap_cache(trigger_paddr, table_size);
348	if (!trigger_tab) {
349		pr_err(EINJ_PFX "Failed to map trigger table!\n");
350		goto out_rel_entry;
351	}
352	trigger_entry = (struct acpi_whea_header *)
353		((char *)trigger_tab + sizeof(struct acpi_einj_trigger));
354	apei_resources_init(&trigger_resources);
355	apei_exec_ctx_init(&trigger_ctx, einj_ins_type,
356			   ARRAY_SIZE(einj_ins_type),
357			   trigger_entry, trigger_tab->entry_count);
358	rc = apei_exec_collect_resources(&trigger_ctx, &trigger_resources);
359	if (rc)
360		goto out_fini;
361	rc = apei_resources_sub(&trigger_resources, &einj_resources);
362	if (rc)
363		goto out_fini;
364	/*
365	 * Some firmware will access target address specified in
366	 * param1 to trigger the error when injecting memory error.
367	 * This will cause resource conflict with regular memory.  So
368	 * remove it from trigger table resources.
369	 */
370	if (param_extension && (type & 0x0038) && param2) {
371		struct apei_resources addr_resources;
372		apei_resources_init(&addr_resources);
373		trigger_param_region = einj_get_trigger_parameter_region(
374			trigger_tab, param1, param2);
375		if (trigger_param_region) {
376			rc = apei_resources_add(&addr_resources,
377				trigger_param_region->address,
378				trigger_param_region->bit_width/8, true);
379			if (rc)
380				goto out_fini;
381			rc = apei_resources_sub(&trigger_resources,
382					&addr_resources);
383		}
384		apei_resources_fini(&addr_resources);
385		if (rc)
386			goto out_fini;
387	}
388	rc = apei_resources_request(&trigger_resources, "APEI EINJ Trigger");
389	if (rc)
390		goto out_fini;
391	rc = apei_exec_pre_map_gars(&trigger_ctx);
392	if (rc)
393		goto out_release;
394
395	rc = apei_exec_run(&trigger_ctx, ACPI_EINJ_TRIGGER_ERROR);
396
397	apei_exec_post_unmap_gars(&trigger_ctx);
398out_release:
399	apei_resources_release(&trigger_resources);
400out_fini:
401	apei_resources_fini(&trigger_resources);
402out_rel_entry:
403	release_mem_region(trigger_paddr + sizeof(*trigger_tab),
404			   table_size - sizeof(*trigger_tab));
405out_rel_header:
406	release_mem_region(trigger_paddr, sizeof(*trigger_tab));
407out:
408	if (trigger_tab)
409		iounmap(trigger_tab);
410
411	return rc;
412}
413
414static int __einj_error_inject(u32 type, u64 param1, u64 param2)
 
415{
416	struct apei_exec_context ctx;
417	u64 val, trigger_paddr, timeout = FIRMWARE_TIMEOUT;
418	int rc;
419
420	einj_exec_ctx_init(&ctx);
421
422	rc = apei_exec_run_optional(&ctx, ACPI_EINJ_BEGIN_OPERATION);
423	if (rc)
424		return rc;
425	apei_exec_ctx_set_input(&ctx, type);
426	if (acpi5) {
427		struct set_error_type_with_address *v5param = einj_param;
428
429		v5param->type = type;
430		if (type & 0x80000000) {
431			switch (vendor_flags) {
432			case SETWA_FLAGS_APICID:
433				v5param->apicid = param1;
434				break;
435			case SETWA_FLAGS_MEM:
436				v5param->memory_address = param1;
437				v5param->memory_address_range = param2;
438				break;
439			case SETWA_FLAGS_PCIE_SBDF:
440				v5param->pcie_sbdf = param1;
441				break;
442			}
443			v5param->flags = vendor_flags;
 
 
 
 
 
 
444		} else {
445			switch (type) {
446			case ACPI_EINJ_PROCESSOR_CORRECTABLE:
447			case ACPI_EINJ_PROCESSOR_UNCORRECTABLE:
448			case ACPI_EINJ_PROCESSOR_FATAL:
449				v5param->apicid = param1;
450				v5param->flags = SETWA_FLAGS_APICID;
451				break;
452			case ACPI_EINJ_MEMORY_CORRECTABLE:
453			case ACPI_EINJ_MEMORY_UNCORRECTABLE:
454			case ACPI_EINJ_MEMORY_FATAL:
455				v5param->memory_address = param1;
456				v5param->memory_address_range = param2;
457				v5param->flags = SETWA_FLAGS_MEM;
458				break;
459			case ACPI_EINJ_PCIX_CORRECTABLE:
460			case ACPI_EINJ_PCIX_UNCORRECTABLE:
461			case ACPI_EINJ_PCIX_FATAL:
462				v5param->pcie_sbdf = param1;
463				v5param->flags = SETWA_FLAGS_PCIE_SBDF;
464				break;
465			}
466		}
467	} else {
468		rc = apei_exec_run(&ctx, ACPI_EINJ_SET_ERROR_TYPE);
469		if (rc)
470			return rc;
471		if (einj_param) {
472			struct einj_parameter *v4param = einj_param;
473			v4param->param1 = param1;
474			v4param->param2 = param2;
475		}
476	}
477	rc = apei_exec_run(&ctx, ACPI_EINJ_EXECUTE_OPERATION);
478	if (rc)
479		return rc;
480	for (;;) {
481		rc = apei_exec_run(&ctx, ACPI_EINJ_CHECK_BUSY_STATUS);
482		if (rc)
483			return rc;
484		val = apei_exec_ctx_get_output(&ctx);
485		if (!(val & EINJ_OP_BUSY))
486			break;
487		if (einj_timedout(&timeout))
488			return -EIO;
489	}
490	rc = apei_exec_run(&ctx, ACPI_EINJ_GET_COMMAND_STATUS);
491	if (rc)
492		return rc;
493	val = apei_exec_ctx_get_output(&ctx);
494	if (val != EINJ_STATUS_SUCCESS)
495		return -EBUSY;
496
497	rc = apei_exec_run(&ctx, ACPI_EINJ_GET_TRIGGER_TABLE);
498	if (rc)
499		return rc;
500	trigger_paddr = apei_exec_ctx_get_output(&ctx);
501	if (notrigger == 0) {
502		rc = __einj_error_trigger(trigger_paddr, type, param1, param2);
503		if (rc)
504			return rc;
505	}
506	rc = apei_exec_run_optional(&ctx, ACPI_EINJ_END_OPERATION);
507
508	return rc;
509}
510
511/* Inject the specified hardware error */
512static int einj_error_inject(u32 type, u64 param1, u64 param2)
 
513{
514	int rc;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
515
 
516	mutex_lock(&einj_mutex);
517	rc = __einj_error_inject(type, param1, param2);
518	mutex_unlock(&einj_mutex);
519
520	return rc;
521}
522
523static u32 error_type;
 
524static u64 error_param1;
525static u64 error_param2;
 
 
526static struct dentry *einj_debug_dir;
527
528static int available_error_type_show(struct seq_file *m, void *v)
529{
530	int rc;
531	u32 available_error_type = 0;
532
533	rc = einj_get_available_error_type(&available_error_type);
534	if (rc)
535		return rc;
536	if (available_error_type & 0x0001)
537		seq_printf(m, "0x00000001\tProcessor Correctable\n");
538	if (available_error_type & 0x0002)
539		seq_printf(m, "0x00000002\tProcessor Uncorrectable non-fatal\n");
540	if (available_error_type & 0x0004)
541		seq_printf(m, "0x00000004\tProcessor Uncorrectable fatal\n");
542	if (available_error_type & 0x0008)
543		seq_printf(m, "0x00000008\tMemory Correctable\n");
544	if (available_error_type & 0x0010)
545		seq_printf(m, "0x00000010\tMemory Uncorrectable non-fatal\n");
546	if (available_error_type & 0x0020)
547		seq_printf(m, "0x00000020\tMemory Uncorrectable fatal\n");
548	if (available_error_type & 0x0040)
549		seq_printf(m, "0x00000040\tPCI Express Correctable\n");
550	if (available_error_type & 0x0080)
551		seq_printf(m, "0x00000080\tPCI Express Uncorrectable non-fatal\n");
552	if (available_error_type & 0x0100)
553		seq_printf(m, "0x00000100\tPCI Express Uncorrectable fatal\n");
554	if (available_error_type & 0x0200)
555		seq_printf(m, "0x00000200\tPlatform Correctable\n");
556	if (available_error_type & 0x0400)
557		seq_printf(m, "0x00000400\tPlatform Uncorrectable non-fatal\n");
558	if (available_error_type & 0x0800)
559		seq_printf(m, "0x00000800\tPlatform Uncorrectable fatal\n");
560
561	return 0;
562}
563
564static int available_error_type_open(struct inode *inode, struct file *file)
565{
566	return single_open(file, available_error_type_show, NULL);
567}
568
569static const struct file_operations available_error_type_fops = {
570	.open		= available_error_type_open,
571	.read		= seq_read,
572	.llseek		= seq_lseek,
573	.release	= single_release,
574};
575
576static int error_type_get(void *data, u64 *val)
577{
578	*val = error_type;
579
580	return 0;
581}
582
583static int error_type_set(void *data, u64 val)
584{
585	int rc;
586	u32 available_error_type = 0;
587	u32 tval, vendor;
588
589	/*
590	 * Vendor defined types have 0x80000000 bit set, and
591	 * are not enumerated by ACPI_EINJ_GET_ERROR_TYPE
592	 */
593	vendor = val & 0x80000000;
594	tval = val & 0x7fffffff;
595
596	/* Only one error type can be specified */
597	if (tval & (tval - 1))
598		return -EINVAL;
599	if (!vendor) {
600		rc = einj_get_available_error_type(&available_error_type);
601		if (rc)
602			return rc;
603		if (!(val & available_error_type))
604			return -EINVAL;
605	}
606	error_type = val;
607
608	return 0;
609}
610
611DEFINE_SIMPLE_ATTRIBUTE(error_type_fops, error_type_get,
612			error_type_set, "0x%llx\n");
613
614static int error_inject_set(void *data, u64 val)
615{
616	if (!error_type)
617		return -EINVAL;
618
619	return einj_error_inject(error_type, error_param1, error_param2);
 
620}
621
622DEFINE_SIMPLE_ATTRIBUTE(error_inject_fops, NULL,
623			error_inject_set, "%llu\n");
624
625static int einj_check_table(struct acpi_table_einj *einj_tab)
626{
627	if ((einj_tab->header_length !=
628	     (sizeof(struct acpi_table_einj) - sizeof(einj_tab->header)))
629	    && (einj_tab->header_length != sizeof(struct acpi_table_einj)))
630		return -EINVAL;
631	if (einj_tab->header.length < sizeof(struct acpi_table_einj))
632		return -EINVAL;
633	if (einj_tab->entries !=
634	    (einj_tab->header.length - sizeof(struct acpi_table_einj)) /
635	    sizeof(struct acpi_einj_entry))
636		return -EINVAL;
637
638	return 0;
639}
640
641static int __init einj_init(void)
642{
643	int rc;
644	acpi_status status;
645	struct dentry *fentry;
646	struct apei_exec_context ctx;
647
648	if (acpi_disabled)
649		return -ENODEV;
650
651	status = acpi_get_table(ACPI_SIG_EINJ, 0,
652				(struct acpi_table_header **)&einj_tab);
653	if (status == AE_NOT_FOUND)
654		return -ENODEV;
655	else if (ACPI_FAILURE(status)) {
656		const char *msg = acpi_format_exception(status);
657		pr_err(EINJ_PFX "Failed to get table, %s\n", msg);
658		return -EINVAL;
659	}
660
661	rc = einj_check_table(einj_tab);
662	if (rc) {
663		pr_warning(FW_BUG EINJ_PFX "EINJ table is invalid\n");
664		return -EINVAL;
665	}
666
667	rc = -ENOMEM;
668	einj_debug_dir = debugfs_create_dir("einj", apei_get_debugfs_dir());
669	if (!einj_debug_dir)
670		goto err_cleanup;
671	fentry = debugfs_create_file("available_error_type", S_IRUSR,
672				     einj_debug_dir, NULL,
673				     &available_error_type_fops);
674	if (!fentry)
675		goto err_cleanup;
676	fentry = debugfs_create_file("error_type", S_IRUSR | S_IWUSR,
677				     einj_debug_dir, NULL, &error_type_fops);
678	if (!fentry)
679		goto err_cleanup;
680	fentry = debugfs_create_file("error_inject", S_IWUSR,
681				     einj_debug_dir, NULL, &error_inject_fops);
682	if (!fentry)
683		goto err_cleanup;
684
685	apei_resources_init(&einj_resources);
686	einj_exec_ctx_init(&ctx);
687	rc = apei_exec_collect_resources(&ctx, &einj_resources);
688	if (rc)
689		goto err_fini;
690	rc = apei_resources_request(&einj_resources, "APEI EINJ");
691	if (rc)
692		goto err_fini;
693	rc = apei_exec_pre_map_gars(&ctx);
694	if (rc)
695		goto err_release;
696
 
697	einj_param = einj_get_parameter_address();
698	if ((param_extension || acpi5) && einj_param) {
 
 
 
 
699		fentry = debugfs_create_x64("param1", S_IRUSR | S_IWUSR,
700					    einj_debug_dir, &error_param1);
701		if (!fentry)
702			goto err_unmap;
703		fentry = debugfs_create_x64("param2", S_IRUSR | S_IWUSR,
704					    einj_debug_dir, &error_param2);
705		if (!fentry)
706			goto err_unmap;
 
 
 
 
 
 
 
 
707
708		fentry = debugfs_create_x32("notrigger", S_IRUSR | S_IWUSR,
709					    einj_debug_dir, &notrigger);
710		if (!fentry)
711			goto err_unmap;
712	}
713
714	if (vendor_dev[0]) {
715		vendor_blob.data = vendor_dev;
716		vendor_blob.size = strlen(vendor_dev);
717		fentry = debugfs_create_blob("vendor", S_IRUSR,
718					     einj_debug_dir, &vendor_blob);
719		if (!fentry)
720			goto err_unmap;
721		fentry = debugfs_create_x32("vendor_flags", S_IRUSR | S_IWUSR,
722					    einj_debug_dir, &vendor_flags);
723		if (!fentry)
724			goto err_unmap;
725	}
726
727	pr_info(EINJ_PFX "Error INJection is initialized.\n");
728
729	return 0;
730
731err_unmap:
732	if (einj_param) {
733		acpi_size size = (acpi5) ?
734			sizeof(struct set_error_type_with_address) :
735			sizeof(struct einj_parameter);
736
737		acpi_os_unmap_memory(einj_param, size);
738	}
739	apei_exec_post_unmap_gars(&ctx);
740err_release:
741	apei_resources_release(&einj_resources);
742err_fini:
743	apei_resources_fini(&einj_resources);
744err_cleanup:
745	debugfs_remove_recursive(einj_debug_dir);
746
747	return rc;
748}
749
750static void __exit einj_exit(void)
751{
752	struct apei_exec_context ctx;
753
754	if (einj_param) {
755		acpi_size size = (acpi5) ?
756			sizeof(struct set_error_type_with_address) :
757			sizeof(struct einj_parameter);
758
759		acpi_os_unmap_memory(einj_param, size);
760	}
761	einj_exec_ctx_init(&ctx);
762	apei_exec_post_unmap_gars(&ctx);
763	apei_resources_release(&einj_resources);
764	apei_resources_fini(&einj_resources);
765	debugfs_remove_recursive(einj_debug_dir);
766}
767
768module_init(einj_init);
769module_exit(einj_exit);
770
771MODULE_AUTHOR("Huang Ying");
772MODULE_DESCRIPTION("APEI Error INJection support");
773MODULE_LICENSE("GPL");
v4.6
  1/*
  2 * APEI Error INJection support
  3 *
  4 * EINJ provides a hardware error injection mechanism, this is useful
  5 * for debugging and testing of other APEI and RAS features.
  6 *
  7 * For more information about EINJ, please refer to ACPI Specification
  8 * version 4.0, section 17.5.
  9 *
 10 * Copyright 2009-2010 Intel Corp.
 11 *   Author: Huang Ying <ying.huang@intel.com>
 12 *
 13 * This program is free software; you can redistribute it and/or
 14 * modify it under the terms of the GNU General Public License version
 15 * 2 as published by the Free Software Foundation.
 16 *
 17 * This program is distributed in the hope that it will be useful,
 18 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 20 * GNU General Public License for more details.
 
 
 
 
 21 */
 22
 23#include <linux/kernel.h>
 24#include <linux/module.h>
 25#include <linux/init.h>
 26#include <linux/io.h>
 27#include <linux/debugfs.h>
 28#include <linux/seq_file.h>
 29#include <linux/nmi.h>
 30#include <linux/delay.h>
 31#include <linux/mm.h>
 32#include <asm/unaligned.h>
 33
 34#include "apei-internal.h"
 35
 36#define EINJ_PFX "EINJ: "
 37
 38#define SPIN_UNIT		100			/* 100ns */
 39/* Firmware should respond within 1 milliseconds */
 40#define FIRMWARE_TIMEOUT	(1 * NSEC_PER_MSEC)
 41#define ACPI5_VENDOR_BIT	BIT(31)
 42#define MEM_ERROR_MASK		(ACPI_EINJ_MEMORY_CORRECTABLE | \
 43				ACPI_EINJ_MEMORY_UNCORRECTABLE | \
 44				ACPI_EINJ_MEMORY_FATAL)
 45
 46/*
 47 * ACPI version 5 provides a SET_ERROR_TYPE_WITH_ADDRESS action.
 48 */
 49static int acpi5;
 50
 51struct set_error_type_with_address {
 52	u32	type;
 53	u32	vendor_extension;
 54	u32	flags;
 55	u32	apicid;
 56	u64	memory_address;
 57	u64	memory_address_range;
 58	u32	pcie_sbdf;
 59};
 60enum {
 61	SETWA_FLAGS_APICID = 1,
 62	SETWA_FLAGS_MEM = 2,
 63	SETWA_FLAGS_PCIE_SBDF = 4,
 64};
 65
 66/*
 67 * Vendor extensions for platform specific operations
 68 */
 69struct vendor_error_type_extension {
 70	u32	length;
 71	u32	pcie_sbdf;
 72	u16	vendor_id;
 73	u16	device_id;
 74	u8	rev_id;
 75	u8	reserved[3];
 76};
 77
 78static u32 notrigger;
 79
 80static u32 vendor_flags;
 81static struct debugfs_blob_wrapper vendor_blob;
 82static char vendor_dev[64];
 83
 84/*
 85 * Some BIOSes allow parameters to the SET_ERROR_TYPE entries in the
 86 * EINJ table through an unpublished extension. Use with caution as
 87 * most will ignore the parameter and make their own choice of address
 88 * for error injection.  This extension is used only if
 89 * param_extension module parameter is specified.
 90 */
 91struct einj_parameter {
 92	u64 type;
 93	u64 reserved1;
 94	u64 reserved2;
 95	u64 param1;
 96	u64 param2;
 97};
 98
 99#define EINJ_OP_BUSY			0x1
100#define EINJ_STATUS_SUCCESS		0x0
101#define EINJ_STATUS_FAIL		0x1
102#define EINJ_STATUS_INVAL		0x2
103
104#define EINJ_TAB_ENTRY(tab)						\
105	((struct acpi_whea_header *)((char *)(tab) +			\
106				    sizeof(struct acpi_table_einj)))
107
108static bool param_extension;
109module_param(param_extension, bool, 0);
110
111static struct acpi_table_einj *einj_tab;
112
113static struct apei_resources einj_resources;
114
115static struct apei_exec_ins_type einj_ins_type[] = {
116	[ACPI_EINJ_READ_REGISTER] = {
117		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
118		.run   = apei_exec_read_register,
119	},
120	[ACPI_EINJ_READ_REGISTER_VALUE] = {
121		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
122		.run   = apei_exec_read_register_value,
123	},
124	[ACPI_EINJ_WRITE_REGISTER] = {
125		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
126		.run   = apei_exec_write_register,
127	},
128	[ACPI_EINJ_WRITE_REGISTER_VALUE] = {
129		.flags = APEI_EXEC_INS_ACCESS_REGISTER,
130		.run   = apei_exec_write_register_value,
131	},
132	[ACPI_EINJ_NOOP] = {
133		.flags = 0,
134		.run   = apei_exec_noop,
135	},
136};
137
138/*
139 * Prevent EINJ interpreter to run simultaneously, because the
140 * corresponding firmware implementation may not work properly when
141 * invoked simultaneously.
142 */
143static DEFINE_MUTEX(einj_mutex);
144
145static void *einj_param;
146
147static void einj_exec_ctx_init(struct apei_exec_context *ctx)
148{
149	apei_exec_ctx_init(ctx, einj_ins_type, ARRAY_SIZE(einj_ins_type),
150			   EINJ_TAB_ENTRY(einj_tab), einj_tab->entries);
151}
152
153static int __einj_get_available_error_type(u32 *type)
154{
155	struct apei_exec_context ctx;
156	int rc;
157
158	einj_exec_ctx_init(&ctx);
159	rc = apei_exec_run(&ctx, ACPI_EINJ_GET_ERROR_TYPE);
160	if (rc)
161		return rc;
162	*type = apei_exec_ctx_get_output(&ctx);
163
164	return 0;
165}
166
167/* Get error injection capabilities of the platform */
168static int einj_get_available_error_type(u32 *type)
169{
170	int rc;
171
172	mutex_lock(&einj_mutex);
173	rc = __einj_get_available_error_type(type);
174	mutex_unlock(&einj_mutex);
175
176	return rc;
177}
178
179static int einj_timedout(u64 *t)
180{
181	if ((s64)*t < SPIN_UNIT) {
182		pr_warning(FW_WARN EINJ_PFX
183			   "Firmware does not respond in time\n");
184		return 1;
185	}
186	*t -= SPIN_UNIT;
187	ndelay(SPIN_UNIT);
188	touch_nmi_watchdog();
189	return 0;
190}
191
192static void check_vendor_extension(u64 paddr,
193				   struct set_error_type_with_address *v5param)
194{
195	int	offset = v5param->vendor_extension;
196	struct	vendor_error_type_extension *v;
197	u32	sbdf;
198
199	if (!offset)
200		return;
201	v = acpi_os_map_iomem(paddr + offset, sizeof(*v));
202	if (!v)
203		return;
204	sbdf = v->pcie_sbdf;
205	sprintf(vendor_dev, "%x:%x:%x.%x vendor_id=%x device_id=%x rev_id=%x\n",
206		sbdf >> 24, (sbdf >> 16) & 0xff,
207		(sbdf >> 11) & 0x1f, (sbdf >> 8) & 0x7,
208		 v->vendor_id, v->device_id, v->rev_id);
209	acpi_os_unmap_iomem(v, sizeof(*v));
210}
211
212static void *einj_get_parameter_address(void)
213{
214	int i;
215	u64 pa_v4 = 0, pa_v5 = 0;
216	struct acpi_whea_header *entry;
217
218	entry = EINJ_TAB_ENTRY(einj_tab);
219	for (i = 0; i < einj_tab->entries; i++) {
220		if (entry->action == ACPI_EINJ_SET_ERROR_TYPE &&
221		    entry->instruction == ACPI_EINJ_WRITE_REGISTER &&
222		    entry->register_region.space_id ==
223		    ACPI_ADR_SPACE_SYSTEM_MEMORY)
224			pa_v4 = get_unaligned(&entry->register_region.address);
 
225		if (entry->action == ACPI_EINJ_SET_ERROR_TYPE_WITH_ADDRESS &&
226		    entry->instruction == ACPI_EINJ_WRITE_REGISTER &&
227		    entry->register_region.space_id ==
228		    ACPI_ADR_SPACE_SYSTEM_MEMORY)
229			pa_v5 = get_unaligned(&entry->register_region.address);
 
230		entry++;
231	}
232	if (pa_v5) {
233		struct set_error_type_with_address *v5param;
234
235		v5param = acpi_os_map_iomem(pa_v5, sizeof(*v5param));
236		if (v5param) {
237			acpi5 = 1;
238			check_vendor_extension(pa_v5, v5param);
239			return v5param;
240		}
241	}
242	if (param_extension && pa_v4) {
243		struct einj_parameter *v4param;
244
245		v4param = acpi_os_map_iomem(pa_v4, sizeof(*v4param));
246		if (!v4param)
247			return NULL;
248		if (v4param->reserved1 || v4param->reserved2) {
249			acpi_os_unmap_iomem(v4param, sizeof(*v4param));
250			return NULL;
251		}
252		return v4param;
253	}
254
255	return NULL;
256}
257
258/* do sanity check to trigger table */
259static int einj_check_trigger_header(struct acpi_einj_trigger *trigger_tab)
260{
261	if (trigger_tab->header_size != sizeof(struct acpi_einj_trigger))
262		return -EINVAL;
263	if (trigger_tab->table_size > PAGE_SIZE ||
264	    trigger_tab->table_size < trigger_tab->header_size)
265		return -EINVAL;
266	if (trigger_tab->entry_count !=
267	    (trigger_tab->table_size - trigger_tab->header_size) /
268	    sizeof(struct acpi_einj_entry))
269		return -EINVAL;
270
271	return 0;
272}
273
274static struct acpi_generic_address *einj_get_trigger_parameter_region(
275	struct acpi_einj_trigger *trigger_tab, u64 param1, u64 param2)
276{
277	int i;
278	struct acpi_whea_header *entry;
279
280	entry = (struct acpi_whea_header *)
281		((char *)trigger_tab + sizeof(struct acpi_einj_trigger));
282	for (i = 0; i < trigger_tab->entry_count; i++) {
283		if (entry->action == ACPI_EINJ_TRIGGER_ERROR &&
284		entry->instruction == ACPI_EINJ_WRITE_REGISTER_VALUE &&
285		entry->register_region.space_id ==
286			ACPI_ADR_SPACE_SYSTEM_MEMORY &&
287		(entry->register_region.address & param2) == (param1 & param2))
288			return &entry->register_region;
289		entry++;
290	}
291
292	return NULL;
293}
294/* Execute instructions in trigger error action table */
295static int __einj_error_trigger(u64 trigger_paddr, u32 type,
296				u64 param1, u64 param2)
297{
298	struct acpi_einj_trigger *trigger_tab = NULL;
299	struct apei_exec_context trigger_ctx;
300	struct apei_resources trigger_resources;
301	struct acpi_whea_header *trigger_entry;
302	struct resource *r;
303	u32 table_size;
304	int rc = -EIO;
305	struct acpi_generic_address *trigger_param_region = NULL;
306
307	r = request_mem_region(trigger_paddr, sizeof(*trigger_tab),
308			       "APEI EINJ Trigger Table");
309	if (!r) {
310		pr_err(EINJ_PFX
311	"Can not request [mem %#010llx-%#010llx] for Trigger table\n",
312		       (unsigned long long)trigger_paddr,
313		       (unsigned long long)trigger_paddr +
314			    sizeof(*trigger_tab) - 1);
315		goto out;
316	}
317	trigger_tab = ioremap_cache(trigger_paddr, sizeof(*trigger_tab));
318	if (!trigger_tab) {
319		pr_err(EINJ_PFX "Failed to map trigger table!\n");
320		goto out_rel_header;
321	}
322	rc = einj_check_trigger_header(trigger_tab);
323	if (rc) {
324		pr_warning(FW_BUG EINJ_PFX
325			   "The trigger error action table is invalid\n");
326		goto out_rel_header;
327	}
328
329	/* No action structures in the TRIGGER_ERROR table, nothing to do */
330	if (!trigger_tab->entry_count)
331		goto out_rel_header;
332
333	rc = -EIO;
334	table_size = trigger_tab->table_size;
335	r = request_mem_region(trigger_paddr + sizeof(*trigger_tab),
336			       table_size - sizeof(*trigger_tab),
337			       "APEI EINJ Trigger Table");
338	if (!r) {
339		pr_err(EINJ_PFX
340"Can not request [mem %#010llx-%#010llx] for Trigger Table Entry\n",
341		       (unsigned long long)trigger_paddr + sizeof(*trigger_tab),
342		       (unsigned long long)trigger_paddr + table_size - 1);
343		goto out_rel_header;
344	}
345	iounmap(trigger_tab);
346	trigger_tab = ioremap_cache(trigger_paddr, table_size);
347	if (!trigger_tab) {
348		pr_err(EINJ_PFX "Failed to map trigger table!\n");
349		goto out_rel_entry;
350	}
351	trigger_entry = (struct acpi_whea_header *)
352		((char *)trigger_tab + sizeof(struct acpi_einj_trigger));
353	apei_resources_init(&trigger_resources);
354	apei_exec_ctx_init(&trigger_ctx, einj_ins_type,
355			   ARRAY_SIZE(einj_ins_type),
356			   trigger_entry, trigger_tab->entry_count);
357	rc = apei_exec_collect_resources(&trigger_ctx, &trigger_resources);
358	if (rc)
359		goto out_fini;
360	rc = apei_resources_sub(&trigger_resources, &einj_resources);
361	if (rc)
362		goto out_fini;
363	/*
364	 * Some firmware will access target address specified in
365	 * param1 to trigger the error when injecting memory error.
366	 * This will cause resource conflict with regular memory.  So
367	 * remove it from trigger table resources.
368	 */
369	if ((param_extension || acpi5) && (type & MEM_ERROR_MASK) && param2) {
370		struct apei_resources addr_resources;
371		apei_resources_init(&addr_resources);
372		trigger_param_region = einj_get_trigger_parameter_region(
373			trigger_tab, param1, param2);
374		if (trigger_param_region) {
375			rc = apei_resources_add(&addr_resources,
376				trigger_param_region->address,
377				trigger_param_region->bit_width/8, true);
378			if (rc)
379				goto out_fini;
380			rc = apei_resources_sub(&trigger_resources,
381					&addr_resources);
382		}
383		apei_resources_fini(&addr_resources);
384		if (rc)
385			goto out_fini;
386	}
387	rc = apei_resources_request(&trigger_resources, "APEI EINJ Trigger");
388	if (rc)
389		goto out_fini;
390	rc = apei_exec_pre_map_gars(&trigger_ctx);
391	if (rc)
392		goto out_release;
393
394	rc = apei_exec_run(&trigger_ctx, ACPI_EINJ_TRIGGER_ERROR);
395
396	apei_exec_post_unmap_gars(&trigger_ctx);
397out_release:
398	apei_resources_release(&trigger_resources);
399out_fini:
400	apei_resources_fini(&trigger_resources);
401out_rel_entry:
402	release_mem_region(trigger_paddr + sizeof(*trigger_tab),
403			   table_size - sizeof(*trigger_tab));
404out_rel_header:
405	release_mem_region(trigger_paddr, sizeof(*trigger_tab));
406out:
407	if (trigger_tab)
408		iounmap(trigger_tab);
409
410	return rc;
411}
412
413static int __einj_error_inject(u32 type, u32 flags, u64 param1, u64 param2,
414			       u64 param3, u64 param4)
415{
416	struct apei_exec_context ctx;
417	u64 val, trigger_paddr, timeout = FIRMWARE_TIMEOUT;
418	int rc;
419
420	einj_exec_ctx_init(&ctx);
421
422	rc = apei_exec_run_optional(&ctx, ACPI_EINJ_BEGIN_OPERATION);
423	if (rc)
424		return rc;
425	apei_exec_ctx_set_input(&ctx, type);
426	if (acpi5) {
427		struct set_error_type_with_address *v5param = einj_param;
428
429		v5param->type = type;
430		if (type & ACPI5_VENDOR_BIT) {
431			switch (vendor_flags) {
432			case SETWA_FLAGS_APICID:
433				v5param->apicid = param1;
434				break;
435			case SETWA_FLAGS_MEM:
436				v5param->memory_address = param1;
437				v5param->memory_address_range = param2;
438				break;
439			case SETWA_FLAGS_PCIE_SBDF:
440				v5param->pcie_sbdf = param1;
441				break;
442			}
443			v5param->flags = vendor_flags;
444		} else if (flags) {
445				v5param->flags = flags;
446				v5param->memory_address = param1;
447				v5param->memory_address_range = param2;
448				v5param->apicid = param3;
449				v5param->pcie_sbdf = param4;
450		} else {
451			switch (type) {
452			case ACPI_EINJ_PROCESSOR_CORRECTABLE:
453			case ACPI_EINJ_PROCESSOR_UNCORRECTABLE:
454			case ACPI_EINJ_PROCESSOR_FATAL:
455				v5param->apicid = param1;
456				v5param->flags = SETWA_FLAGS_APICID;
457				break;
458			case ACPI_EINJ_MEMORY_CORRECTABLE:
459			case ACPI_EINJ_MEMORY_UNCORRECTABLE:
460			case ACPI_EINJ_MEMORY_FATAL:
461				v5param->memory_address = param1;
462				v5param->memory_address_range = param2;
463				v5param->flags = SETWA_FLAGS_MEM;
464				break;
465			case ACPI_EINJ_PCIX_CORRECTABLE:
466			case ACPI_EINJ_PCIX_UNCORRECTABLE:
467			case ACPI_EINJ_PCIX_FATAL:
468				v5param->pcie_sbdf = param1;
469				v5param->flags = SETWA_FLAGS_PCIE_SBDF;
470				break;
471			}
472		}
473	} else {
474		rc = apei_exec_run(&ctx, ACPI_EINJ_SET_ERROR_TYPE);
475		if (rc)
476			return rc;
477		if (einj_param) {
478			struct einj_parameter *v4param = einj_param;
479			v4param->param1 = param1;
480			v4param->param2 = param2;
481		}
482	}
483	rc = apei_exec_run(&ctx, ACPI_EINJ_EXECUTE_OPERATION);
484	if (rc)
485		return rc;
486	for (;;) {
487		rc = apei_exec_run(&ctx, ACPI_EINJ_CHECK_BUSY_STATUS);
488		if (rc)
489			return rc;
490		val = apei_exec_ctx_get_output(&ctx);
491		if (!(val & EINJ_OP_BUSY))
492			break;
493		if (einj_timedout(&timeout))
494			return -EIO;
495	}
496	rc = apei_exec_run(&ctx, ACPI_EINJ_GET_COMMAND_STATUS);
497	if (rc)
498		return rc;
499	val = apei_exec_ctx_get_output(&ctx);
500	if (val != EINJ_STATUS_SUCCESS)
501		return -EBUSY;
502
503	rc = apei_exec_run(&ctx, ACPI_EINJ_GET_TRIGGER_TABLE);
504	if (rc)
505		return rc;
506	trigger_paddr = apei_exec_ctx_get_output(&ctx);
507	if (notrigger == 0) {
508		rc = __einj_error_trigger(trigger_paddr, type, param1, param2);
509		if (rc)
510			return rc;
511	}
512	rc = apei_exec_run_optional(&ctx, ACPI_EINJ_END_OPERATION);
513
514	return rc;
515}
516
517/* Inject the specified hardware error */
518static int einj_error_inject(u32 type, u32 flags, u64 param1, u64 param2,
519			     u64 param3, u64 param4)
520{
521	int rc;
522	u64 base_addr, size;
523
524	/* If user manually set "flags", make sure it is legal */
525	if (flags && (flags &
526		~(SETWA_FLAGS_APICID|SETWA_FLAGS_MEM|SETWA_FLAGS_PCIE_SBDF)))
527		return -EINVAL;
528
529	/*
530	 * We need extra sanity checks for memory errors.
531	 * Other types leap directly to injection.
532	 */
533
534	/* ensure param1/param2 existed */
535	if (!(param_extension || acpi5))
536		goto inject;
537
538	/* ensure injection is memory related */
539	if (type & ACPI5_VENDOR_BIT) {
540		if (vendor_flags != SETWA_FLAGS_MEM)
541			goto inject;
542	} else if (!(type & MEM_ERROR_MASK) && !(flags & SETWA_FLAGS_MEM))
543		goto inject;
544
545	/*
546	 * Disallow crazy address masks that give BIOS leeway to pick
547	 * injection address almost anywhere. Insist on page or
548	 * better granularity and that target address is normal RAM or
549	 * NVDIMM.
550	 */
551	base_addr = param1 & param2;
552	size = ~param2 + 1;
553
554	if (((param2 & PAGE_MASK) != PAGE_MASK) ||
555	    ((region_intersects(base_addr, size, IORESOURCE_SYSTEM_RAM, IORES_DESC_NONE)
556				!= REGION_INTERSECTS) &&
557	     (region_intersects(base_addr, size, IORESOURCE_MEM, IORES_DESC_PERSISTENT_MEMORY)
558				!= REGION_INTERSECTS)))
559		return -EINVAL;
560
561inject:
562	mutex_lock(&einj_mutex);
563	rc = __einj_error_inject(type, flags, param1, param2, param3, param4);
564	mutex_unlock(&einj_mutex);
565
566	return rc;
567}
568
569static u32 error_type;
570static u32 error_flags;
571static u64 error_param1;
572static u64 error_param2;
573static u64 error_param3;
574static u64 error_param4;
575static struct dentry *einj_debug_dir;
576
577static int available_error_type_show(struct seq_file *m, void *v)
578{
579	int rc;
580	u32 available_error_type = 0;
581
582	rc = einj_get_available_error_type(&available_error_type);
583	if (rc)
584		return rc;
585	if (available_error_type & 0x0001)
586		seq_printf(m, "0x00000001\tProcessor Correctable\n");
587	if (available_error_type & 0x0002)
588		seq_printf(m, "0x00000002\tProcessor Uncorrectable non-fatal\n");
589	if (available_error_type & 0x0004)
590		seq_printf(m, "0x00000004\tProcessor Uncorrectable fatal\n");
591	if (available_error_type & 0x0008)
592		seq_printf(m, "0x00000008\tMemory Correctable\n");
593	if (available_error_type & 0x0010)
594		seq_printf(m, "0x00000010\tMemory Uncorrectable non-fatal\n");
595	if (available_error_type & 0x0020)
596		seq_printf(m, "0x00000020\tMemory Uncorrectable fatal\n");
597	if (available_error_type & 0x0040)
598		seq_printf(m, "0x00000040\tPCI Express Correctable\n");
599	if (available_error_type & 0x0080)
600		seq_printf(m, "0x00000080\tPCI Express Uncorrectable non-fatal\n");
601	if (available_error_type & 0x0100)
602		seq_printf(m, "0x00000100\tPCI Express Uncorrectable fatal\n");
603	if (available_error_type & 0x0200)
604		seq_printf(m, "0x00000200\tPlatform Correctable\n");
605	if (available_error_type & 0x0400)
606		seq_printf(m, "0x00000400\tPlatform Uncorrectable non-fatal\n");
607	if (available_error_type & 0x0800)
608		seq_printf(m, "0x00000800\tPlatform Uncorrectable fatal\n");
609
610	return 0;
611}
612
613static int available_error_type_open(struct inode *inode, struct file *file)
614{
615	return single_open(file, available_error_type_show, NULL);
616}
617
618static const struct file_operations available_error_type_fops = {
619	.open		= available_error_type_open,
620	.read		= seq_read,
621	.llseek		= seq_lseek,
622	.release	= single_release,
623};
624
625static int error_type_get(void *data, u64 *val)
626{
627	*val = error_type;
628
629	return 0;
630}
631
632static int error_type_set(void *data, u64 val)
633{
634	int rc;
635	u32 available_error_type = 0;
636	u32 tval, vendor;
637
638	/*
639	 * Vendor defined types have 0x80000000 bit set, and
640	 * are not enumerated by ACPI_EINJ_GET_ERROR_TYPE
641	 */
642	vendor = val & ACPI5_VENDOR_BIT;
643	tval = val & 0x7fffffff;
644
645	/* Only one error type can be specified */
646	if (tval & (tval - 1))
647		return -EINVAL;
648	if (!vendor) {
649		rc = einj_get_available_error_type(&available_error_type);
650		if (rc)
651			return rc;
652		if (!(val & available_error_type))
653			return -EINVAL;
654	}
655	error_type = val;
656
657	return 0;
658}
659
660DEFINE_SIMPLE_ATTRIBUTE(error_type_fops, error_type_get,
661			error_type_set, "0x%llx\n");
662
663static int error_inject_set(void *data, u64 val)
664{
665	if (!error_type)
666		return -EINVAL;
667
668	return einj_error_inject(error_type, error_flags, error_param1, error_param2,
669		error_param3, error_param4);
670}
671
672DEFINE_SIMPLE_ATTRIBUTE(error_inject_fops, NULL,
673			error_inject_set, "%llu\n");
674
675static int einj_check_table(struct acpi_table_einj *einj_tab)
676{
677	if ((einj_tab->header_length !=
678	     (sizeof(struct acpi_table_einj) - sizeof(einj_tab->header)))
679	    && (einj_tab->header_length != sizeof(struct acpi_table_einj)))
680		return -EINVAL;
681	if (einj_tab->header.length < sizeof(struct acpi_table_einj))
682		return -EINVAL;
683	if (einj_tab->entries !=
684	    (einj_tab->header.length - sizeof(struct acpi_table_einj)) /
685	    sizeof(struct acpi_einj_entry))
686		return -EINVAL;
687
688	return 0;
689}
690
691static int __init einj_init(void)
692{
693	int rc;
694	acpi_status status;
695	struct dentry *fentry;
696	struct apei_exec_context ctx;
697
698	if (acpi_disabled)
699		return -ENODEV;
700
701	status = acpi_get_table(ACPI_SIG_EINJ, 0,
702				(struct acpi_table_header **)&einj_tab);
703	if (status == AE_NOT_FOUND)
704		return -ENODEV;
705	else if (ACPI_FAILURE(status)) {
706		const char *msg = acpi_format_exception(status);
707		pr_err(EINJ_PFX "Failed to get table, %s\n", msg);
708		return -EINVAL;
709	}
710
711	rc = einj_check_table(einj_tab);
712	if (rc) {
713		pr_warning(FW_BUG EINJ_PFX "EINJ table is invalid\n");
714		return -EINVAL;
715	}
716
717	rc = -ENOMEM;
718	einj_debug_dir = debugfs_create_dir("einj", apei_get_debugfs_dir());
719	if (!einj_debug_dir)
720		goto err_cleanup;
721	fentry = debugfs_create_file("available_error_type", S_IRUSR,
722				     einj_debug_dir, NULL,
723				     &available_error_type_fops);
724	if (!fentry)
725		goto err_cleanup;
726	fentry = debugfs_create_file("error_type", S_IRUSR | S_IWUSR,
727				     einj_debug_dir, NULL, &error_type_fops);
728	if (!fentry)
729		goto err_cleanup;
730	fentry = debugfs_create_file("error_inject", S_IWUSR,
731				     einj_debug_dir, NULL, &error_inject_fops);
732	if (!fentry)
733		goto err_cleanup;
734
735	apei_resources_init(&einj_resources);
736	einj_exec_ctx_init(&ctx);
737	rc = apei_exec_collect_resources(&ctx, &einj_resources);
738	if (rc)
739		goto err_fini;
740	rc = apei_resources_request(&einj_resources, "APEI EINJ");
741	if (rc)
742		goto err_fini;
743	rc = apei_exec_pre_map_gars(&ctx);
744	if (rc)
745		goto err_release;
746
747	rc = -ENOMEM;
748	einj_param = einj_get_parameter_address();
749	if ((param_extension || acpi5) && einj_param) {
750		fentry = debugfs_create_x32("flags", S_IRUSR | S_IWUSR,
751					    einj_debug_dir, &error_flags);
752		if (!fentry)
753			goto err_unmap;
754		fentry = debugfs_create_x64("param1", S_IRUSR | S_IWUSR,
755					    einj_debug_dir, &error_param1);
756		if (!fentry)
757			goto err_unmap;
758		fentry = debugfs_create_x64("param2", S_IRUSR | S_IWUSR,
759					    einj_debug_dir, &error_param2);
760		if (!fentry)
761			goto err_unmap;
762		fentry = debugfs_create_x64("param3", S_IRUSR | S_IWUSR,
763					    einj_debug_dir, &error_param3);
764		if (!fentry)
765			goto err_unmap;
766		fentry = debugfs_create_x64("param4", S_IRUSR | S_IWUSR,
767					    einj_debug_dir, &error_param4);
768		if (!fentry)
769			goto err_unmap;
770
771		fentry = debugfs_create_x32("notrigger", S_IRUSR | S_IWUSR,
772					    einj_debug_dir, &notrigger);
773		if (!fentry)
774			goto err_unmap;
775	}
776
777	if (vendor_dev[0]) {
778		vendor_blob.data = vendor_dev;
779		vendor_blob.size = strlen(vendor_dev);
780		fentry = debugfs_create_blob("vendor", S_IRUSR,
781					     einj_debug_dir, &vendor_blob);
782		if (!fentry)
783			goto err_unmap;
784		fentry = debugfs_create_x32("vendor_flags", S_IRUSR | S_IWUSR,
785					    einj_debug_dir, &vendor_flags);
786		if (!fentry)
787			goto err_unmap;
788	}
789
790	pr_info(EINJ_PFX "Error INJection is initialized.\n");
791
792	return 0;
793
794err_unmap:
795	if (einj_param) {
796		acpi_size size = (acpi5) ?
797			sizeof(struct set_error_type_with_address) :
798			sizeof(struct einj_parameter);
799
800		acpi_os_unmap_iomem(einj_param, size);
801	}
802	apei_exec_post_unmap_gars(&ctx);
803err_release:
804	apei_resources_release(&einj_resources);
805err_fini:
806	apei_resources_fini(&einj_resources);
807err_cleanup:
808	debugfs_remove_recursive(einj_debug_dir);
809
810	return rc;
811}
812
813static void __exit einj_exit(void)
814{
815	struct apei_exec_context ctx;
816
817	if (einj_param) {
818		acpi_size size = (acpi5) ?
819			sizeof(struct set_error_type_with_address) :
820			sizeof(struct einj_parameter);
821
822		acpi_os_unmap_iomem(einj_param, size);
823	}
824	einj_exec_ctx_init(&ctx);
825	apei_exec_post_unmap_gars(&ctx);
826	apei_resources_release(&einj_resources);
827	apei_resources_fini(&einj_resources);
828	debugfs_remove_recursive(einj_debug_dir);
829}
830
831module_init(einj_init);
832module_exit(einj_exit);
833
834MODULE_AUTHOR("Huang Ying");
835MODULE_DESCRIPTION("APEI Error INJection support");
836MODULE_LICENSE("GPL");