Linux Audio

Check our new training course

Yocto distribution development and maintenance

Need a Yocto distribution for your embedded project?
Loading...
v6.2
  1/* SPDX-License-Identifier: GPL-2.0 */
  2#if !defined(_TRACE_KVM_MAIN_H) || defined(TRACE_HEADER_MULTI_READ)
  3#define _TRACE_KVM_MAIN_H
  4
  5#include <linux/tracepoint.h>
  6
  7#undef TRACE_SYSTEM
  8#define TRACE_SYSTEM kvm
  9
 10#define ERSN(x) { KVM_EXIT_##x, "KVM_EXIT_" #x }
 11
 12#define kvm_trace_exit_reason						\
 13	ERSN(UNKNOWN), ERSN(EXCEPTION), ERSN(IO), ERSN(HYPERCALL),	\
 14	ERSN(DEBUG), ERSN(HLT), ERSN(MMIO), ERSN(IRQ_WINDOW_OPEN),	\
 15	ERSN(SHUTDOWN), ERSN(FAIL_ENTRY), ERSN(INTR), ERSN(SET_TPR),	\
 16	ERSN(TPR_ACCESS), ERSN(S390_SIEIC), ERSN(S390_RESET), ERSN(DCR),\
 17	ERSN(NMI), ERSN(INTERNAL_ERROR), ERSN(OSI), ERSN(PAPR_HCALL),	\
 18	ERSN(S390_UCONTROL), ERSN(WATCHDOG), ERSN(S390_TSCH), ERSN(EPR),\
 19	ERSN(SYSTEM_EVENT), ERSN(S390_STSI), ERSN(IOAPIC_EOI),          \
 20	ERSN(HYPERV), ERSN(ARM_NISV), ERSN(X86_RDMSR), ERSN(X86_WRMSR)
 21
 22TRACE_EVENT(kvm_userspace_exit,
 23	    TP_PROTO(__u32 reason, int errno),
 24	    TP_ARGS(reason, errno),
 25
 26	TP_STRUCT__entry(
 27		__field(	__u32,		reason		)
 28		__field(	int,		errno		)
 29	),
 30
 31	TP_fast_assign(
 32		__entry->reason		= reason;
 33		__entry->errno		= errno;
 34	),
 35
 36	TP_printk("reason %s (%d)",
 37		  __entry->errno < 0 ?
 38		  (__entry->errno == -EINTR ? "restart" : "error") :
 39		  __print_symbolic(__entry->reason, kvm_trace_exit_reason),
 40		  __entry->errno < 0 ? -__entry->errno : __entry->reason)
 41);
 42
 43TRACE_EVENT(kvm_vcpu_wakeup,
 44	    TP_PROTO(__u64 ns, bool waited, bool valid),
 45	    TP_ARGS(ns, waited, valid),
 46
 47	TP_STRUCT__entry(
 48		__field(	__u64,		ns		)
 49		__field(	bool,		waited		)
 50		__field(	bool,		valid		)
 51	),
 52
 53	TP_fast_assign(
 54		__entry->ns		= ns;
 55		__entry->waited		= waited;
 56		__entry->valid		= valid;
 57	),
 58
 59	TP_printk("%s time %lld ns, polling %s",
 60		  __entry->waited ? "wait" : "poll",
 61		  __entry->ns,
 62		  __entry->valid ? "valid" : "invalid")
 63);
 64
 65#if defined(CONFIG_HAVE_KVM_IRQFD)
 66TRACE_EVENT(kvm_set_irq,
 67	TP_PROTO(unsigned int gsi, int level, int irq_source_id),
 68	TP_ARGS(gsi, level, irq_source_id),
 69
 70	TP_STRUCT__entry(
 71		__field(	unsigned int,	gsi		)
 72		__field(	int,		level		)
 73		__field(	int,		irq_source_id	)
 74	),
 75
 76	TP_fast_assign(
 77		__entry->gsi		= gsi;
 78		__entry->level		= level;
 79		__entry->irq_source_id	= irq_source_id;
 80	),
 81
 82	TP_printk("gsi %u level %d source %d",
 83		  __entry->gsi, __entry->level, __entry->irq_source_id)
 84);
 85#endif /* defined(CONFIG_HAVE_KVM_IRQFD) */
 86
 87#if defined(__KVM_HAVE_IOAPIC)
 88#define kvm_deliver_mode		\
 89	{0x0, "Fixed"},			\
 90	{0x1, "LowPrio"},		\
 91	{0x2, "SMI"},			\
 92	{0x3, "Res3"},			\
 93	{0x4, "NMI"},			\
 94	{0x5, "INIT"},			\
 95	{0x6, "SIPI"},			\
 96	{0x7, "ExtINT"}
 97
 98TRACE_EVENT(kvm_ioapic_set_irq,
 99	    TP_PROTO(__u64 e, int pin, bool coalesced),
100	    TP_ARGS(e, pin, coalesced),
101
102	TP_STRUCT__entry(
103		__field(	__u64,		e		)
104		__field(	int,		pin		)
105		__field(	bool,		coalesced	)
106	),
107
108	TP_fast_assign(
109		__entry->e		= e;
110		__entry->pin		= pin;
111		__entry->coalesced	= coalesced;
112	),
113
114	TP_printk("pin %u dst %x vec %u (%s|%s|%s%s)%s",
115		  __entry->pin, (u8)(__entry->e >> 56), (u8)__entry->e,
116		  __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
117		  (__entry->e & (1<<11)) ? "logical" : "physical",
118		  (__entry->e & (1<<15)) ? "level" : "edge",
119		  (__entry->e & (1<<16)) ? "|masked" : "",
120		  __entry->coalesced ? " (coalesced)" : "")
121);
122
123TRACE_EVENT(kvm_ioapic_delayed_eoi_inj,
124	    TP_PROTO(__u64 e),
125	    TP_ARGS(e),
126
127	TP_STRUCT__entry(
128		__field(	__u64,		e		)
129	),
130
131	TP_fast_assign(
132		__entry->e		= e;
133	),
134
135	TP_printk("dst %x vec %u (%s|%s|%s%s)",
136		  (u8)(__entry->e >> 56), (u8)__entry->e,
137		  __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
138		  (__entry->e & (1<<11)) ? "logical" : "physical",
139		  (__entry->e & (1<<15)) ? "level" : "edge",
140		  (__entry->e & (1<<16)) ? "|masked" : "")
141);
142
143TRACE_EVENT(kvm_msi_set_irq,
144	    TP_PROTO(__u64 address, __u64 data),
145	    TP_ARGS(address, data),
146
147	TP_STRUCT__entry(
148		__field(	__u64,		address		)
149		__field(	__u64,		data		)
150	),
151
152	TP_fast_assign(
153		__entry->address	= address;
154		__entry->data		= data;
155	),
156
157	TP_printk("dst %llx vec %u (%s|%s|%s%s)",
158		  (u8)(__entry->address >> 12) | ((__entry->address >> 32) & 0xffffff00),
159		  (u8)__entry->data,
160		  __print_symbolic((__entry->data >> 8 & 0x7), kvm_deliver_mode),
161		  (__entry->address & (1<<2)) ? "logical" : "physical",
162		  (__entry->data & (1<<15)) ? "level" : "edge",
163		  (__entry->address & (1<<3)) ? "|rh" : "")
164);
165
166#define kvm_irqchips						\
167	{KVM_IRQCHIP_PIC_MASTER,	"PIC master"},		\
168	{KVM_IRQCHIP_PIC_SLAVE,		"PIC slave"},		\
169	{KVM_IRQCHIP_IOAPIC,		"IOAPIC"}
170
171#endif /* defined(__KVM_HAVE_IOAPIC) */
172
173#if defined(CONFIG_HAVE_KVM_IRQFD)
174
175#ifdef kvm_irqchips
176#define kvm_ack_irq_string "irqchip %s pin %u"
177#define kvm_ack_irq_parm  __print_symbolic(__entry->irqchip, kvm_irqchips), __entry->pin
178#else
179#define kvm_ack_irq_string "irqchip %d pin %u"
180#define kvm_ack_irq_parm  __entry->irqchip, __entry->pin
181#endif
182
183TRACE_EVENT(kvm_ack_irq,
184	TP_PROTO(unsigned int irqchip, unsigned int pin),
185	TP_ARGS(irqchip, pin),
186
187	TP_STRUCT__entry(
188		__field(	unsigned int,	irqchip		)
189		__field(	unsigned int,	pin		)
190	),
191
192	TP_fast_assign(
193		__entry->irqchip	= irqchip;
194		__entry->pin		= pin;
195	),
196
197	TP_printk(kvm_ack_irq_string, kvm_ack_irq_parm)
 
 
 
 
 
 
198);
199
200#endif /* defined(CONFIG_HAVE_KVM_IRQFD) */
201
202
203
204#define KVM_TRACE_MMIO_READ_UNSATISFIED 0
205#define KVM_TRACE_MMIO_READ 1
206#define KVM_TRACE_MMIO_WRITE 2
207
208#define kvm_trace_symbol_mmio \
209	{ KVM_TRACE_MMIO_READ_UNSATISFIED, "unsatisfied-read" }, \
210	{ KVM_TRACE_MMIO_READ, "read" }, \
211	{ KVM_TRACE_MMIO_WRITE, "write" }
212
213TRACE_EVENT(kvm_mmio,
214	TP_PROTO(int type, int len, u64 gpa, void *val),
215	TP_ARGS(type, len, gpa, val),
216
217	TP_STRUCT__entry(
218		__field(	u32,	type		)
219		__field(	u32,	len		)
220		__field(	u64,	gpa		)
221		__field(	u64,	val		)
222	),
223
224	TP_fast_assign(
225		__entry->type		= type;
226		__entry->len		= len;
227		__entry->gpa		= gpa;
228		__entry->val		= 0;
229		if (val)
230			memcpy(&__entry->val, val,
231			       min_t(u32, sizeof(__entry->val), len));
232	),
233
234	TP_printk("mmio %s len %u gpa 0x%llx val 0x%llx",
235		  __print_symbolic(__entry->type, kvm_trace_symbol_mmio),
236		  __entry->len, __entry->gpa, __entry->val)
237);
238
239#define kvm_fpu_load_symbol	\
240	{0, "unload"},		\
241	{1, "load"}
242
243TRACE_EVENT(kvm_fpu,
244	TP_PROTO(int load),
245	TP_ARGS(load),
246
247	TP_STRUCT__entry(
248		__field(	u32,	        load		)
249	),
250
251	TP_fast_assign(
252		__entry->load		= load;
253	),
254
255	TP_printk("%s", __print_symbolic(__entry->load, kvm_fpu_load_symbol))
256);
257
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
258#ifdef CONFIG_KVM_ASYNC_PF
259DECLARE_EVENT_CLASS(kvm_async_get_page_class,
260
261	TP_PROTO(u64 gva, u64 gfn),
262
263	TP_ARGS(gva, gfn),
264
265	TP_STRUCT__entry(
266		__field(__u64, gva)
267		__field(u64, gfn)
268	),
269
270	TP_fast_assign(
271		__entry->gva = gva;
272		__entry->gfn = gfn;
273	),
274
275	TP_printk("gva = %#llx, gfn = %#llx", __entry->gva, __entry->gfn)
276);
277
278DEFINE_EVENT(kvm_async_get_page_class, kvm_try_async_get_page,
279
280	TP_PROTO(u64 gva, u64 gfn),
281
282	TP_ARGS(gva, gfn)
283);
284
285DEFINE_EVENT(kvm_async_get_page_class, kvm_async_pf_repeated_fault,
286
287	TP_PROTO(u64 gva, u64 gfn),
288
289	TP_ARGS(gva, gfn)
290);
291
292DECLARE_EVENT_CLASS(kvm_async_pf_nopresent_ready,
293
294	TP_PROTO(u64 token, u64 gva),
295
296	TP_ARGS(token, gva),
297
298	TP_STRUCT__entry(
299		__field(__u64, token)
300		__field(__u64, gva)
301	),
302
303	TP_fast_assign(
304		__entry->token = token;
305		__entry->gva = gva;
306	),
307
308	TP_printk("token %#llx gva %#llx", __entry->token, __entry->gva)
309
310);
311
312DEFINE_EVENT(kvm_async_pf_nopresent_ready, kvm_async_pf_not_present,
313
314	TP_PROTO(u64 token, u64 gva),
315
316	TP_ARGS(token, gva)
317);
318
319DEFINE_EVENT(kvm_async_pf_nopresent_ready, kvm_async_pf_ready,
320
321	TP_PROTO(u64 token, u64 gva),
322
323	TP_ARGS(token, gva)
324);
325
326TRACE_EVENT(
327	kvm_async_pf_completed,
328	TP_PROTO(unsigned long address, u64 gva),
329	TP_ARGS(address, gva),
330
331	TP_STRUCT__entry(
332		__field(unsigned long, address)
333		__field(u64, gva)
334		),
335
336	TP_fast_assign(
337		__entry->address = address;
338		__entry->gva = gva;
339		),
340
341	TP_printk("gva %#llx address %#lx",  __entry->gva,
342		  __entry->address)
343);
344
345#endif
346
347TRACE_EVENT(kvm_halt_poll_ns,
348	TP_PROTO(bool grow, unsigned int vcpu_id, unsigned int new,
349		 unsigned int old),
350	TP_ARGS(grow, vcpu_id, new, old),
351
352	TP_STRUCT__entry(
353		__field(bool, grow)
354		__field(unsigned int, vcpu_id)
355		__field(unsigned int, new)
356		__field(unsigned int, old)
357	),
358
359	TP_fast_assign(
360		__entry->grow           = grow;
361		__entry->vcpu_id        = vcpu_id;
362		__entry->new            = new;
363		__entry->old            = old;
364	),
365
366	TP_printk("vcpu %u: halt_poll_ns %u (%s %u)",
367			__entry->vcpu_id,
368			__entry->new,
369			__entry->grow ? "grow" : "shrink",
370			__entry->old)
371);
372
373#define trace_kvm_halt_poll_ns_grow(vcpu_id, new, old) \
374	trace_kvm_halt_poll_ns(true, vcpu_id, new, old)
375#define trace_kvm_halt_poll_ns_shrink(vcpu_id, new, old) \
376	trace_kvm_halt_poll_ns(false, vcpu_id, new, old)
377
378TRACE_EVENT(kvm_dirty_ring_push,
379	TP_PROTO(struct kvm_dirty_ring *ring, u32 slot, u64 offset),
380	TP_ARGS(ring, slot, offset),
381
382	TP_STRUCT__entry(
383		__field(int, index)
384		__field(u32, dirty_index)
385		__field(u32, reset_index)
386		__field(u32, slot)
387		__field(u64, offset)
388	),
389
390	TP_fast_assign(
391		__entry->index          = ring->index;
392		__entry->dirty_index    = ring->dirty_index;
393		__entry->reset_index    = ring->reset_index;
394		__entry->slot           = slot;
395		__entry->offset         = offset;
396	),
397
398	TP_printk("ring %d: dirty 0x%x reset 0x%x "
399		  "slot %u offset 0x%llx (used %u)",
400		  __entry->index, __entry->dirty_index,
401		  __entry->reset_index,  __entry->slot, __entry->offset,
402		  __entry->dirty_index - __entry->reset_index)
403);
404
405TRACE_EVENT(kvm_dirty_ring_reset,
406	TP_PROTO(struct kvm_dirty_ring *ring),
407	TP_ARGS(ring),
408
409	TP_STRUCT__entry(
410		__field(int, index)
411		__field(u32, dirty_index)
412		__field(u32, reset_index)
413	),
414
415	TP_fast_assign(
416		__entry->index          = ring->index;
417		__entry->dirty_index    = ring->dirty_index;
418		__entry->reset_index    = ring->reset_index;
419	),
420
421	TP_printk("ring %d: dirty 0x%x reset 0x%x (used %u)",
422		  __entry->index, __entry->dirty_index, __entry->reset_index,
423		  __entry->dirty_index - __entry->reset_index)
424);
425
426TRACE_EVENT(kvm_dirty_ring_exit,
427	TP_PROTO(struct kvm_vcpu *vcpu),
428	TP_ARGS(vcpu),
429
430	TP_STRUCT__entry(
431	    __field(int, vcpu_id)
432	),
433
434	TP_fast_assign(
435	    __entry->vcpu_id = vcpu->vcpu_id;
436	),
437
438	TP_printk("vcpu %d", __entry->vcpu_id)
439);
440
441TRACE_EVENT(kvm_unmap_hva_range,
442	TP_PROTO(unsigned long start, unsigned long end),
443	TP_ARGS(start, end),
444
445	TP_STRUCT__entry(
446		__field(	unsigned long,	start		)
447		__field(	unsigned long,	end		)
448	),
449
450	TP_fast_assign(
451		__entry->start		= start;
452		__entry->end		= end;
453	),
454
455	TP_printk("mmu notifier unmap range: %#016lx -- %#016lx",
456		  __entry->start, __entry->end)
457);
458
459TRACE_EVENT(kvm_set_spte_hva,
460	TP_PROTO(unsigned long hva),
461	TP_ARGS(hva),
462
463	TP_STRUCT__entry(
464		__field(	unsigned long,	hva		)
465	),
466
467	TP_fast_assign(
468		__entry->hva		= hva;
469	),
470
471	TP_printk("mmu notifier set pte hva: %#016lx", __entry->hva)
472);
473
474TRACE_EVENT(kvm_age_hva,
475	TP_PROTO(unsigned long start, unsigned long end),
476	TP_ARGS(start, end),
477
478	TP_STRUCT__entry(
479		__field(	unsigned long,	start		)
480		__field(	unsigned long,	end		)
481	),
482
483	TP_fast_assign(
484		__entry->start		= start;
485		__entry->end		= end;
486	),
487
488	TP_printk("mmu notifier age hva: %#016lx -- %#016lx",
489		  __entry->start, __entry->end)
490);
491
492TRACE_EVENT(kvm_test_age_hva,
493	TP_PROTO(unsigned long hva),
494	TP_ARGS(hva),
495
496	TP_STRUCT__entry(
497		__field(	unsigned long,	hva		)
498	),
499
500	TP_fast_assign(
501		__entry->hva		= hva;
502	),
503
504	TP_printk("mmu notifier test age hva: %#016lx", __entry->hva)
505);
506
507#endif /* _TRACE_KVM_MAIN_H */
508
509/* This part must be outside protection */
510#include <trace/define_trace.h>
v3.15
 
  1#if !defined(_TRACE_KVM_MAIN_H) || defined(TRACE_HEADER_MULTI_READ)
  2#define _TRACE_KVM_MAIN_H
  3
  4#include <linux/tracepoint.h>
  5
  6#undef TRACE_SYSTEM
  7#define TRACE_SYSTEM kvm
  8
  9#define ERSN(x) { KVM_EXIT_##x, "KVM_EXIT_" #x }
 10
 11#define kvm_trace_exit_reason						\
 12	ERSN(UNKNOWN), ERSN(EXCEPTION), ERSN(IO), ERSN(HYPERCALL),	\
 13	ERSN(DEBUG), ERSN(HLT), ERSN(MMIO), ERSN(IRQ_WINDOW_OPEN),	\
 14	ERSN(SHUTDOWN), ERSN(FAIL_ENTRY), ERSN(INTR), ERSN(SET_TPR),	\
 15	ERSN(TPR_ACCESS), ERSN(S390_SIEIC), ERSN(S390_RESET), ERSN(DCR),\
 16	ERSN(NMI), ERSN(INTERNAL_ERROR), ERSN(OSI), ERSN(PAPR_HCALL),	\
 17	ERSN(S390_UCONTROL), ERSN(WATCHDOG), ERSN(S390_TSCH)
 
 
 18
 19TRACE_EVENT(kvm_userspace_exit,
 20	    TP_PROTO(__u32 reason, int errno),
 21	    TP_ARGS(reason, errno),
 22
 23	TP_STRUCT__entry(
 24		__field(	__u32,		reason		)
 25		__field(	int,		errno		)
 26	),
 27
 28	TP_fast_assign(
 29		__entry->reason		= reason;
 30		__entry->errno		= errno;
 31	),
 32
 33	TP_printk("reason %s (%d)",
 34		  __entry->errno < 0 ?
 35		  (__entry->errno == -EINTR ? "restart" : "error") :
 36		  __print_symbolic(__entry->reason, kvm_trace_exit_reason),
 37		  __entry->errno < 0 ? -__entry->errno : __entry->reason)
 38);
 39
 40#if defined(CONFIG_HAVE_KVM_IRQCHIP)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 41TRACE_EVENT(kvm_set_irq,
 42	TP_PROTO(unsigned int gsi, int level, int irq_source_id),
 43	TP_ARGS(gsi, level, irq_source_id),
 44
 45	TP_STRUCT__entry(
 46		__field(	unsigned int,	gsi		)
 47		__field(	int,		level		)
 48		__field(	int,		irq_source_id	)
 49	),
 50
 51	TP_fast_assign(
 52		__entry->gsi		= gsi;
 53		__entry->level		= level;
 54		__entry->irq_source_id	= irq_source_id;
 55	),
 56
 57	TP_printk("gsi %u level %d source %d",
 58		  __entry->gsi, __entry->level, __entry->irq_source_id)
 59);
 60#endif
 61
 62#if defined(__KVM_HAVE_IOAPIC)
 63#define kvm_deliver_mode		\
 64	{0x0, "Fixed"},			\
 65	{0x1, "LowPrio"},		\
 66	{0x2, "SMI"},			\
 67	{0x3, "Res3"},			\
 68	{0x4, "NMI"},			\
 69	{0x5, "INIT"},			\
 70	{0x6, "SIPI"},			\
 71	{0x7, "ExtINT"}
 72
 73TRACE_EVENT(kvm_ioapic_set_irq,
 74	    TP_PROTO(__u64 e, int pin, bool coalesced),
 75	    TP_ARGS(e, pin, coalesced),
 76
 77	TP_STRUCT__entry(
 78		__field(	__u64,		e		)
 79		__field(	int,		pin		)
 80		__field(	bool,		coalesced	)
 81	),
 82
 83	TP_fast_assign(
 84		__entry->e		= e;
 85		__entry->pin		= pin;
 86		__entry->coalesced	= coalesced;
 87	),
 88
 89	TP_printk("pin %u dst %x vec=%u (%s|%s|%s%s)%s",
 90		  __entry->pin, (u8)(__entry->e >> 56), (u8)__entry->e,
 91		  __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
 92		  (__entry->e & (1<<11)) ? "logical" : "physical",
 93		  (__entry->e & (1<<15)) ? "level" : "edge",
 94		  (__entry->e & (1<<16)) ? "|masked" : "",
 95		  __entry->coalesced ? " (coalesced)" : "")
 96);
 97
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 98TRACE_EVENT(kvm_msi_set_irq,
 99	    TP_PROTO(__u64 address, __u64 data),
100	    TP_ARGS(address, data),
101
102	TP_STRUCT__entry(
103		__field(	__u64,		address		)
104		__field(	__u64,		data		)
105	),
106
107	TP_fast_assign(
108		__entry->address	= address;
109		__entry->data		= data;
110	),
111
112	TP_printk("dst %u vec %x (%s|%s|%s%s)",
113		  (u8)(__entry->address >> 12), (u8)__entry->data,
 
114		  __print_symbolic((__entry->data >> 8 & 0x7), kvm_deliver_mode),
115		  (__entry->address & (1<<2)) ? "logical" : "physical",
116		  (__entry->data & (1<<15)) ? "level" : "edge",
117		  (__entry->address & (1<<3)) ? "|rh" : "")
118);
119
120#define kvm_irqchips						\
121	{KVM_IRQCHIP_PIC_MASTER,	"PIC master"},		\
122	{KVM_IRQCHIP_PIC_SLAVE,		"PIC slave"},		\
123	{KVM_IRQCHIP_IOAPIC,		"IOAPIC"}
124
125#endif /* defined(__KVM_HAVE_IOAPIC) */
126
127#if defined(CONFIG_HAVE_KVM_IRQCHIP)
 
 
 
 
 
 
 
 
128
129TRACE_EVENT(kvm_ack_irq,
130	TP_PROTO(unsigned int irqchip, unsigned int pin),
131	TP_ARGS(irqchip, pin),
132
133	TP_STRUCT__entry(
134		__field(	unsigned int,	irqchip		)
135		__field(	unsigned int,	pin		)
136	),
137
138	TP_fast_assign(
139		__entry->irqchip	= irqchip;
140		__entry->pin		= pin;
141	),
142
143#ifdef kvm_irqchips
144	TP_printk("irqchip %s pin %u",
145		  __print_symbolic(__entry->irqchip, kvm_irqchips),
146		 __entry->pin)
147#else
148	TP_printk("irqchip %d pin %u", __entry->irqchip, __entry->pin)
149#endif
150);
151
152#endif /* defined(CONFIG_HAVE_KVM_IRQCHIP) */
153
154
155
156#define KVM_TRACE_MMIO_READ_UNSATISFIED 0
157#define KVM_TRACE_MMIO_READ 1
158#define KVM_TRACE_MMIO_WRITE 2
159
160#define kvm_trace_symbol_mmio \
161	{ KVM_TRACE_MMIO_READ_UNSATISFIED, "unsatisfied-read" }, \
162	{ KVM_TRACE_MMIO_READ, "read" }, \
163	{ KVM_TRACE_MMIO_WRITE, "write" }
164
165TRACE_EVENT(kvm_mmio,
166	TP_PROTO(int type, int len, u64 gpa, u64 val),
167	TP_ARGS(type, len, gpa, val),
168
169	TP_STRUCT__entry(
170		__field(	u32,	type		)
171		__field(	u32,	len		)
172		__field(	u64,	gpa		)
173		__field(	u64,	val		)
174	),
175
176	TP_fast_assign(
177		__entry->type		= type;
178		__entry->len		= len;
179		__entry->gpa		= gpa;
180		__entry->val		= val;
 
 
 
181	),
182
183	TP_printk("mmio %s len %u gpa 0x%llx val 0x%llx",
184		  __print_symbolic(__entry->type, kvm_trace_symbol_mmio),
185		  __entry->len, __entry->gpa, __entry->val)
186);
187
188#define kvm_fpu_load_symbol	\
189	{0, "unload"},		\
190	{1, "load"}
191
192TRACE_EVENT(kvm_fpu,
193	TP_PROTO(int load),
194	TP_ARGS(load),
195
196	TP_STRUCT__entry(
197		__field(	u32,	        load		)
198	),
199
200	TP_fast_assign(
201		__entry->load		= load;
202	),
203
204	TP_printk("%s", __print_symbolic(__entry->load, kvm_fpu_load_symbol))
205);
206
207TRACE_EVENT(kvm_age_page,
208	TP_PROTO(ulong hva, struct kvm_memory_slot *slot, int ref),
209	TP_ARGS(hva, slot, ref),
210
211	TP_STRUCT__entry(
212		__field(	u64,	hva		)
213		__field(	u64,	gfn		)
214		__field(	u8,	referenced	)
215	),
216
217	TP_fast_assign(
218		__entry->hva		= hva;
219		__entry->gfn		=
220		  slot->base_gfn + ((hva - slot->userspace_addr) >> PAGE_SHIFT);
221		__entry->referenced	= ref;
222	),
223
224	TP_printk("hva %llx gfn %llx %s",
225		  __entry->hva, __entry->gfn,
226		  __entry->referenced ? "YOUNG" : "OLD")
227);
228
229#ifdef CONFIG_KVM_ASYNC_PF
230DECLARE_EVENT_CLASS(kvm_async_get_page_class,
231
232	TP_PROTO(u64 gva, u64 gfn),
233
234	TP_ARGS(gva, gfn),
235
236	TP_STRUCT__entry(
237		__field(__u64, gva)
238		__field(u64, gfn)
239	),
240
241	TP_fast_assign(
242		__entry->gva = gva;
243		__entry->gfn = gfn;
244	),
245
246	TP_printk("gva = %#llx, gfn = %#llx", __entry->gva, __entry->gfn)
247);
248
249DEFINE_EVENT(kvm_async_get_page_class, kvm_try_async_get_page,
250
251	TP_PROTO(u64 gva, u64 gfn),
252
253	TP_ARGS(gva, gfn)
254);
255
256DEFINE_EVENT(kvm_async_get_page_class, kvm_async_pf_doublefault,
257
258	TP_PROTO(u64 gva, u64 gfn),
259
260	TP_ARGS(gva, gfn)
261);
262
263DECLARE_EVENT_CLASS(kvm_async_pf_nopresent_ready,
264
265	TP_PROTO(u64 token, u64 gva),
266
267	TP_ARGS(token, gva),
268
269	TP_STRUCT__entry(
270		__field(__u64, token)
271		__field(__u64, gva)
272	),
273
274	TP_fast_assign(
275		__entry->token = token;
276		__entry->gva = gva;
277	),
278
279	TP_printk("token %#llx gva %#llx", __entry->token, __entry->gva)
280
281);
282
283DEFINE_EVENT(kvm_async_pf_nopresent_ready, kvm_async_pf_not_present,
284
285	TP_PROTO(u64 token, u64 gva),
286
287	TP_ARGS(token, gva)
288);
289
290DEFINE_EVENT(kvm_async_pf_nopresent_ready, kvm_async_pf_ready,
291
292	TP_PROTO(u64 token, u64 gva),
293
294	TP_ARGS(token, gva)
295);
296
297TRACE_EVENT(
298	kvm_async_pf_completed,
299	TP_PROTO(unsigned long address, u64 gva),
300	TP_ARGS(address, gva),
301
302	TP_STRUCT__entry(
303		__field(unsigned long, address)
304		__field(u64, gva)
305		),
306
307	TP_fast_assign(
308		__entry->address = address;
309		__entry->gva = gva;
310		),
311
312	TP_printk("gva %#llx address %#lx",  __entry->gva,
313		  __entry->address)
314);
315
316#endif
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
317
318#endif /* _TRACE_KVM_MAIN_H */
319
320/* This part must be outside protection */
321#include <trace/define_trace.h>