Linux Audio

Check our new training course

Loading...
v6.8
  1/* SPDX-License-Identifier: GPL-2.0 */
  2/*
  3 * xHCI host controller driver
  4 *
  5 * Copyright (C) 2013 Xenia Ragiadakou
  6 *
  7 * Author: Xenia Ragiadakou
  8 * Email : burzalodowa@gmail.com
  9 */
 10
 11#undef TRACE_SYSTEM
 12#define TRACE_SYSTEM xhci-hcd
 13
 14/*
 15 * The TRACE_SYSTEM_VAR defaults to TRACE_SYSTEM, but must be a
 16 * legitimate C variable. It is not exported to user space.
 17 */
 18#undef TRACE_SYSTEM_VAR
 19#define TRACE_SYSTEM_VAR xhci_hcd
 20
 21#if !defined(__XHCI_TRACE_H) || defined(TRACE_HEADER_MULTI_READ)
 22#define __XHCI_TRACE_H
 23
 24#include <linux/tracepoint.h>
 25#include "xhci.h"
 26#include "xhci-dbgcap.h"
 27
 
 
 28DECLARE_EVENT_CLASS(xhci_log_msg,
 29	TP_PROTO(struct va_format *vaf),
 30	TP_ARGS(vaf),
 31	TP_STRUCT__entry(__vstring(msg, vaf->fmt, vaf->va)),
 32	TP_fast_assign(
 33		__assign_vstr(msg, vaf->fmt, vaf->va);
 34	),
 35	TP_printk("%s", __get_str(msg))
 36);
 37
 38DEFINE_EVENT(xhci_log_msg, xhci_dbg_address,
 39	TP_PROTO(struct va_format *vaf),
 40	TP_ARGS(vaf)
 41);
 42
 43DEFINE_EVENT(xhci_log_msg, xhci_dbg_context_change,
 44	TP_PROTO(struct va_format *vaf),
 45	TP_ARGS(vaf)
 46);
 47
 48DEFINE_EVENT(xhci_log_msg, xhci_dbg_quirks,
 49	TP_PROTO(struct va_format *vaf),
 50	TP_ARGS(vaf)
 51);
 52
 53DEFINE_EVENT(xhci_log_msg, xhci_dbg_reset_ep,
 54	TP_PROTO(struct va_format *vaf),
 55	TP_ARGS(vaf)
 56);
 57
 58DEFINE_EVENT(xhci_log_msg, xhci_dbg_cancel_urb,
 59	TP_PROTO(struct va_format *vaf),
 60	TP_ARGS(vaf)
 61);
 62
 63DEFINE_EVENT(xhci_log_msg, xhci_dbg_init,
 64	TP_PROTO(struct va_format *vaf),
 65	TP_ARGS(vaf)
 66);
 67
 68DEFINE_EVENT(xhci_log_msg, xhci_dbg_ring_expansion,
 69	TP_PROTO(struct va_format *vaf),
 70	TP_ARGS(vaf)
 71);
 72
 73DECLARE_EVENT_CLASS(xhci_log_ctx,
 74	TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx,
 75		 unsigned int ep_num),
 76	TP_ARGS(xhci, ctx, ep_num),
 77	TP_STRUCT__entry(
 78		__field(int, ctx_64)
 79		__field(unsigned, ctx_type)
 80		__field(dma_addr_t, ctx_dma)
 81		__field(u8 *, ctx_va)
 82		__field(unsigned, ctx_ep_num)
 
 83		__dynamic_array(u32, ctx_data,
 84			((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 8) *
 85			((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1))
 86	),
 87	TP_fast_assign(
 
 88
 
 89		__entry->ctx_64 = HCC_64BYTE_CONTEXT(xhci->hcc_params);
 90		__entry->ctx_type = ctx->type;
 91		__entry->ctx_dma = ctx->dma;
 92		__entry->ctx_va = ctx->bytes;
 
 93		__entry->ctx_ep_num = ep_num;
 94		memcpy(__get_dynamic_array(ctx_data), ctx->bytes,
 95			((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 32) *
 96			((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1));
 97	),
 98	TP_printk("ctx_64=%d, ctx_type=%u, ctx_dma=@%llx, ctx_va=@%p",
 99			__entry->ctx_64, __entry->ctx_type,
100			(unsigned long long) __entry->ctx_dma, __entry->ctx_va
101	)
102);
103
104DEFINE_EVENT(xhci_log_ctx, xhci_address_ctx,
105	TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx,
106		 unsigned int ep_num),
107	TP_ARGS(xhci, ctx, ep_num)
108);
109
110DECLARE_EVENT_CLASS(xhci_log_trb,
111	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
112	TP_ARGS(ring, trb),
113	TP_STRUCT__entry(
114		__field(u32, type)
115		__field(u32, field0)
116		__field(u32, field1)
117		__field(u32, field2)
118		__field(u32, field3)
119	),
120	TP_fast_assign(
121		__entry->type = ring->type;
122		__entry->field0 = le32_to_cpu(trb->field[0]);
123		__entry->field1 = le32_to_cpu(trb->field[1]);
124		__entry->field2 = le32_to_cpu(trb->field[2]);
125		__entry->field3 = le32_to_cpu(trb->field[3]);
126	),
127	TP_printk("%s: %s", xhci_ring_type_string(__entry->type),
128		  xhci_decode_trb(__get_buf(XHCI_MSG_MAX), XHCI_MSG_MAX, __entry->field0,
129				  __entry->field1, __entry->field2, __entry->field3)
130	)
131);
132
133DEFINE_EVENT(xhci_log_trb, xhci_handle_event,
134	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
135	TP_ARGS(ring, trb)
136);
137
138DEFINE_EVENT(xhci_log_trb, xhci_handle_command,
139	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
140	TP_ARGS(ring, trb)
141);
142
143DEFINE_EVENT(xhci_log_trb, xhci_handle_transfer,
144	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
145	TP_ARGS(ring, trb)
146);
147
148DEFINE_EVENT(xhci_log_trb, xhci_queue_trb,
149	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
150	TP_ARGS(ring, trb)
151);
152
153DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_event,
154	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
155	TP_ARGS(ring, trb)
156);
157
158DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_transfer,
159	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
160	TP_ARGS(ring, trb)
161);
162
163DEFINE_EVENT(xhci_log_trb, xhci_dbc_gadget_ep_queue,
164	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
165	TP_ARGS(ring, trb)
166);
167
168DECLARE_EVENT_CLASS(xhci_log_free_virt_dev,
169	TP_PROTO(struct xhci_virt_device *vdev),
170	TP_ARGS(vdev),
171	TP_STRUCT__entry(
172		__field(void *, vdev)
173		__field(unsigned long long, out_ctx)
174		__field(unsigned long long, in_ctx)
175		__field(u8, fake_port)
176		__field(u8, real_port)
177		__field(u16, current_mel)
178
179	),
180	TP_fast_assign(
181		__entry->vdev = vdev;
182		__entry->in_ctx = (unsigned long long) vdev->in_ctx->dma;
183		__entry->out_ctx = (unsigned long long) vdev->out_ctx->dma;
184		__entry->fake_port = (u8) vdev->fake_port;
185		__entry->real_port = (u8) vdev->real_port;
186		__entry->current_mel = (u16) vdev->current_mel;
187		),
188	TP_printk("vdev %p ctx %llx | %llx fake_port %d real_port %d current_mel %d",
189		__entry->vdev, __entry->in_ctx, __entry->out_ctx,
190		__entry->fake_port, __entry->real_port, __entry->current_mel
191	)
192);
193
194DEFINE_EVENT(xhci_log_free_virt_dev, xhci_free_virt_device,
195	TP_PROTO(struct xhci_virt_device *vdev),
196	TP_ARGS(vdev)
197);
198
199DECLARE_EVENT_CLASS(xhci_log_virt_dev,
200	TP_PROTO(struct xhci_virt_device *vdev),
201	TP_ARGS(vdev),
202	TP_STRUCT__entry(
203		__field(void *, vdev)
204		__field(unsigned long long, out_ctx)
205		__field(unsigned long long, in_ctx)
206		__field(int, devnum)
207		__field(int, state)
208		__field(int, speed)
209		__field(u8, portnum)
210		__field(u8, level)
211		__field(int, slot_id)
212	),
213	TP_fast_assign(
214		__entry->vdev = vdev;
215		__entry->in_ctx = (unsigned long long) vdev->in_ctx->dma;
216		__entry->out_ctx = (unsigned long long) vdev->out_ctx->dma;
217		__entry->devnum = vdev->udev->devnum;
218		__entry->state = vdev->udev->state;
219		__entry->speed = vdev->udev->speed;
220		__entry->portnum = vdev->udev->portnum;
221		__entry->level = vdev->udev->level;
222		__entry->slot_id = vdev->udev->slot_id;
223	),
224	TP_printk("vdev %p ctx %llx | %llx num %d state %d speed %d port %d level %d slot %d",
225		__entry->vdev, __entry->in_ctx, __entry->out_ctx,
226		__entry->devnum, __entry->state, __entry->speed,
227		__entry->portnum, __entry->level, __entry->slot_id
228	)
229);
230
231DEFINE_EVENT(xhci_log_virt_dev, xhci_alloc_virt_device,
232	TP_PROTO(struct xhci_virt_device *vdev),
233	TP_ARGS(vdev)
234);
235
236DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_device,
237	TP_PROTO(struct xhci_virt_device *vdev),
238	TP_ARGS(vdev)
239);
240
241DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_addressable_virt_device,
242	TP_PROTO(struct xhci_virt_device *vdev),
243	TP_ARGS(vdev)
244);
245
246DEFINE_EVENT(xhci_log_virt_dev, xhci_stop_device,
247	TP_PROTO(struct xhci_virt_device *vdev),
248	TP_ARGS(vdev)
249);
250
251DECLARE_EVENT_CLASS(xhci_log_urb,
252	TP_PROTO(struct urb *urb),
253	TP_ARGS(urb),
254	TP_STRUCT__entry(
255		__field(void *, urb)
256		__field(unsigned int, pipe)
257		__field(unsigned int, stream)
258		__field(int, status)
259		__field(unsigned int, flags)
260		__field(int, num_mapped_sgs)
261		__field(int, num_sgs)
262		__field(int, length)
263		__field(int, actual)
264		__field(int, epnum)
265		__field(int, dir_in)
266		__field(int, type)
267		__field(int, slot_id)
268	),
269	TP_fast_assign(
270		__entry->urb = urb;
271		__entry->pipe = urb->pipe;
272		__entry->stream = urb->stream_id;
273		__entry->status = urb->status;
274		__entry->flags = urb->transfer_flags;
275		__entry->num_mapped_sgs = urb->num_mapped_sgs;
276		__entry->num_sgs = urb->num_sgs;
277		__entry->length = urb->transfer_buffer_length;
278		__entry->actual = urb->actual_length;
279		__entry->epnum = usb_endpoint_num(&urb->ep->desc);
280		__entry->dir_in = usb_endpoint_dir_in(&urb->ep->desc);
281		__entry->type = usb_endpoint_type(&urb->ep->desc);
282		__entry->slot_id = urb->dev->slot_id;
283	),
284	TP_printk("ep%d%s-%s: urb %p pipe %u slot %d length %d/%d sgs %d/%d stream %d flags %08x",
285			__entry->epnum, __entry->dir_in ? "in" : "out",
286			__print_symbolic(__entry->type,
287				   { USB_ENDPOINT_XFER_INT,	"intr" },
288				   { USB_ENDPOINT_XFER_CONTROL,	"control" },
289				   { USB_ENDPOINT_XFER_BULK,	"bulk" },
290				   { USB_ENDPOINT_XFER_ISOC,	"isoc" }),
291			__entry->urb, __entry->pipe, __entry->slot_id,
292			__entry->actual, __entry->length, __entry->num_mapped_sgs,
293			__entry->num_sgs, __entry->stream, __entry->flags
294		)
295);
296
297DEFINE_EVENT(xhci_log_urb, xhci_urb_enqueue,
298	TP_PROTO(struct urb *urb),
299	TP_ARGS(urb)
300);
301
302DEFINE_EVENT(xhci_log_urb, xhci_urb_giveback,
303	TP_PROTO(struct urb *urb),
304	TP_ARGS(urb)
305);
306
307DEFINE_EVENT(xhci_log_urb, xhci_urb_dequeue,
308	TP_PROTO(struct urb *urb),
309	TP_ARGS(urb)
310);
311
312DECLARE_EVENT_CLASS(xhci_log_ep_ctx,
313	TP_PROTO(struct xhci_ep_ctx *ctx),
314	TP_ARGS(ctx),
315	TP_STRUCT__entry(
316		__field(u32, info)
317		__field(u32, info2)
318		__field(u64, deq)
319		__field(u32, tx_info)
320	),
321	TP_fast_assign(
322		__entry->info = le32_to_cpu(ctx->ep_info);
323		__entry->info2 = le32_to_cpu(ctx->ep_info2);
324		__entry->deq = le64_to_cpu(ctx->deq);
325		__entry->tx_info = le32_to_cpu(ctx->tx_info);
326	),
327	TP_printk("%s", xhci_decode_ep_context(__get_buf(XHCI_MSG_MAX),
328		__entry->info, __entry->info2, __entry->deq, __entry->tx_info)
329	)
330);
331
332DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_stop_ep,
333	TP_PROTO(struct xhci_ep_ctx *ctx),
334	TP_ARGS(ctx)
335);
336
337DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_set_deq_ep,
338	TP_PROTO(struct xhci_ep_ctx *ctx),
339	TP_ARGS(ctx)
340);
341
342DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_reset_ep,
343	TP_PROTO(struct xhci_ep_ctx *ctx),
344	TP_ARGS(ctx)
345);
346
347DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_config_ep,
348	TP_PROTO(struct xhci_ep_ctx *ctx),
349	TP_ARGS(ctx)
350);
351
352DEFINE_EVENT(xhci_log_ep_ctx, xhci_add_endpoint,
353	TP_PROTO(struct xhci_ep_ctx *ctx),
354	TP_ARGS(ctx)
355);
356
357DECLARE_EVENT_CLASS(xhci_log_slot_ctx,
358	TP_PROTO(struct xhci_slot_ctx *ctx),
359	TP_ARGS(ctx),
360	TP_STRUCT__entry(
361		__field(u32, info)
362		__field(u32, info2)
363		__field(u32, tt_info)
364		__field(u32, state)
365	),
366	TP_fast_assign(
367		__entry->info = le32_to_cpu(ctx->dev_info);
368		__entry->info2 = le32_to_cpu(ctx->dev_info2);
369		__entry->tt_info = le64_to_cpu(ctx->tt_info);
370		__entry->state = le32_to_cpu(ctx->dev_state);
371	),
372	TP_printk("%s", xhci_decode_slot_context(__get_buf(XHCI_MSG_MAX),
373			__entry->info, __entry->info2,
374			__entry->tt_info, __entry->state)
375	)
376);
377
378DEFINE_EVENT(xhci_log_slot_ctx, xhci_alloc_dev,
379	TP_PROTO(struct xhci_slot_ctx *ctx),
380	TP_ARGS(ctx)
381);
382
383DEFINE_EVENT(xhci_log_slot_ctx, xhci_free_dev,
384	TP_PROTO(struct xhci_slot_ctx *ctx),
385	TP_ARGS(ctx)
386);
387
388DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_disable_slot,
389	TP_PROTO(struct xhci_slot_ctx *ctx),
390	TP_ARGS(ctx)
391);
392
393DEFINE_EVENT(xhci_log_slot_ctx, xhci_discover_or_reset_device,
394	TP_PROTO(struct xhci_slot_ctx *ctx),
395	TP_ARGS(ctx)
396);
397
398DEFINE_EVENT(xhci_log_slot_ctx, xhci_setup_device_slot,
399	TP_PROTO(struct xhci_slot_ctx *ctx),
400	TP_ARGS(ctx)
401);
402
403DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_addr_dev,
404	TP_PROTO(struct xhci_slot_ctx *ctx),
405	TP_ARGS(ctx)
406);
407
408DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_reset_dev,
409	TP_PROTO(struct xhci_slot_ctx *ctx),
410	TP_ARGS(ctx)
411);
412
413DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_set_deq,
414	TP_PROTO(struct xhci_slot_ctx *ctx),
415	TP_ARGS(ctx)
416);
417
418DEFINE_EVENT(xhci_log_slot_ctx, xhci_configure_endpoint,
419	TP_PROTO(struct xhci_slot_ctx *ctx),
420	TP_ARGS(ctx)
421);
422
423DECLARE_EVENT_CLASS(xhci_log_ctrl_ctx,
424	TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
425	TP_ARGS(ctrl_ctx),
426	TP_STRUCT__entry(
427		__field(u32, drop)
428		__field(u32, add)
429	),
430	TP_fast_assign(
431		__entry->drop = le32_to_cpu(ctrl_ctx->drop_flags);
432		__entry->add = le32_to_cpu(ctrl_ctx->add_flags);
433	),
434	TP_printk("%s", xhci_decode_ctrl_ctx(__get_buf(XHCI_MSG_MAX), __entry->drop, __entry->add)
435	)
436);
437
438DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_address_ctrl_ctx,
439	TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
440	TP_ARGS(ctrl_ctx)
441);
442
443DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_configure_endpoint_ctrl_ctx,
444	TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
445	TP_ARGS(ctrl_ctx)
446);
447
448DECLARE_EVENT_CLASS(xhci_log_ring,
449	TP_PROTO(struct xhci_ring *ring),
450	TP_ARGS(ring),
451	TP_STRUCT__entry(
452		__field(u32, type)
453		__field(void *, ring)
454		__field(dma_addr_t, enq)
455		__field(dma_addr_t, deq)
456		__field(dma_addr_t, enq_seg)
457		__field(dma_addr_t, deq_seg)
458		__field(unsigned int, num_segs)
459		__field(unsigned int, stream_id)
460		__field(unsigned int, cycle_state)
 
461		__field(unsigned int, bounce_buf_len)
462	),
463	TP_fast_assign(
464		__entry->ring = ring;
465		__entry->type = ring->type;
466		__entry->num_segs = ring->num_segs;
467		__entry->stream_id = ring->stream_id;
468		__entry->enq_seg = ring->enq_seg->dma;
469		__entry->deq_seg = ring->deq_seg->dma;
470		__entry->cycle_state = ring->cycle_state;
 
471		__entry->bounce_buf_len = ring->bounce_buf_len;
472		__entry->enq = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue);
473		__entry->deq = xhci_trb_virt_to_dma(ring->deq_seg, ring->dequeue);
474	),
475	TP_printk("%s %p: enq %pad(%pad) deq %pad(%pad) segs %d stream %d bounce %d cycle %d",
476			xhci_ring_type_string(__entry->type), __entry->ring,
477			&__entry->enq, &__entry->enq_seg,
478			&__entry->deq, &__entry->deq_seg,
479			__entry->num_segs,
480			__entry->stream_id,
 
481			__entry->bounce_buf_len,
482			__entry->cycle_state
483		)
484);
485
486DEFINE_EVENT(xhci_log_ring, xhci_ring_alloc,
487	TP_PROTO(struct xhci_ring *ring),
488	TP_ARGS(ring)
489);
490
491DEFINE_EVENT(xhci_log_ring, xhci_ring_free,
492	TP_PROTO(struct xhci_ring *ring),
493	TP_ARGS(ring)
494);
495
496DEFINE_EVENT(xhci_log_ring, xhci_ring_expansion,
497	TP_PROTO(struct xhci_ring *ring),
498	TP_ARGS(ring)
499);
500
501DEFINE_EVENT(xhci_log_ring, xhci_inc_enq,
502	TP_PROTO(struct xhci_ring *ring),
503	TP_ARGS(ring)
504);
505
506DEFINE_EVENT(xhci_log_ring, xhci_inc_deq,
507	TP_PROTO(struct xhci_ring *ring),
508	TP_ARGS(ring)
509);
510
511DECLARE_EVENT_CLASS(xhci_log_portsc,
512		    TP_PROTO(struct xhci_port *port, u32 portsc),
513		    TP_ARGS(port, portsc),
514		    TP_STRUCT__entry(
515				     __field(u32, busnum)
516				     __field(u32, portnum)
517				     __field(u32, portsc)
518				     ),
519		    TP_fast_assign(
520				   __entry->busnum = port->rhub->hcd->self.busnum;
521				   __entry->portnum = port->hcd_portnum;
522				   __entry->portsc = portsc;
523				   ),
524		    TP_printk("port %d-%d: %s",
525			      __entry->busnum,
526			      __entry->portnum,
527			      xhci_decode_portsc(__get_buf(XHCI_MSG_MAX), __entry->portsc)
528			      )
529);
530
531DEFINE_EVENT(xhci_log_portsc, xhci_handle_port_status,
532	     TP_PROTO(struct xhci_port *port, u32 portsc),
533	     TP_ARGS(port, portsc)
534);
535
536DEFINE_EVENT(xhci_log_portsc, xhci_get_port_status,
537	     TP_PROTO(struct xhci_port *port, u32 portsc),
538	     TP_ARGS(port, portsc)
539);
540
541DEFINE_EVENT(xhci_log_portsc, xhci_hub_status_data,
542	     TP_PROTO(struct xhci_port *port, u32 portsc),
543	     TP_ARGS(port, portsc)
544);
545
546DECLARE_EVENT_CLASS(xhci_log_doorbell,
547	TP_PROTO(u32 slot, u32 doorbell),
548	TP_ARGS(slot, doorbell),
549	TP_STRUCT__entry(
550		__field(u32, slot)
551		__field(u32, doorbell)
552	),
553	TP_fast_assign(
554		__entry->slot = slot;
555		__entry->doorbell = doorbell;
556	),
557	TP_printk("Ring doorbell for %s",
558		  xhci_decode_doorbell(__get_buf(XHCI_MSG_MAX), __entry->slot, __entry->doorbell)
559	)
560);
561
562DEFINE_EVENT(xhci_log_doorbell, xhci_ring_ep_doorbell,
563	     TP_PROTO(u32 slot, u32 doorbell),
564	     TP_ARGS(slot, doorbell)
565);
566
567DEFINE_EVENT(xhci_log_doorbell, xhci_ring_host_doorbell,
568	     TP_PROTO(u32 slot, u32 doorbell),
569	     TP_ARGS(slot, doorbell)
570);
571
572DECLARE_EVENT_CLASS(xhci_dbc_log_request,
573	TP_PROTO(struct dbc_request *req),
574	TP_ARGS(req),
575	TP_STRUCT__entry(
576		__field(struct dbc_request *, req)
577		__field(bool, dir)
578		__field(unsigned int, actual)
579		__field(unsigned int, length)
580		__field(int, status)
581	),
582	TP_fast_assign(
583		__entry->req = req;
584		__entry->dir = req->direction;
585		__entry->actual = req->actual;
586		__entry->length = req->length;
587		__entry->status = req->status;
588	),
589	TP_printk("%s: req %p length %u/%u ==> %d",
590		__entry->dir ? "bulk-in" : "bulk-out",
591		__entry->req, __entry->actual,
592		__entry->length, __entry->status
593	)
594);
595
596DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_alloc_request,
597	TP_PROTO(struct dbc_request *req),
598	TP_ARGS(req)
599);
600
601DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_free_request,
602	TP_PROTO(struct dbc_request *req),
603	TP_ARGS(req)
604);
605
606DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_queue_request,
607	TP_PROTO(struct dbc_request *req),
608	TP_ARGS(req)
609);
610
611DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_giveback_request,
612	TP_PROTO(struct dbc_request *req),
613	TP_ARGS(req)
614);
615#endif /* __XHCI_TRACE_H */
616
617/* this part must be outside header guard */
618
619#undef TRACE_INCLUDE_PATH
620#define TRACE_INCLUDE_PATH .
621
622#undef TRACE_INCLUDE_FILE
623#define TRACE_INCLUDE_FILE xhci-trace
624
625#include <trace/define_trace.h>
v5.9
  1/* SPDX-License-Identifier: GPL-2.0 */
  2/*
  3 * xHCI host controller driver
  4 *
  5 * Copyright (C) 2013 Xenia Ragiadakou
  6 *
  7 * Author: Xenia Ragiadakou
  8 * Email : burzalodowa@gmail.com
  9 */
 10
 11#undef TRACE_SYSTEM
 12#define TRACE_SYSTEM xhci-hcd
 13
 14/*
 15 * The TRACE_SYSTEM_VAR defaults to TRACE_SYSTEM, but must be a
 16 * legitimate C variable. It is not exported to user space.
 17 */
 18#undef TRACE_SYSTEM_VAR
 19#define TRACE_SYSTEM_VAR xhci_hcd
 20
 21#if !defined(__XHCI_TRACE_H) || defined(TRACE_HEADER_MULTI_READ)
 22#define __XHCI_TRACE_H
 23
 24#include <linux/tracepoint.h>
 25#include "xhci.h"
 26#include "xhci-dbgcap.h"
 27
 28#define XHCI_MSG_MAX	500
 29
 30DECLARE_EVENT_CLASS(xhci_log_msg,
 31	TP_PROTO(struct va_format *vaf),
 32	TP_ARGS(vaf),
 33	TP_STRUCT__entry(__dynamic_array(char, msg, XHCI_MSG_MAX)),
 34	TP_fast_assign(
 35		vsnprintf(__get_str(msg), XHCI_MSG_MAX, vaf->fmt, *vaf->va);
 36	),
 37	TP_printk("%s", __get_str(msg))
 38);
 39
 40DEFINE_EVENT(xhci_log_msg, xhci_dbg_address,
 41	TP_PROTO(struct va_format *vaf),
 42	TP_ARGS(vaf)
 43);
 44
 45DEFINE_EVENT(xhci_log_msg, xhci_dbg_context_change,
 46	TP_PROTO(struct va_format *vaf),
 47	TP_ARGS(vaf)
 48);
 49
 50DEFINE_EVENT(xhci_log_msg, xhci_dbg_quirks,
 51	TP_PROTO(struct va_format *vaf),
 52	TP_ARGS(vaf)
 53);
 54
 55DEFINE_EVENT(xhci_log_msg, xhci_dbg_reset_ep,
 56	TP_PROTO(struct va_format *vaf),
 57	TP_ARGS(vaf)
 58);
 59
 60DEFINE_EVENT(xhci_log_msg, xhci_dbg_cancel_urb,
 61	TP_PROTO(struct va_format *vaf),
 62	TP_ARGS(vaf)
 63);
 64
 65DEFINE_EVENT(xhci_log_msg, xhci_dbg_init,
 66	TP_PROTO(struct va_format *vaf),
 67	TP_ARGS(vaf)
 68);
 69
 70DEFINE_EVENT(xhci_log_msg, xhci_dbg_ring_expansion,
 71	TP_PROTO(struct va_format *vaf),
 72	TP_ARGS(vaf)
 73);
 74
 75DECLARE_EVENT_CLASS(xhci_log_ctx,
 76	TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx,
 77		 unsigned int ep_num),
 78	TP_ARGS(xhci, ctx, ep_num),
 79	TP_STRUCT__entry(
 80		__field(int, ctx_64)
 81		__field(unsigned, ctx_type)
 82		__field(dma_addr_t, ctx_dma)
 83		__field(u8 *, ctx_va)
 84		__field(unsigned, ctx_ep_num)
 85		__field(int, slot_id)
 86		__dynamic_array(u32, ctx_data,
 87			((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 8) *
 88			((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1))
 89	),
 90	TP_fast_assign(
 91		struct usb_device *udev;
 92
 93		udev = to_usb_device(xhci_to_hcd(xhci)->self.controller);
 94		__entry->ctx_64 = HCC_64BYTE_CONTEXT(xhci->hcc_params);
 95		__entry->ctx_type = ctx->type;
 96		__entry->ctx_dma = ctx->dma;
 97		__entry->ctx_va = ctx->bytes;
 98		__entry->slot_id = udev->slot_id;
 99		__entry->ctx_ep_num = ep_num;
100		memcpy(__get_dynamic_array(ctx_data), ctx->bytes,
101			((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 32) *
102			((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1));
103	),
104	TP_printk("ctx_64=%d, ctx_type=%u, ctx_dma=@%llx, ctx_va=@%p",
105			__entry->ctx_64, __entry->ctx_type,
106			(unsigned long long) __entry->ctx_dma, __entry->ctx_va
107	)
108);
109
110DEFINE_EVENT(xhci_log_ctx, xhci_address_ctx,
111	TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx,
112		 unsigned int ep_num),
113	TP_ARGS(xhci, ctx, ep_num)
114);
115
116DECLARE_EVENT_CLASS(xhci_log_trb,
117	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
118	TP_ARGS(ring, trb),
119	TP_STRUCT__entry(
120		__field(u32, type)
121		__field(u32, field0)
122		__field(u32, field1)
123		__field(u32, field2)
124		__field(u32, field3)
125	),
126	TP_fast_assign(
127		__entry->type = ring->type;
128		__entry->field0 = le32_to_cpu(trb->field[0]);
129		__entry->field1 = le32_to_cpu(trb->field[1]);
130		__entry->field2 = le32_to_cpu(trb->field[2]);
131		__entry->field3 = le32_to_cpu(trb->field[3]);
132	),
133	TP_printk("%s: %s", xhci_ring_type_string(__entry->type),
134			xhci_decode_trb(__entry->field0, __entry->field1,
135					__entry->field2, __entry->field3)
136	)
137);
138
139DEFINE_EVENT(xhci_log_trb, xhci_handle_event,
140	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
141	TP_ARGS(ring, trb)
142);
143
144DEFINE_EVENT(xhci_log_trb, xhci_handle_command,
145	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
146	TP_ARGS(ring, trb)
147);
148
149DEFINE_EVENT(xhci_log_trb, xhci_handle_transfer,
150	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
151	TP_ARGS(ring, trb)
152);
153
154DEFINE_EVENT(xhci_log_trb, xhci_queue_trb,
155	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
156	TP_ARGS(ring, trb)
157);
158
159DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_event,
160	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
161	TP_ARGS(ring, trb)
162);
163
164DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_transfer,
165	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
166	TP_ARGS(ring, trb)
167);
168
169DEFINE_EVENT(xhci_log_trb, xhci_dbc_gadget_ep_queue,
170	TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
171	TP_ARGS(ring, trb)
172);
173
174DECLARE_EVENT_CLASS(xhci_log_free_virt_dev,
175	TP_PROTO(struct xhci_virt_device *vdev),
176	TP_ARGS(vdev),
177	TP_STRUCT__entry(
178		__field(void *, vdev)
179		__field(unsigned long long, out_ctx)
180		__field(unsigned long long, in_ctx)
181		__field(u8, fake_port)
182		__field(u8, real_port)
183		__field(u16, current_mel)
184
185	),
186	TP_fast_assign(
187		__entry->vdev = vdev;
188		__entry->in_ctx = (unsigned long long) vdev->in_ctx->dma;
189		__entry->out_ctx = (unsigned long long) vdev->out_ctx->dma;
190		__entry->fake_port = (u8) vdev->fake_port;
191		__entry->real_port = (u8) vdev->real_port;
192		__entry->current_mel = (u16) vdev->current_mel;
193		),
194	TP_printk("vdev %p ctx %llx | %llx fake_port %d real_port %d current_mel %d",
195		__entry->vdev, __entry->in_ctx, __entry->out_ctx,
196		__entry->fake_port, __entry->real_port, __entry->current_mel
197	)
198);
199
200DEFINE_EVENT(xhci_log_free_virt_dev, xhci_free_virt_device,
201	TP_PROTO(struct xhci_virt_device *vdev),
202	TP_ARGS(vdev)
203);
204
205DECLARE_EVENT_CLASS(xhci_log_virt_dev,
206	TP_PROTO(struct xhci_virt_device *vdev),
207	TP_ARGS(vdev),
208	TP_STRUCT__entry(
209		__field(void *, vdev)
210		__field(unsigned long long, out_ctx)
211		__field(unsigned long long, in_ctx)
212		__field(int, devnum)
213		__field(int, state)
214		__field(int, speed)
215		__field(u8, portnum)
216		__field(u8, level)
217		__field(int, slot_id)
218	),
219	TP_fast_assign(
220		__entry->vdev = vdev;
221		__entry->in_ctx = (unsigned long long) vdev->in_ctx->dma;
222		__entry->out_ctx = (unsigned long long) vdev->out_ctx->dma;
223		__entry->devnum = vdev->udev->devnum;
224		__entry->state = vdev->udev->state;
225		__entry->speed = vdev->udev->speed;
226		__entry->portnum = vdev->udev->portnum;
227		__entry->level = vdev->udev->level;
228		__entry->slot_id = vdev->udev->slot_id;
229	),
230	TP_printk("vdev %p ctx %llx | %llx num %d state %d speed %d port %d level %d slot %d",
231		__entry->vdev, __entry->in_ctx, __entry->out_ctx,
232		__entry->devnum, __entry->state, __entry->speed,
233		__entry->portnum, __entry->level, __entry->slot_id
234	)
235);
236
237DEFINE_EVENT(xhci_log_virt_dev, xhci_alloc_virt_device,
238	TP_PROTO(struct xhci_virt_device *vdev),
239	TP_ARGS(vdev)
240);
241
242DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_device,
243	TP_PROTO(struct xhci_virt_device *vdev),
244	TP_ARGS(vdev)
245);
246
247DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_addressable_virt_device,
248	TP_PROTO(struct xhci_virt_device *vdev),
249	TP_ARGS(vdev)
250);
251
252DEFINE_EVENT(xhci_log_virt_dev, xhci_stop_device,
253	TP_PROTO(struct xhci_virt_device *vdev),
254	TP_ARGS(vdev)
255);
256
257DECLARE_EVENT_CLASS(xhci_log_urb,
258	TP_PROTO(struct urb *urb),
259	TP_ARGS(urb),
260	TP_STRUCT__entry(
261		__field(void *, urb)
262		__field(unsigned int, pipe)
263		__field(unsigned int, stream)
264		__field(int, status)
265		__field(unsigned int, flags)
266		__field(int, num_mapped_sgs)
267		__field(int, num_sgs)
268		__field(int, length)
269		__field(int, actual)
270		__field(int, epnum)
271		__field(int, dir_in)
272		__field(int, type)
273		__field(int, slot_id)
274	),
275	TP_fast_assign(
276		__entry->urb = urb;
277		__entry->pipe = urb->pipe;
278		__entry->stream = urb->stream_id;
279		__entry->status = urb->status;
280		__entry->flags = urb->transfer_flags;
281		__entry->num_mapped_sgs = urb->num_mapped_sgs;
282		__entry->num_sgs = urb->num_sgs;
283		__entry->length = urb->transfer_buffer_length;
284		__entry->actual = urb->actual_length;
285		__entry->epnum = usb_endpoint_num(&urb->ep->desc);
286		__entry->dir_in = usb_endpoint_dir_in(&urb->ep->desc);
287		__entry->type = usb_endpoint_type(&urb->ep->desc);
288		__entry->slot_id = urb->dev->slot_id;
289	),
290	TP_printk("ep%d%s-%s: urb %p pipe %u slot %d length %d/%d sgs %d/%d stream %d flags %08x",
291			__entry->epnum, __entry->dir_in ? "in" : "out",
292			__print_symbolic(__entry->type,
293				   { USB_ENDPOINT_XFER_INT,	"intr" },
294				   { USB_ENDPOINT_XFER_CONTROL,	"control" },
295				   { USB_ENDPOINT_XFER_BULK,	"bulk" },
296				   { USB_ENDPOINT_XFER_ISOC,	"isoc" }),
297			__entry->urb, __entry->pipe, __entry->slot_id,
298			__entry->actual, __entry->length, __entry->num_mapped_sgs,
299			__entry->num_sgs, __entry->stream, __entry->flags
300		)
301);
302
303DEFINE_EVENT(xhci_log_urb, xhci_urb_enqueue,
304	TP_PROTO(struct urb *urb),
305	TP_ARGS(urb)
306);
307
308DEFINE_EVENT(xhci_log_urb, xhci_urb_giveback,
309	TP_PROTO(struct urb *urb),
310	TP_ARGS(urb)
311);
312
313DEFINE_EVENT(xhci_log_urb, xhci_urb_dequeue,
314	TP_PROTO(struct urb *urb),
315	TP_ARGS(urb)
316);
317
318DECLARE_EVENT_CLASS(xhci_log_ep_ctx,
319	TP_PROTO(struct xhci_ep_ctx *ctx),
320	TP_ARGS(ctx),
321	TP_STRUCT__entry(
322		__field(u32, info)
323		__field(u32, info2)
324		__field(u64, deq)
325		__field(u32, tx_info)
326	),
327	TP_fast_assign(
328		__entry->info = le32_to_cpu(ctx->ep_info);
329		__entry->info2 = le32_to_cpu(ctx->ep_info2);
330		__entry->deq = le64_to_cpu(ctx->deq);
331		__entry->tx_info = le32_to_cpu(ctx->tx_info);
332	),
333	TP_printk("%s", xhci_decode_ep_context(__entry->info,
334		__entry->info2, __entry->deq, __entry->tx_info)
335	)
336);
337
338DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_stop_ep,
339	TP_PROTO(struct xhci_ep_ctx *ctx),
340	TP_ARGS(ctx)
341);
342
343DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_set_deq_ep,
344	TP_PROTO(struct xhci_ep_ctx *ctx),
345	TP_ARGS(ctx)
346);
347
348DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_reset_ep,
349	TP_PROTO(struct xhci_ep_ctx *ctx),
350	TP_ARGS(ctx)
351);
352
353DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_config_ep,
354	TP_PROTO(struct xhci_ep_ctx *ctx),
355	TP_ARGS(ctx)
356);
357
358DEFINE_EVENT(xhci_log_ep_ctx, xhci_add_endpoint,
359	TP_PROTO(struct xhci_ep_ctx *ctx),
360	TP_ARGS(ctx)
361);
362
363DECLARE_EVENT_CLASS(xhci_log_slot_ctx,
364	TP_PROTO(struct xhci_slot_ctx *ctx),
365	TP_ARGS(ctx),
366	TP_STRUCT__entry(
367		__field(u32, info)
368		__field(u32, info2)
369		__field(u32, tt_info)
370		__field(u32, state)
371	),
372	TP_fast_assign(
373		__entry->info = le32_to_cpu(ctx->dev_info);
374		__entry->info2 = le32_to_cpu(ctx->dev_info2);
375		__entry->tt_info = le64_to_cpu(ctx->tt_info);
376		__entry->state = le32_to_cpu(ctx->dev_state);
377	),
378	TP_printk("%s", xhci_decode_slot_context(__entry->info,
379			__entry->info2, __entry->tt_info,
380			__entry->state)
381	)
382);
383
384DEFINE_EVENT(xhci_log_slot_ctx, xhci_alloc_dev,
385	TP_PROTO(struct xhci_slot_ctx *ctx),
386	TP_ARGS(ctx)
387);
388
389DEFINE_EVENT(xhci_log_slot_ctx, xhci_free_dev,
390	TP_PROTO(struct xhci_slot_ctx *ctx),
391	TP_ARGS(ctx)
392);
393
394DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_disable_slot,
395	TP_PROTO(struct xhci_slot_ctx *ctx),
396	TP_ARGS(ctx)
397);
398
399DEFINE_EVENT(xhci_log_slot_ctx, xhci_discover_or_reset_device,
400	TP_PROTO(struct xhci_slot_ctx *ctx),
401	TP_ARGS(ctx)
402);
403
404DEFINE_EVENT(xhci_log_slot_ctx, xhci_setup_device_slot,
405	TP_PROTO(struct xhci_slot_ctx *ctx),
406	TP_ARGS(ctx)
407);
408
409DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_addr_dev,
410	TP_PROTO(struct xhci_slot_ctx *ctx),
411	TP_ARGS(ctx)
412);
413
414DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_reset_dev,
415	TP_PROTO(struct xhci_slot_ctx *ctx),
416	TP_ARGS(ctx)
417);
418
419DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_set_deq,
420	TP_PROTO(struct xhci_slot_ctx *ctx),
421	TP_ARGS(ctx)
422);
423
424DEFINE_EVENT(xhci_log_slot_ctx, xhci_configure_endpoint,
425	TP_PROTO(struct xhci_slot_ctx *ctx),
426	TP_ARGS(ctx)
427);
428
429DECLARE_EVENT_CLASS(xhci_log_ctrl_ctx,
430	TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
431	TP_ARGS(ctrl_ctx),
432	TP_STRUCT__entry(
433		__field(u32, drop)
434		__field(u32, add)
435	),
436	TP_fast_assign(
437		__entry->drop = le32_to_cpu(ctrl_ctx->drop_flags);
438		__entry->add = le32_to_cpu(ctrl_ctx->add_flags);
439	),
440	TP_printk("%s", xhci_decode_ctrl_ctx(__entry->drop, __entry->add)
441	)
442);
443
444DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_address_ctrl_ctx,
445	TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
446	TP_ARGS(ctrl_ctx)
447);
448
449DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_configure_endpoint_ctrl_ctx,
450	TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
451	TP_ARGS(ctrl_ctx)
452);
453
454DECLARE_EVENT_CLASS(xhci_log_ring,
455	TP_PROTO(struct xhci_ring *ring),
456	TP_ARGS(ring),
457	TP_STRUCT__entry(
458		__field(u32, type)
459		__field(void *, ring)
460		__field(dma_addr_t, enq)
461		__field(dma_addr_t, deq)
462		__field(dma_addr_t, enq_seg)
463		__field(dma_addr_t, deq_seg)
464		__field(unsigned int, num_segs)
465		__field(unsigned int, stream_id)
466		__field(unsigned int, cycle_state)
467		__field(unsigned int, num_trbs_free)
468		__field(unsigned int, bounce_buf_len)
469	),
470	TP_fast_assign(
471		__entry->ring = ring;
472		__entry->type = ring->type;
473		__entry->num_segs = ring->num_segs;
474		__entry->stream_id = ring->stream_id;
475		__entry->enq_seg = ring->enq_seg->dma;
476		__entry->deq_seg = ring->deq_seg->dma;
477		__entry->cycle_state = ring->cycle_state;
478		__entry->num_trbs_free = ring->num_trbs_free;
479		__entry->bounce_buf_len = ring->bounce_buf_len;
480		__entry->enq = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue);
481		__entry->deq = xhci_trb_virt_to_dma(ring->deq_seg, ring->dequeue);
482	),
483	TP_printk("%s %p: enq %pad(%pad) deq %pad(%pad) segs %d stream %d free_trbs %d bounce %d cycle %d",
484			xhci_ring_type_string(__entry->type), __entry->ring,
485			&__entry->enq, &__entry->enq_seg,
486			&__entry->deq, &__entry->deq_seg,
487			__entry->num_segs,
488			__entry->stream_id,
489			__entry->num_trbs_free,
490			__entry->bounce_buf_len,
491			__entry->cycle_state
492		)
493);
494
495DEFINE_EVENT(xhci_log_ring, xhci_ring_alloc,
496	TP_PROTO(struct xhci_ring *ring),
497	TP_ARGS(ring)
498);
499
500DEFINE_EVENT(xhci_log_ring, xhci_ring_free,
501	TP_PROTO(struct xhci_ring *ring),
502	TP_ARGS(ring)
503);
504
505DEFINE_EVENT(xhci_log_ring, xhci_ring_expansion,
506	TP_PROTO(struct xhci_ring *ring),
507	TP_ARGS(ring)
508);
509
510DEFINE_EVENT(xhci_log_ring, xhci_inc_enq,
511	TP_PROTO(struct xhci_ring *ring),
512	TP_ARGS(ring)
513);
514
515DEFINE_EVENT(xhci_log_ring, xhci_inc_deq,
516	TP_PROTO(struct xhci_ring *ring),
517	TP_ARGS(ring)
518);
519
520DECLARE_EVENT_CLASS(xhci_log_portsc,
521		    TP_PROTO(u32 portnum, u32 portsc),
522		    TP_ARGS(portnum, portsc),
523		    TP_STRUCT__entry(
 
524				     __field(u32, portnum)
525				     __field(u32, portsc)
526				     ),
527		    TP_fast_assign(
528				   __entry->portnum = portnum;
 
529				   __entry->portsc = portsc;
530				   ),
531		    TP_printk("port-%d: %s",
 
532			      __entry->portnum,
533			      xhci_decode_portsc(__entry->portsc)
534			      )
535);
536
537DEFINE_EVENT(xhci_log_portsc, xhci_handle_port_status,
538	     TP_PROTO(u32 portnum, u32 portsc),
539	     TP_ARGS(portnum, portsc)
540);
541
542DEFINE_EVENT(xhci_log_portsc, xhci_get_port_status,
543	     TP_PROTO(u32 portnum, u32 portsc),
544	     TP_ARGS(portnum, portsc)
545);
546
547DEFINE_EVENT(xhci_log_portsc, xhci_hub_status_data,
548	     TP_PROTO(u32 portnum, u32 portsc),
549	     TP_ARGS(portnum, portsc)
550);
551
552DECLARE_EVENT_CLASS(xhci_log_doorbell,
553	TP_PROTO(u32 slot, u32 doorbell),
554	TP_ARGS(slot, doorbell),
555	TP_STRUCT__entry(
556		__field(u32, slot)
557		__field(u32, doorbell)
558	),
559	TP_fast_assign(
560		__entry->slot = slot;
561		__entry->doorbell = doorbell;
562	),
563	TP_printk("Ring doorbell for %s",
564		xhci_decode_doorbell(__entry->slot, __entry->doorbell)
565	)
566);
567
568DEFINE_EVENT(xhci_log_doorbell, xhci_ring_ep_doorbell,
569	     TP_PROTO(u32 slot, u32 doorbell),
570	     TP_ARGS(slot, doorbell)
571);
572
573DEFINE_EVENT(xhci_log_doorbell, xhci_ring_host_doorbell,
574	     TP_PROTO(u32 slot, u32 doorbell),
575	     TP_ARGS(slot, doorbell)
576);
577
578DECLARE_EVENT_CLASS(xhci_dbc_log_request,
579	TP_PROTO(struct dbc_request *req),
580	TP_ARGS(req),
581	TP_STRUCT__entry(
582		__field(struct dbc_request *, req)
583		__field(bool, dir)
584		__field(unsigned int, actual)
585		__field(unsigned int, length)
586		__field(int, status)
587	),
588	TP_fast_assign(
589		__entry->req = req;
590		__entry->dir = req->direction;
591		__entry->actual = req->actual;
592		__entry->length = req->length;
593		__entry->status = req->status;
594	),
595	TP_printk("%s: req %p length %u/%u ==> %d",
596		__entry->dir ? "bulk-in" : "bulk-out",
597		__entry->req, __entry->actual,
598		__entry->length, __entry->status
599	)
600);
601
602DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_alloc_request,
603	TP_PROTO(struct dbc_request *req),
604	TP_ARGS(req)
605);
606
607DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_free_request,
608	TP_PROTO(struct dbc_request *req),
609	TP_ARGS(req)
610);
611
612DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_queue_request,
613	TP_PROTO(struct dbc_request *req),
614	TP_ARGS(req)
615);
616
617DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_giveback_request,
618	TP_PROTO(struct dbc_request *req),
619	TP_ARGS(req)
620);
621#endif /* __XHCI_TRACE_H */
622
623/* this part must be outside header guard */
624
625#undef TRACE_INCLUDE_PATH
626#define TRACE_INCLUDE_PATH .
627
628#undef TRACE_INCLUDE_FILE
629#define TRACE_INCLUDE_FILE xhci-trace
630
631#include <trace/define_trace.h>