Linux Audio

Check our new training course

Embedded Linux training

Mar 10-20, 2025, special US time zones
Register
Loading...
v5.4
  1/* SPDX-License-Identifier: GPL-2.0-only */
  2/*
  3 * auxtrace.h: AUX area trace support
  4 * Copyright (c) 2013-2015, Intel Corporation.
  5 */
  6
  7#ifndef __PERF_AUXTRACE_H
  8#define __PERF_AUXTRACE_H
  9
 10#include <sys/types.h>
 11#include <errno.h>
 12#include <stdbool.h>
 13#include <stddef.h>
 14#include <stdio.h> // FILE
 15#include <linux/list.h>
 16#include <linux/perf_event.h>
 17#include <linux/types.h>
 18#include <asm/bitsperlong.h>
 19#include <asm/barrier.h>
 20
 21union perf_event;
 22struct perf_session;
 23struct evlist;
 
 24struct perf_tool;
 25struct mmap;
 26struct perf_sample;
 27struct option;
 28struct record_opts;
 29struct perf_record_auxtrace_error;
 30struct perf_record_auxtrace_info;
 31struct events_stats;
 
 32
 33enum auxtrace_error_type {
 34       PERF_AUXTRACE_ERROR_ITRACE  = 1,
 35       PERF_AUXTRACE_ERROR_MAX
 36};
 37
 38/* Auxtrace records must have the same alignment as perf event records */
 39#define PERF_AUXTRACE_RECORD_ALIGNMENT 8
 40
 41enum auxtrace_type {
 42	PERF_AUXTRACE_UNKNOWN,
 43	PERF_AUXTRACE_INTEL_PT,
 44	PERF_AUXTRACE_INTEL_BTS,
 45	PERF_AUXTRACE_CS_ETM,
 46	PERF_AUXTRACE_ARM_SPE,
 47	PERF_AUXTRACE_S390_CPUMSF,
 48};
 49
 50enum itrace_period_type {
 51	PERF_ITRACE_PERIOD_INSTRUCTIONS,
 52	PERF_ITRACE_PERIOD_TICKS,
 53	PERF_ITRACE_PERIOD_NANOSECS,
 54};
 55
 
 
 
 
 
 56/**
 57 * struct itrace_synth_opts - AUX area tracing synthesis options.
 58 * @set: indicates whether or not options have been set
 59 * @default_no_sample: Default to no sampling.
 60 * @inject: indicates the event (not just the sample) must be fully synthesized
 61 *          because 'perf inject' will write it out
 62 * @instructions: whether to synthesize 'instructions' events
 63 * @branches: whether to synthesize 'branches' events
 
 64 * @transactions: whether to synthesize events for transactions
 65 * @ptwrites: whether to synthesize events for ptwrites
 66 * @pwr_events: whether to synthesize power events
 67 * @other_events: whether to synthesize other events recorded due to the use of
 68 *                aux_output
 69 * @errors: whether to synthesize decoder error events
 70 * @dont_decode: whether to skip decoding entirely
 71 * @log: write a decoding log
 72 * @calls: limit branch samples to calls (can be combined with @returns)
 73 * @returns: limit branch samples to returns (can be combined with @calls)
 74 * @callchain: add callchain to 'instructions' events
 
 75 * @thread_stack: feed branches to the thread_stack
 76 * @last_branch: add branch context to 'instruction' events
 
 
 
 
 
 
 
 
 
 
 77 * @callchain_sz: maximum callchain size
 78 * @last_branch_sz: branch context size
 79 * @period: 'instructions' events period
 80 * @period_type: 'instructions' events period type
 81 * @initial_skip: skip N events at the beginning.
 82 * @cpu_bitmap: CPUs for which to synthesize events, or NULL for all
 83 * @ptime_range: time intervals to trace or NULL
 84 * @range_num: number of time intervals to trace
 
 
 
 
 
 85 */
 86struct itrace_synth_opts {
 87	bool			set;
 88	bool			default_no_sample;
 89	bool			inject;
 90	bool			instructions;
 91	bool			branches;
 92	bool			transactions;
 93	bool			ptwrites;
 94	bool			pwr_events;
 95	bool			other_events;
 96	bool			errors;
 97	bool			dont_decode;
 98	bool			log;
 99	bool			calls;
100	bool			returns;
101	bool			callchain;
 
102	bool			thread_stack;
103	bool			last_branch;
 
 
 
 
 
 
 
 
 
 
104	unsigned int		callchain_sz;
105	unsigned int		last_branch_sz;
106	unsigned long long	period;
107	enum itrace_period_type	period_type;
108	unsigned long		initial_skip;
109	unsigned long		*cpu_bitmap;
110	struct perf_time_interval *ptime_range;
111	int			range_num;
 
 
 
 
 
112};
113
114/**
115 * struct auxtrace_index_entry - indexes a AUX area tracing event within a
116 *                               perf.data file.
117 * @file_offset: offset within the perf.data file
118 * @sz: size of the event
119 */
120struct auxtrace_index_entry {
121	u64			file_offset;
122	u64			sz;
123};
124
125#define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256
126
127/**
128 * struct auxtrace_index - index of AUX area tracing events within a perf.data
129 *                         file.
130 * @list: linking a number of arrays of entries
131 * @nr: number of entries
132 * @entries: array of entries
133 */
134struct auxtrace_index {
135	struct list_head	list;
136	size_t			nr;
137	struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT];
138};
139
140/**
141 * struct auxtrace - session callbacks to allow AUX area data decoding.
142 * @process_event: lets the decoder see all session events
143 * @process_auxtrace_event: process a PERF_RECORD_AUXTRACE event
 
 
 
144 * @flush_events: process any remaining data
145 * @free_events: free resources associated with event processing
146 * @free: free resources associated with the session
147 */
148struct auxtrace {
149	int (*process_event)(struct perf_session *session,
150			     union perf_event *event,
151			     struct perf_sample *sample,
152			     struct perf_tool *tool);
153	int (*process_auxtrace_event)(struct perf_session *session,
154				      union perf_event *event,
155				      struct perf_tool *tool);
 
 
 
 
 
156	int (*flush_events)(struct perf_session *session,
157			    struct perf_tool *tool);
158	void (*free_events)(struct perf_session *session);
159	void (*free)(struct perf_session *session);
 
 
160};
161
162/**
163 * struct auxtrace_buffer - a buffer containing AUX area tracing data.
164 * @list: buffers are queued in a list held by struct auxtrace_queue
165 * @size: size of the buffer in bytes
166 * @pid: in per-thread mode, the pid this buffer is associated with
167 * @tid: in per-thread mode, the tid this buffer is associated with
168 * @cpu: in per-cpu mode, the cpu this buffer is associated with
169 * @data: actual buffer data (can be null if the data has not been loaded)
170 * @data_offset: file offset at which the buffer can be read
171 * @mmap_addr: mmap address at which the buffer can be read
172 * @mmap_size: size of the mmap at @mmap_addr
173 * @data_needs_freeing: @data was malloc'd so free it when it is no longer
174 *                      needed
175 * @consecutive: the original data was split up and this buffer is consecutive
176 *               to the previous buffer
177 * @offset: offset as determined by aux_head / aux_tail members of struct
178 *          perf_event_mmap_page
179 * @reference: an implementation-specific reference determined when the data is
180 *             recorded
181 * @buffer_nr: used to number each buffer
182 * @use_size: implementation actually only uses this number of bytes
183 * @use_data: implementation actually only uses data starting at this address
184 */
185struct auxtrace_buffer {
186	struct list_head	list;
187	size_t			size;
188	pid_t			pid;
189	pid_t			tid;
190	int			cpu;
191	void			*data;
192	off_t			data_offset;
193	void			*mmap_addr;
194	size_t			mmap_size;
195	bool			data_needs_freeing;
196	bool			consecutive;
197	u64			offset;
198	u64			reference;
199	u64			buffer_nr;
200	size_t			use_size;
201	void			*use_data;
202};
203
204/**
205 * struct auxtrace_queue - a queue of AUX area tracing data buffers.
206 * @head: head of buffer list
207 * @tid: in per-thread mode, the tid this queue is associated with
208 * @cpu: in per-cpu mode, the cpu this queue is associated with
209 * @set: %true once this queue has been dedicated to a specific thread or cpu
210 * @priv: implementation-specific data
211 */
212struct auxtrace_queue {
213	struct list_head	head;
214	pid_t			tid;
215	int			cpu;
216	bool			set;
217	void			*priv;
218};
219
220/**
221 * struct auxtrace_queues - an array of AUX area tracing queues.
222 * @queue_array: array of queues
223 * @nr_queues: number of queues
224 * @new_data: set whenever new data is queued
225 * @populated: queues have been fully populated using the auxtrace_index
226 * @next_buffer_nr: used to number each buffer
227 */
228struct auxtrace_queues {
229	struct auxtrace_queue	*queue_array;
230	unsigned int		nr_queues;
231	bool			new_data;
232	bool			populated;
233	u64			next_buffer_nr;
234};
235
236/**
237 * struct auxtrace_heap_item - element of struct auxtrace_heap.
238 * @queue_nr: queue number
239 * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected
240 *           to be a timestamp
241 */
242struct auxtrace_heap_item {
243	unsigned int		queue_nr;
244	u64			ordinal;
245};
246
247/**
248 * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues.
249 * @heap_array: the heap
250 * @heap_cnt: the number of elements in the heap
251 * @heap_sz: maximum number of elements (grows as needed)
252 */
253struct auxtrace_heap {
254	struct auxtrace_heap_item	*heap_array;
255	unsigned int		heap_cnt;
256	unsigned int		heap_sz;
257};
258
259/**
260 * struct auxtrace_mmap - records an mmap of the auxtrace buffer.
261 * @base: address of mapped area
262 * @userpg: pointer to buffer's perf_event_mmap_page
263 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
264 * @len: size of mapped area
265 * @prev: previous aux_head
266 * @idx: index of this mmap
267 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
268 *       mmap) otherwise %0
269 * @cpu: cpu number for a per-cpu mmap otherwise %-1
270 */
271struct auxtrace_mmap {
272	void		*base;
273	void		*userpg;
274	size_t		mask;
275	size_t		len;
276	u64		prev;
277	int		idx;
278	pid_t		tid;
279	int		cpu;
280};
281
282/**
283 * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap.
284 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
285 * @offset: file offset of mapped area
286 * @len: size of mapped area
287 * @prot: mmap memory protection
288 * @idx: index of this mmap
289 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
290 *       mmap) otherwise %0
291 * @cpu: cpu number for a per-cpu mmap otherwise %-1
292 */
293struct auxtrace_mmap_params {
294	size_t		mask;
295	off_t		offset;
296	size_t		len;
297	int		prot;
298	int		idx;
299	pid_t		tid;
300	int		cpu;
301};
302
303/**
304 * struct auxtrace_record - callbacks for recording AUX area data.
305 * @recording_options: validate and process recording options
306 * @info_priv_size: return the size of the private data in auxtrace_info_event
307 * @info_fill: fill-in the private data in auxtrace_info_event
308 * @free: free this auxtrace record structure
309 * @snapshot_start: starting a snapshot
310 * @snapshot_finish: finishing a snapshot
311 * @find_snapshot: find data to snapshot within auxtrace mmap
312 * @parse_snapshot_options: parse snapshot options
313 * @reference: provide a 64-bit reference number for auxtrace_event
314 * @read_finish: called after reading from an auxtrace mmap
315 * @alignment: alignment (if any) for AUX area data
 
 
 
316 */
317struct auxtrace_record {
318	int (*recording_options)(struct auxtrace_record *itr,
319				 struct evlist *evlist,
320				 struct record_opts *opts);
321	size_t (*info_priv_size)(struct auxtrace_record *itr,
322				 struct evlist *evlist);
323	int (*info_fill)(struct auxtrace_record *itr,
324			 struct perf_session *session,
325			 struct perf_record_auxtrace_info *auxtrace_info,
326			 size_t priv_size);
327	void (*free)(struct auxtrace_record *itr);
328	int (*snapshot_start)(struct auxtrace_record *itr);
329	int (*snapshot_finish)(struct auxtrace_record *itr);
330	int (*find_snapshot)(struct auxtrace_record *itr, int idx,
331			     struct auxtrace_mmap *mm, unsigned char *data,
332			     u64 *head, u64 *old);
333	int (*parse_snapshot_options)(struct auxtrace_record *itr,
334				      struct record_opts *opts,
335				      const char *str);
336	u64 (*reference)(struct auxtrace_record *itr);
337	int (*read_finish)(struct auxtrace_record *itr, int idx);
338	unsigned int alignment;
 
 
 
339};
340
341/**
342 * struct addr_filter - address filter.
343 * @list: list node
344 * @range: true if it is a range filter
345 * @start: true if action is 'filter' or 'start'
346 * @action: 'filter', 'start' or 'stop' ('tracestop' is accepted but converted
347 *          to 'stop')
348 * @sym_from: symbol name for the filter address
349 * @sym_to: symbol name that determines the filter size
350 * @sym_from_idx: selects n'th from symbols with the same name (0 means global
351 *                and less than 0 means symbol must be unique)
352 * @sym_to_idx: same as @sym_from_idx but for @sym_to
353 * @addr: filter address
354 * @size: filter region size (for range filters)
355 * @filename: DSO file name or NULL for the kernel
356 * @str: allocated string that contains the other string members
357 */
358struct addr_filter {
359	struct list_head	list;
360	bool			range;
361	bool			start;
362	const char		*action;
363	const char		*sym_from;
364	const char		*sym_to;
365	int			sym_from_idx;
366	int			sym_to_idx;
367	u64			addr;
368	u64			size;
369	const char		*filename;
370	char			*str;
371};
372
373/**
374 * struct addr_filters - list of address filters.
375 * @head: list of address filters
376 * @cnt: number of address filters
377 */
378struct addr_filters {
379	struct list_head	head;
380	int			cnt;
381};
382
383struct auxtrace_cache;
384
385#ifdef HAVE_AUXTRACE_SUPPORT
386
387/*
388 * In snapshot mode the mmapped page is read-only which makes using
389 * __sync_val_compare_and_swap() problematic.  However, snapshot mode expects
390 * the buffer is not updated while the snapshot is made (e.g. Intel PT disables
391 * the event) so there is not a race anyway.
392 */
393static inline u64 auxtrace_mmap__read_snapshot_head(struct auxtrace_mmap *mm)
394{
395	struct perf_event_mmap_page *pc = mm->userpg;
396	u64 head = READ_ONCE(pc->aux_head);
397
398	/* Ensure all reads are done after we read the head */
399	rmb();
400	return head;
401}
402
403static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm)
404{
405	struct perf_event_mmap_page *pc = mm->userpg;
406#if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
407	u64 head = READ_ONCE(pc->aux_head);
408#else
409	u64 head = __sync_val_compare_and_swap(&pc->aux_head, 0, 0);
410#endif
411
412	/* Ensure all reads are done after we read the head */
413	rmb();
414	return head;
415}
416
417static inline void auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail)
418{
419	struct perf_event_mmap_page *pc = mm->userpg;
420#if BITS_PER_LONG != 64 && defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
421	u64 old_tail;
422#endif
423
424	/* Ensure all reads are done before we write the tail out */
425	mb();
426#if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
427	pc->aux_tail = tail;
428#else
429	do {
430		old_tail = __sync_val_compare_and_swap(&pc->aux_tail, 0, 0);
431	} while (!__sync_bool_compare_and_swap(&pc->aux_tail, old_tail, tail));
432#endif
433}
434
435int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
436			struct auxtrace_mmap_params *mp,
437			void *userpg, int fd);
438void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
439void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
440				off_t auxtrace_offset,
441				unsigned int auxtrace_pages,
442				bool auxtrace_overwrite);
443void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
444				   struct evlist *evlist, int idx,
445				   bool per_cpu);
446
447typedef int (*process_auxtrace_t)(struct perf_tool *tool,
448				  struct mmap *map,
449				  union perf_event *event, void *data1,
450				  size_t len1, void *data2, size_t len2);
451
452int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr,
453			struct perf_tool *tool, process_auxtrace_t fn);
454
455int auxtrace_mmap__read_snapshot(struct mmap *map,
456				 struct auxtrace_record *itr,
457				 struct perf_tool *tool, process_auxtrace_t fn,
458				 size_t snapshot_size);
459
460int auxtrace_queues__init(struct auxtrace_queues *queues);
461int auxtrace_queues__add_event(struct auxtrace_queues *queues,
462			       struct perf_session *session,
463			       union perf_event *event, off_t data_offset,
464			       struct auxtrace_buffer **buffer_ptr);
 
 
 
 
 
 
 
 
465void auxtrace_queues__free(struct auxtrace_queues *queues);
466int auxtrace_queues__process_index(struct auxtrace_queues *queues,
467				   struct perf_session *session);
 
 
468struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
469					      struct auxtrace_buffer *buffer);
470void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd);
 
 
 
 
471void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer);
472void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer);
473void auxtrace_buffer__free(struct auxtrace_buffer *buffer);
474
475int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr,
476		       u64 ordinal);
477void auxtrace_heap__pop(struct auxtrace_heap *heap);
478void auxtrace_heap__free(struct auxtrace_heap *heap);
479
480struct auxtrace_cache_entry {
481	struct hlist_node hash;
482	u32 key;
483};
484
485struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size,
486					   unsigned int limit_percent);
487void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache);
488void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c);
489void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry);
490int auxtrace_cache__add(struct auxtrace_cache *c, u32 key,
491			struct auxtrace_cache_entry *entry);
 
492void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key);
493
494struct auxtrace_record *auxtrace_record__init(struct evlist *evlist,
495					      int *err);
496
497int auxtrace_parse_snapshot_options(struct auxtrace_record *itr,
498				    struct record_opts *opts,
499				    const char *str);
 
 
 
 
500int auxtrace_record__options(struct auxtrace_record *itr,
501			     struct evlist *evlist,
502			     struct record_opts *opts);
503size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr,
504				       struct evlist *evlist);
505int auxtrace_record__info_fill(struct auxtrace_record *itr,
506			       struct perf_session *session,
507			       struct perf_record_auxtrace_info *auxtrace_info,
508			       size_t priv_size);
509void auxtrace_record__free(struct auxtrace_record *itr);
510int auxtrace_record__snapshot_start(struct auxtrace_record *itr);
511int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit);
512int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx,
513				   struct auxtrace_mmap *mm,
514				   unsigned char *data, u64 *head, u64 *old);
515u64 auxtrace_record__reference(struct auxtrace_record *itr);
 
516
517int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event,
518				   off_t file_offset);
519int auxtrace_index__write(int fd, struct list_head *head);
520int auxtrace_index__process(int fd, u64 size, struct perf_session *session,
521			    bool needs_swap);
522void auxtrace_index__free(struct list_head *head);
523
524void auxtrace_synth_error(struct perf_record_auxtrace_error *auxtrace_error, int type,
525			  int code, int cpu, pid_t pid, pid_t tid, u64 ip,
526			  const char *msg, u64 timestamp);
527
528int perf_event__process_auxtrace_info(struct perf_session *session,
529				      union perf_event *event);
530s64 perf_event__process_auxtrace(struct perf_session *session,
531				 union perf_event *event);
532int perf_event__process_auxtrace_error(struct perf_session *session,
533				       union perf_event *event);
 
 
534int itrace_parse_synth_opts(const struct option *opt, const char *str,
535			    int unset);
536void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts,
537				    bool no_sample);
538
539size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp);
540void perf_session__auxtrace_error_inc(struct perf_session *session,
541				      union perf_event *event);
542void events_stats__auxtrace_error_warn(const struct events_stats *stats);
543
544void addr_filters__init(struct addr_filters *filts);
545void addr_filters__exit(struct addr_filters *filts);
546int addr_filters__parse_bare_filter(struct addr_filters *filts,
547				    const char *filter);
548int auxtrace_parse_filters(struct evlist *evlist);
549
550int auxtrace__process_event(struct perf_session *session, union perf_event *event,
551			    struct perf_sample *sample, struct perf_tool *tool);
 
 
552int auxtrace__flush_events(struct perf_session *session, struct perf_tool *tool);
553void auxtrace__free_events(struct perf_session *session);
554void auxtrace__free(struct perf_session *session);
 
 
555
556#define ITRACE_HELP \
557"				i:	    		synthesize instructions events\n"		\
558"				b:	    		synthesize branches events\n"		\
559"				c:	    		synthesize branches events (calls only)\n"	\
560"				r:	    		synthesize branches events (returns only)\n" \
561"				x:	    		synthesize transactions events\n"		\
562"				w:	    		synthesize ptwrite events\n"		\
563"				p:	    		synthesize power events\n"			\
564"				e:	    		synthesize error events\n"			\
565"				d:	    		create a debug log\n"			\
 
 
 
 
 
 
 
 
 
 
 
566"				g[len]:     		synthesize a call chain (use with i or x)\n" \
 
567"				l[len]:     		synthesize last branch entries (use with i or x)\n" \
 
568"				sNUMBER:    		skip initial number of events\n"		\
 
569"				PERIOD[ns|us|ms|i|t]:   specify period to sample stream\n" \
570"				concatenate multiple options. Default is ibxwpe or cewp\n"
571
572static inline
573void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts,
574				       struct perf_time_interval *ptime_range,
575				       int range_num)
576{
577	opts->ptime_range = ptime_range;
578	opts->range_num = range_num;
579}
580
581static inline
582void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts)
583{
584	opts->ptime_range = NULL;
585	opts->range_num = 0;
586}
587
588#else
589#include "debug.h"
590
591static inline struct auxtrace_record *
592auxtrace_record__init(struct evlist *evlist __maybe_unused,
593		      int *err)
594{
595	*err = 0;
596	return NULL;
597}
598
599static inline
600void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused)
601{
602}
603
604static inline
605int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused,
606			     struct evlist *evlist __maybe_unused,
607			     struct record_opts *opts __maybe_unused)
608{
609	return 0;
610}
611
612#define perf_event__process_auxtrace_info		0
613#define perf_event__process_auxtrace			0
614#define perf_event__process_auxtrace_error		0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
615
616static inline
617void perf_session__auxtrace_error_inc(struct perf_session *session
618				      __maybe_unused,
619				      union perf_event *event
620				      __maybe_unused)
621{
622}
623
624static inline
625void events_stats__auxtrace_error_warn(const struct events_stats *stats
626				       __maybe_unused)
627{
628}
629
630static inline
 
 
 
 
 
 
 
 
631int itrace_parse_synth_opts(const struct option *opt __maybe_unused,
632			    const char *str __maybe_unused,
633			    int unset __maybe_unused)
634{
635	pr_err("AUX area tracing not supported\n");
636	return -EINVAL;
637}
638
639static inline
640int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused,
641				    struct record_opts *opts __maybe_unused,
642				    const char *str)
643{
644	if (!str)
645		return 0;
646	pr_err("AUX area tracing not supported\n");
647	return -EINVAL;
648}
649
650static inline
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
651int auxtrace__process_event(struct perf_session *session __maybe_unused,
652			    union perf_event *event __maybe_unused,
653			    struct perf_sample *sample __maybe_unused,
654			    struct perf_tool *tool __maybe_unused)
655{
656	return 0;
657}
658
659static inline
 
 
 
 
 
 
660int auxtrace__flush_events(struct perf_session *session __maybe_unused,
661			   struct perf_tool *tool __maybe_unused)
662{
663	return 0;
664}
665
666static inline
667void auxtrace__free_events(struct perf_session *session __maybe_unused)
668{
669}
670
671static inline
672void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused)
673{
674}
675
676static inline
677void auxtrace__free(struct perf_session *session __maybe_unused)
678{
679}
680
681static inline
682int auxtrace_index__write(int fd __maybe_unused,
683			  struct list_head *head __maybe_unused)
684{
685	return -EINVAL;
686}
687
688static inline
689int auxtrace_index__process(int fd __maybe_unused,
690			    u64 size __maybe_unused,
691			    struct perf_session *session __maybe_unused,
692			    bool needs_swap __maybe_unused)
693{
694	return -EINVAL;
695}
696
697static inline
698void auxtrace_index__free(struct list_head *head __maybe_unused)
699{
 
 
 
 
 
 
 
700}
701
702static inline
703int auxtrace_parse_filters(struct evlist *evlist __maybe_unused)
704{
705	return 0;
706}
707
708int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
709			struct auxtrace_mmap_params *mp,
710			void *userpg, int fd);
711void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
712void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
713				off_t auxtrace_offset,
714				unsigned int auxtrace_pages,
715				bool auxtrace_overwrite);
716void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
717				   struct evlist *evlist, int idx,
718				   bool per_cpu);
719
720#define ITRACE_HELP ""
721
722static inline
723void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts
724				       __maybe_unused,
725				       struct perf_time_interval *ptime_range
726				       __maybe_unused,
727				       int range_num __maybe_unused)
728{
729}
730
731static inline
732void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts
733					 __maybe_unused)
734{
735}
736
737#endif
738
739#endif
v5.14.15
  1/* SPDX-License-Identifier: GPL-2.0-only */
  2/*
  3 * auxtrace.h: AUX area trace support
  4 * Copyright (c) 2013-2015, Intel Corporation.
  5 */
  6
  7#ifndef __PERF_AUXTRACE_H
  8#define __PERF_AUXTRACE_H
  9
 10#include <sys/types.h>
 11#include <errno.h>
 12#include <stdbool.h>
 13#include <stddef.h>
 14#include <stdio.h> // FILE
 15#include <linux/list.h>
 16#include <linux/perf_event.h>
 17#include <linux/types.h>
 18#include <asm/bitsperlong.h>
 19#include <asm/barrier.h>
 20
 21union perf_event;
 22struct perf_session;
 23struct evlist;
 24struct evsel;
 25struct perf_tool;
 26struct mmap;
 27struct perf_sample;
 28struct option;
 29struct record_opts;
 30struct perf_record_auxtrace_error;
 31struct perf_record_auxtrace_info;
 32struct events_stats;
 33struct perf_pmu;
 34
 35enum auxtrace_error_type {
 36       PERF_AUXTRACE_ERROR_ITRACE  = 1,
 37       PERF_AUXTRACE_ERROR_MAX
 38};
 39
 40/* Auxtrace records must have the same alignment as perf event records */
 41#define PERF_AUXTRACE_RECORD_ALIGNMENT 8
 42
 43enum auxtrace_type {
 44	PERF_AUXTRACE_UNKNOWN,
 45	PERF_AUXTRACE_INTEL_PT,
 46	PERF_AUXTRACE_INTEL_BTS,
 47	PERF_AUXTRACE_CS_ETM,
 48	PERF_AUXTRACE_ARM_SPE,
 49	PERF_AUXTRACE_S390_CPUMSF,
 50};
 51
 52enum itrace_period_type {
 53	PERF_ITRACE_PERIOD_INSTRUCTIONS,
 54	PERF_ITRACE_PERIOD_TICKS,
 55	PERF_ITRACE_PERIOD_NANOSECS,
 56};
 57
 58#define AUXTRACE_ERR_FLG_OVERFLOW	(1 << ('o' - 'a'))
 59#define AUXTRACE_ERR_FLG_DATA_LOST	(1 << ('l' - 'a'))
 60
 61#define AUXTRACE_LOG_FLG_ALL_PERF_EVTS	(1 << ('a' - 'a'))
 62
 63/**
 64 * struct itrace_synth_opts - AUX area tracing synthesis options.
 65 * @set: indicates whether or not options have been set
 66 * @default_no_sample: Default to no sampling.
 67 * @inject: indicates the event (not just the sample) must be fully synthesized
 68 *          because 'perf inject' will write it out
 69 * @instructions: whether to synthesize 'instructions' events
 70 * @branches: whether to synthesize 'branches' events
 71 *            (branch misses only for Arm SPE)
 72 * @transactions: whether to synthesize events for transactions
 73 * @ptwrites: whether to synthesize events for ptwrites
 74 * @pwr_events: whether to synthesize power events
 75 * @other_events: whether to synthesize other events recorded due to the use of
 76 *                aux_output
 77 * @errors: whether to synthesize decoder error events
 78 * @dont_decode: whether to skip decoding entirely
 79 * @log: write a decoding log
 80 * @calls: limit branch samples to calls (can be combined with @returns)
 81 * @returns: limit branch samples to returns (can be combined with @calls)
 82 * @callchain: add callchain to 'instructions' events
 83 * @add_callchain: add callchain to existing event records
 84 * @thread_stack: feed branches to the thread_stack
 85 * @last_branch: add branch context to 'instruction' events
 86 * @add_last_branch: add branch context to existing event records
 87 * @flc: whether to synthesize first level cache events
 88 * @llc: whether to synthesize last level cache events
 89 * @tlb: whether to synthesize TLB events
 90 * @remote_access: whether to synthesize remote access events
 91 * @mem: whether to synthesize memory events
 92 * @timeless_decoding: prefer "timeless" decoding i.e. ignore timestamps
 93 * @vm_time_correlation: perform VM Time Correlation
 94 * @vm_tm_corr_dry_run: VM Time Correlation dry-run
 95 * @vm_tm_corr_args:  VM Time Correlation implementation-specific arguments
 96 * @callchain_sz: maximum callchain size
 97 * @last_branch_sz: branch context size
 98 * @period: 'instructions' events period
 99 * @period_type: 'instructions' events period type
100 * @initial_skip: skip N events at the beginning.
101 * @cpu_bitmap: CPUs for which to synthesize events, or NULL for all
102 * @ptime_range: time intervals to trace or NULL
103 * @range_num: number of time intervals to trace
104 * @error_plus_flags: flags to affect what errors are reported
105 * @error_minus_flags: flags to affect what errors are reported
106 * @log_plus_flags: flags to affect what is logged
107 * @log_minus_flags: flags to affect what is logged
108 * @quick: quicker (less detailed) decoding
109 */
110struct itrace_synth_opts {
111	bool			set;
112	bool			default_no_sample;
113	bool			inject;
114	bool			instructions;
115	bool			branches;
116	bool			transactions;
117	bool			ptwrites;
118	bool			pwr_events;
119	bool			other_events;
120	bool			errors;
121	bool			dont_decode;
122	bool			log;
123	bool			calls;
124	bool			returns;
125	bool			callchain;
126	bool			add_callchain;
127	bool			thread_stack;
128	bool			last_branch;
129	bool			add_last_branch;
130	bool			flc;
131	bool			llc;
132	bool			tlb;
133	bool			remote_access;
134	bool			mem;
135	bool			timeless_decoding;
136	bool			vm_time_correlation;
137	bool			vm_tm_corr_dry_run;
138	char			*vm_tm_corr_args;
139	unsigned int		callchain_sz;
140	unsigned int		last_branch_sz;
141	unsigned long long	period;
142	enum itrace_period_type	period_type;
143	unsigned long		initial_skip;
144	unsigned long		*cpu_bitmap;
145	struct perf_time_interval *ptime_range;
146	int			range_num;
147	unsigned int		error_plus_flags;
148	unsigned int		error_minus_flags;
149	unsigned int		log_plus_flags;
150	unsigned int		log_minus_flags;
151	unsigned int		quick;
152};
153
154/**
155 * struct auxtrace_index_entry - indexes a AUX area tracing event within a
156 *                               perf.data file.
157 * @file_offset: offset within the perf.data file
158 * @sz: size of the event
159 */
160struct auxtrace_index_entry {
161	u64			file_offset;
162	u64			sz;
163};
164
165#define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256
166
167/**
168 * struct auxtrace_index - index of AUX area tracing events within a perf.data
169 *                         file.
170 * @list: linking a number of arrays of entries
171 * @nr: number of entries
172 * @entries: array of entries
173 */
174struct auxtrace_index {
175	struct list_head	list;
176	size_t			nr;
177	struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT];
178};
179
180/**
181 * struct auxtrace - session callbacks to allow AUX area data decoding.
182 * @process_event: lets the decoder see all session events
183 * @process_auxtrace_event: process a PERF_RECORD_AUXTRACE event
184 * @queue_data: queue an AUX sample or PERF_RECORD_AUXTRACE event for later
185 *              processing
186 * @dump_auxtrace_sample: dump AUX area sample data
187 * @flush_events: process any remaining data
188 * @free_events: free resources associated with event processing
189 * @free: free resources associated with the session
190 */
191struct auxtrace {
192	int (*process_event)(struct perf_session *session,
193			     union perf_event *event,
194			     struct perf_sample *sample,
195			     struct perf_tool *tool);
196	int (*process_auxtrace_event)(struct perf_session *session,
197				      union perf_event *event,
198				      struct perf_tool *tool);
199	int (*queue_data)(struct perf_session *session,
200			  struct perf_sample *sample, union perf_event *event,
201			  u64 data_offset);
202	void (*dump_auxtrace_sample)(struct perf_session *session,
203				     struct perf_sample *sample);
204	int (*flush_events)(struct perf_session *session,
205			    struct perf_tool *tool);
206	void (*free_events)(struct perf_session *session);
207	void (*free)(struct perf_session *session);
208	bool (*evsel_is_auxtrace)(struct perf_session *session,
209				  struct evsel *evsel);
210};
211
212/**
213 * struct auxtrace_buffer - a buffer containing AUX area tracing data.
214 * @list: buffers are queued in a list held by struct auxtrace_queue
215 * @size: size of the buffer in bytes
216 * @pid: in per-thread mode, the pid this buffer is associated with
217 * @tid: in per-thread mode, the tid this buffer is associated with
218 * @cpu: in per-cpu mode, the cpu this buffer is associated with
219 * @data: actual buffer data (can be null if the data has not been loaded)
220 * @data_offset: file offset at which the buffer can be read
221 * @mmap_addr: mmap address at which the buffer can be read
222 * @mmap_size: size of the mmap at @mmap_addr
223 * @data_needs_freeing: @data was malloc'd so free it when it is no longer
224 *                      needed
225 * @consecutive: the original data was split up and this buffer is consecutive
226 *               to the previous buffer
227 * @offset: offset as determined by aux_head / aux_tail members of struct
228 *          perf_event_mmap_page
229 * @reference: an implementation-specific reference determined when the data is
230 *             recorded
231 * @buffer_nr: used to number each buffer
232 * @use_size: implementation actually only uses this number of bytes
233 * @use_data: implementation actually only uses data starting at this address
234 */
235struct auxtrace_buffer {
236	struct list_head	list;
237	size_t			size;
238	pid_t			pid;
239	pid_t			tid;
240	int			cpu;
241	void			*data;
242	off_t			data_offset;
243	void			*mmap_addr;
244	size_t			mmap_size;
245	bool			data_needs_freeing;
246	bool			consecutive;
247	u64			offset;
248	u64			reference;
249	u64			buffer_nr;
250	size_t			use_size;
251	void			*use_data;
252};
253
254/**
255 * struct auxtrace_queue - a queue of AUX area tracing data buffers.
256 * @head: head of buffer list
257 * @tid: in per-thread mode, the tid this queue is associated with
258 * @cpu: in per-cpu mode, the cpu this queue is associated with
259 * @set: %true once this queue has been dedicated to a specific thread or cpu
260 * @priv: implementation-specific data
261 */
262struct auxtrace_queue {
263	struct list_head	head;
264	pid_t			tid;
265	int			cpu;
266	bool			set;
267	void			*priv;
268};
269
270/**
271 * struct auxtrace_queues - an array of AUX area tracing queues.
272 * @queue_array: array of queues
273 * @nr_queues: number of queues
274 * @new_data: set whenever new data is queued
275 * @populated: queues have been fully populated using the auxtrace_index
276 * @next_buffer_nr: used to number each buffer
277 */
278struct auxtrace_queues {
279	struct auxtrace_queue	*queue_array;
280	unsigned int		nr_queues;
281	bool			new_data;
282	bool			populated;
283	u64			next_buffer_nr;
284};
285
286/**
287 * struct auxtrace_heap_item - element of struct auxtrace_heap.
288 * @queue_nr: queue number
289 * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected
290 *           to be a timestamp
291 */
292struct auxtrace_heap_item {
293	unsigned int		queue_nr;
294	u64			ordinal;
295};
296
297/**
298 * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues.
299 * @heap_array: the heap
300 * @heap_cnt: the number of elements in the heap
301 * @heap_sz: maximum number of elements (grows as needed)
302 */
303struct auxtrace_heap {
304	struct auxtrace_heap_item	*heap_array;
305	unsigned int		heap_cnt;
306	unsigned int		heap_sz;
307};
308
309/**
310 * struct auxtrace_mmap - records an mmap of the auxtrace buffer.
311 * @base: address of mapped area
312 * @userpg: pointer to buffer's perf_event_mmap_page
313 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
314 * @len: size of mapped area
315 * @prev: previous aux_head
316 * @idx: index of this mmap
317 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
318 *       mmap) otherwise %0
319 * @cpu: cpu number for a per-cpu mmap otherwise %-1
320 */
321struct auxtrace_mmap {
322	void		*base;
323	void		*userpg;
324	size_t		mask;
325	size_t		len;
326	u64		prev;
327	int		idx;
328	pid_t		tid;
329	int		cpu;
330};
331
332/**
333 * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap.
334 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
335 * @offset: file offset of mapped area
336 * @len: size of mapped area
337 * @prot: mmap memory protection
338 * @idx: index of this mmap
339 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
340 *       mmap) otherwise %0
341 * @cpu: cpu number for a per-cpu mmap otherwise %-1
342 */
343struct auxtrace_mmap_params {
344	size_t		mask;
345	off_t		offset;
346	size_t		len;
347	int		prot;
348	int		idx;
349	pid_t		tid;
350	int		cpu;
351};
352
353/**
354 * struct auxtrace_record - callbacks for recording AUX area data.
355 * @recording_options: validate and process recording options
356 * @info_priv_size: return the size of the private data in auxtrace_info_event
357 * @info_fill: fill-in the private data in auxtrace_info_event
358 * @free: free this auxtrace record structure
359 * @snapshot_start: starting a snapshot
360 * @snapshot_finish: finishing a snapshot
361 * @find_snapshot: find data to snapshot within auxtrace mmap
362 * @parse_snapshot_options: parse snapshot options
363 * @reference: provide a 64-bit reference number for auxtrace_event
364 * @read_finish: called after reading from an auxtrace mmap
365 * @alignment: alignment (if any) for AUX area data
366 * @default_aux_sample_size: default sample size for --aux sample option
367 * @pmu: associated pmu
368 * @evlist: selected events list
369 */
370struct auxtrace_record {
371	int (*recording_options)(struct auxtrace_record *itr,
372				 struct evlist *evlist,
373				 struct record_opts *opts);
374	size_t (*info_priv_size)(struct auxtrace_record *itr,
375				 struct evlist *evlist);
376	int (*info_fill)(struct auxtrace_record *itr,
377			 struct perf_session *session,
378			 struct perf_record_auxtrace_info *auxtrace_info,
379			 size_t priv_size);
380	void (*free)(struct auxtrace_record *itr);
381	int (*snapshot_start)(struct auxtrace_record *itr);
382	int (*snapshot_finish)(struct auxtrace_record *itr);
383	int (*find_snapshot)(struct auxtrace_record *itr, int idx,
384			     struct auxtrace_mmap *mm, unsigned char *data,
385			     u64 *head, u64 *old);
386	int (*parse_snapshot_options)(struct auxtrace_record *itr,
387				      struct record_opts *opts,
388				      const char *str);
389	u64 (*reference)(struct auxtrace_record *itr);
390	int (*read_finish)(struct auxtrace_record *itr, int idx);
391	unsigned int alignment;
392	unsigned int default_aux_sample_size;
393	struct perf_pmu *pmu;
394	struct evlist *evlist;
395};
396
397/**
398 * struct addr_filter - address filter.
399 * @list: list node
400 * @range: true if it is a range filter
401 * @start: true if action is 'filter' or 'start'
402 * @action: 'filter', 'start' or 'stop' ('tracestop' is accepted but converted
403 *          to 'stop')
404 * @sym_from: symbol name for the filter address
405 * @sym_to: symbol name that determines the filter size
406 * @sym_from_idx: selects n'th from symbols with the same name (0 means global
407 *                and less than 0 means symbol must be unique)
408 * @sym_to_idx: same as @sym_from_idx but for @sym_to
409 * @addr: filter address
410 * @size: filter region size (for range filters)
411 * @filename: DSO file name or NULL for the kernel
412 * @str: allocated string that contains the other string members
413 */
414struct addr_filter {
415	struct list_head	list;
416	bool			range;
417	bool			start;
418	const char		*action;
419	const char		*sym_from;
420	const char		*sym_to;
421	int			sym_from_idx;
422	int			sym_to_idx;
423	u64			addr;
424	u64			size;
425	const char		*filename;
426	char			*str;
427};
428
429/**
430 * struct addr_filters - list of address filters.
431 * @head: list of address filters
432 * @cnt: number of address filters
433 */
434struct addr_filters {
435	struct list_head	head;
436	int			cnt;
437};
438
439struct auxtrace_cache;
440
441#ifdef HAVE_AUXTRACE_SUPPORT
442
443/*
444 * In snapshot mode the mmapped page is read-only which makes using
445 * __sync_val_compare_and_swap() problematic.  However, snapshot mode expects
446 * the buffer is not updated while the snapshot is made (e.g. Intel PT disables
447 * the event) so there is not a race anyway.
448 */
449static inline u64 auxtrace_mmap__read_snapshot_head(struct auxtrace_mmap *mm)
450{
451	struct perf_event_mmap_page *pc = mm->userpg;
452	u64 head = READ_ONCE(pc->aux_head);
453
454	/* Ensure all reads are done after we read the head */
455	smp_rmb();
456	return head;
457}
458
459static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm)
460{
461	struct perf_event_mmap_page *pc = mm->userpg;
462#if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
463	u64 head = READ_ONCE(pc->aux_head);
464#else
465	u64 head = __sync_val_compare_and_swap(&pc->aux_head, 0, 0);
466#endif
467
468	/* Ensure all reads are done after we read the head */
469	smp_rmb();
470	return head;
471}
472
473static inline void auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail)
474{
475	struct perf_event_mmap_page *pc = mm->userpg;
476#if BITS_PER_LONG != 64 && defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
477	u64 old_tail;
478#endif
479
480	/* Ensure all reads are done before we write the tail out */
481	smp_mb();
482#if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
483	pc->aux_tail = tail;
484#else
485	do {
486		old_tail = __sync_val_compare_and_swap(&pc->aux_tail, 0, 0);
487	} while (!__sync_bool_compare_and_swap(&pc->aux_tail, old_tail, tail));
488#endif
489}
490
491int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
492			struct auxtrace_mmap_params *mp,
493			void *userpg, int fd);
494void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
495void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
496				off_t auxtrace_offset,
497				unsigned int auxtrace_pages,
498				bool auxtrace_overwrite);
499void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
500				   struct evlist *evlist, int idx,
501				   bool per_cpu);
502
503typedef int (*process_auxtrace_t)(struct perf_tool *tool,
504				  struct mmap *map,
505				  union perf_event *event, void *data1,
506				  size_t len1, void *data2, size_t len2);
507
508int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr,
509			struct perf_tool *tool, process_auxtrace_t fn);
510
511int auxtrace_mmap__read_snapshot(struct mmap *map,
512				 struct auxtrace_record *itr,
513				 struct perf_tool *tool, process_auxtrace_t fn,
514				 size_t snapshot_size);
515
516int auxtrace_queues__init(struct auxtrace_queues *queues);
517int auxtrace_queues__add_event(struct auxtrace_queues *queues,
518			       struct perf_session *session,
519			       union perf_event *event, off_t data_offset,
520			       struct auxtrace_buffer **buffer_ptr);
521struct auxtrace_queue *
522auxtrace_queues__sample_queue(struct auxtrace_queues *queues,
523			      struct perf_sample *sample,
524			      struct perf_session *session);
525int auxtrace_queues__add_sample(struct auxtrace_queues *queues,
526				struct perf_session *session,
527				struct perf_sample *sample, u64 data_offset,
528				u64 reference);
529void auxtrace_queues__free(struct auxtrace_queues *queues);
530int auxtrace_queues__process_index(struct auxtrace_queues *queues,
531				   struct perf_session *session);
532int auxtrace_queue_data(struct perf_session *session, bool samples,
533			bool events);
534struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
535					      struct auxtrace_buffer *buffer);
536void *auxtrace_buffer__get_data_rw(struct auxtrace_buffer *buffer, int fd, bool rw);
537static inline void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd)
538{
539	return auxtrace_buffer__get_data_rw(buffer, fd, false);
540}
541void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer);
542void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer);
543void auxtrace_buffer__free(struct auxtrace_buffer *buffer);
544
545int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr,
546		       u64 ordinal);
547void auxtrace_heap__pop(struct auxtrace_heap *heap);
548void auxtrace_heap__free(struct auxtrace_heap *heap);
549
550struct auxtrace_cache_entry {
551	struct hlist_node hash;
552	u32 key;
553};
554
555struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size,
556					   unsigned int limit_percent);
557void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache);
558void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c);
559void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry);
560int auxtrace_cache__add(struct auxtrace_cache *c, u32 key,
561			struct auxtrace_cache_entry *entry);
562void auxtrace_cache__remove(struct auxtrace_cache *c, u32 key);
563void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key);
564
565struct auxtrace_record *auxtrace_record__init(struct evlist *evlist,
566					      int *err);
567
568int auxtrace_parse_snapshot_options(struct auxtrace_record *itr,
569				    struct record_opts *opts,
570				    const char *str);
571int auxtrace_parse_sample_options(struct auxtrace_record *itr,
572				  struct evlist *evlist,
573				  struct record_opts *opts, const char *str);
574void auxtrace_regroup_aux_output(struct evlist *evlist);
575int auxtrace_record__options(struct auxtrace_record *itr,
576			     struct evlist *evlist,
577			     struct record_opts *opts);
578size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr,
579				       struct evlist *evlist);
580int auxtrace_record__info_fill(struct auxtrace_record *itr,
581			       struct perf_session *session,
582			       struct perf_record_auxtrace_info *auxtrace_info,
583			       size_t priv_size);
584void auxtrace_record__free(struct auxtrace_record *itr);
585int auxtrace_record__snapshot_start(struct auxtrace_record *itr);
586int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit);
587int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx,
588				   struct auxtrace_mmap *mm,
589				   unsigned char *data, u64 *head, u64 *old);
590u64 auxtrace_record__reference(struct auxtrace_record *itr);
591int auxtrace_record__read_finish(struct auxtrace_record *itr, int idx);
592
593int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event,
594				   off_t file_offset);
595int auxtrace_index__write(int fd, struct list_head *head);
596int auxtrace_index__process(int fd, u64 size, struct perf_session *session,
597			    bool needs_swap);
598void auxtrace_index__free(struct list_head *head);
599
600void auxtrace_synth_error(struct perf_record_auxtrace_error *auxtrace_error, int type,
601			  int code, int cpu, pid_t pid, pid_t tid, u64 ip,
602			  const char *msg, u64 timestamp);
603
604int perf_event__process_auxtrace_info(struct perf_session *session,
605				      union perf_event *event);
606s64 perf_event__process_auxtrace(struct perf_session *session,
607				 union perf_event *event);
608int perf_event__process_auxtrace_error(struct perf_session *session,
609				       union perf_event *event);
610int itrace_do_parse_synth_opts(struct itrace_synth_opts *synth_opts,
611			       const char *str, int unset);
612int itrace_parse_synth_opts(const struct option *opt, const char *str,
613			    int unset);
614void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts,
615				    bool no_sample);
616
617size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp);
618void perf_session__auxtrace_error_inc(struct perf_session *session,
619				      union perf_event *event);
620void events_stats__auxtrace_error_warn(const struct events_stats *stats);
621
622void addr_filters__init(struct addr_filters *filts);
623void addr_filters__exit(struct addr_filters *filts);
624int addr_filters__parse_bare_filter(struct addr_filters *filts,
625				    const char *filter);
626int auxtrace_parse_filters(struct evlist *evlist);
627
628int auxtrace__process_event(struct perf_session *session, union perf_event *event,
629			    struct perf_sample *sample, struct perf_tool *tool);
630void auxtrace__dump_auxtrace_sample(struct perf_session *session,
631				    struct perf_sample *sample);
632int auxtrace__flush_events(struct perf_session *session, struct perf_tool *tool);
633void auxtrace__free_events(struct perf_session *session);
634void auxtrace__free(struct perf_session *session);
635bool auxtrace__evsel_is_auxtrace(struct perf_session *session,
636				 struct evsel *evsel);
637
638#define ITRACE_HELP \
639"				i[period]:    		synthesize instructions events\n" \
640"				b:	    		synthesize branches events (branch misses for Arm SPE)\n" \
641"				c:	    		synthesize branches events (calls only)\n"	\
642"				r:	    		synthesize branches events (returns only)\n" \
643"				x:	    		synthesize transactions events\n"		\
644"				w:	    		synthesize ptwrite events\n"		\
645"				p:	    		synthesize power events\n"			\
646"				o:			synthesize other events recorded due to the use\n" \
647"							of aux-output (refer to perf record)\n"	\
648"				e[flags]:		synthesize error events\n" \
649"							each flag must be preceded by + or -\n" \
650"							error flags are: o (overflow)\n" \
651"									 l (data lost)\n" \
652"				d[flags]:		create a debug log\n" \
653"							each flag must be preceded by + or -\n" \
654"							log flags are: a (all perf events)\n" \
655"				f:	    		synthesize first level cache events\n" \
656"				m:	    		synthesize last level cache events\n" \
657"				t:	    		synthesize TLB events\n" \
658"				a:	    		synthesize remote access events\n" \
659"				g[len]:     		synthesize a call chain (use with i or x)\n" \
660"				G[len]:			synthesize a call chain on existing event records\n" \
661"				l[len]:     		synthesize last branch entries (use with i or x)\n" \
662"				L[len]:			synthesize last branch entries on existing event records\n" \
663"				sNUMBER:    		skip initial number of events\n"		\
664"				q:			quicker (less detailed) decoding\n" \
665"				PERIOD[ns|us|ms|i|t]:   specify period to sample stream\n" \
666"				concatenate multiple options. Default is ibxwpe or cewp\n"
667
668static inline
669void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts,
670				       struct perf_time_interval *ptime_range,
671				       int range_num)
672{
673	opts->ptime_range = ptime_range;
674	opts->range_num = range_num;
675}
676
677static inline
678void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts)
679{
680	opts->ptime_range = NULL;
681	opts->range_num = 0;
682}
683
684#else
685#include "debug.h"
686
687static inline struct auxtrace_record *
688auxtrace_record__init(struct evlist *evlist __maybe_unused,
689		      int *err)
690{
691	*err = 0;
692	return NULL;
693}
694
695static inline
696void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused)
697{
698}
699
700static inline
701int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused,
702			     struct evlist *evlist __maybe_unused,
703			     struct record_opts *opts __maybe_unused)
704{
705	return 0;
706}
707
708static inline
709int perf_event__process_auxtrace_info(struct perf_session *session __maybe_unused,
710				      union perf_event *event __maybe_unused)
711{
712	return 0;
713}
714
715static inline
716s64 perf_event__process_auxtrace(struct perf_session *session __maybe_unused,
717				 union perf_event *event __maybe_unused)
718{
719	return 0;
720}
721
722static inline
723int perf_event__process_auxtrace_error(struct perf_session *session __maybe_unused,
724				       union perf_event *event __maybe_unused)
725{
726	return 0;
727}
728
729static inline
730void perf_session__auxtrace_error_inc(struct perf_session *session
731				      __maybe_unused,
732				      union perf_event *event
733				      __maybe_unused)
734{
735}
736
737static inline
738void events_stats__auxtrace_error_warn(const struct events_stats *stats
739				       __maybe_unused)
740{
741}
742
743static inline
744int itrace_do_parse_synth_opts(struct itrace_synth_opts *synth_opts __maybe_unused,
745			       const char *str __maybe_unused, int unset __maybe_unused)
746{
747	pr_err("AUX area tracing not supported\n");
748	return -EINVAL;
749}
750
751static inline
752int itrace_parse_synth_opts(const struct option *opt __maybe_unused,
753			    const char *str __maybe_unused,
754			    int unset __maybe_unused)
755{
756	pr_err("AUX area tracing not supported\n");
757	return -EINVAL;
758}
759
760static inline
761int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused,
762				    struct record_opts *opts __maybe_unused,
763				    const char *str)
764{
765	if (!str)
766		return 0;
767	pr_err("AUX area tracing not supported\n");
768	return -EINVAL;
769}
770
771static inline
772int auxtrace_parse_sample_options(struct auxtrace_record *itr __maybe_unused,
773				  struct evlist *evlist __maybe_unused,
774				  struct record_opts *opts __maybe_unused,
775				  const char *str)
776{
777	if (!str)
778		return 0;
779	pr_err("AUX area tracing not supported\n");
780	return -EINVAL;
781}
782
783static inline
784void auxtrace_regroup_aux_output(struct evlist *evlist __maybe_unused)
785{
786}
787
788static inline
789int auxtrace__process_event(struct perf_session *session __maybe_unused,
790			    union perf_event *event __maybe_unused,
791			    struct perf_sample *sample __maybe_unused,
792			    struct perf_tool *tool __maybe_unused)
793{
794	return 0;
795}
796
797static inline
798void auxtrace__dump_auxtrace_sample(struct perf_session *session __maybe_unused,
799				    struct perf_sample *sample __maybe_unused)
800{
801}
802
803static inline
804int auxtrace__flush_events(struct perf_session *session __maybe_unused,
805			   struct perf_tool *tool __maybe_unused)
806{
807	return 0;
808}
809
810static inline
811void auxtrace__free_events(struct perf_session *session __maybe_unused)
812{
813}
814
815static inline
816void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused)
817{
818}
819
820static inline
821void auxtrace__free(struct perf_session *session __maybe_unused)
822{
823}
824
825static inline
826int auxtrace_index__write(int fd __maybe_unused,
827			  struct list_head *head __maybe_unused)
828{
829	return -EINVAL;
830}
831
832static inline
833int auxtrace_index__process(int fd __maybe_unused,
834			    u64 size __maybe_unused,
835			    struct perf_session *session __maybe_unused,
836			    bool needs_swap __maybe_unused)
837{
838	return -EINVAL;
839}
840
841static inline
842void auxtrace_index__free(struct list_head *head __maybe_unused)
843{
844}
845
846static inline
847bool auxtrace__evsel_is_auxtrace(struct perf_session *session __maybe_unused,
848				 struct evsel *evsel __maybe_unused)
849{
850	return false;
851}
852
853static inline
854int auxtrace_parse_filters(struct evlist *evlist __maybe_unused)
855{
856	return 0;
857}
858
859int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
860			struct auxtrace_mmap_params *mp,
861			void *userpg, int fd);
862void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
863void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
864				off_t auxtrace_offset,
865				unsigned int auxtrace_pages,
866				bool auxtrace_overwrite);
867void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
868				   struct evlist *evlist, int idx,
869				   bool per_cpu);
870
871#define ITRACE_HELP ""
872
873static inline
874void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts
875				       __maybe_unused,
876				       struct perf_time_interval *ptime_range
877				       __maybe_unused,
878				       int range_num __maybe_unused)
879{
880}
881
882static inline
883void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts
884					 __maybe_unused)
885{
886}
887
888#endif
889
890#endif