Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.15.
   1// SPDX-License-Identifier: GPL-2.0-only
   2#define _GNU_SOURCE
   3
   4#include <arpa/inet.h>
   5#include <bpf/bpf.h>
   6#include <bpf/libbpf.h>
   7#include <errno.h>
   8#include <fcntl.h>
   9#include <getopt.h>
  10#include <linux/ethtool.h>
  11#include <linux/hashtable.h>
  12#include <linux/if_link.h>
  13#include <linux/jhash.h>
  14#include <linux/limits.h>
  15#include <linux/list.h>
  16#include <linux/sockios.h>
  17#include <locale.h>
  18#include <math.h>
  19#include <net/if.h>
  20#include <poll.h>
  21#include <signal.h>
  22#include <stdbool.h>
  23#include <stdio.h>
  24#include <stdlib.h>
  25#include <string.h>
  26#include <sys/ioctl.h>
  27#include <sys/mman.h>
  28#include <sys/signalfd.h>
  29#include <sys/sysinfo.h>
  30#include <sys/timerfd.h>
  31#include <sys/utsname.h>
  32#include <time.h>
  33#include <unistd.h>
  34
  35#include "bpf_util.h"
  36#include "xdp_sample_user.h"
  37
  38#define __sample_print(fmt, cond, ...)                                         \
  39	({                                                                     \
  40		if (cond)                                                      \
  41			printf(fmt, ##__VA_ARGS__);                            \
  42	})
  43
  44#define print_always(fmt, ...) __sample_print(fmt, 1, ##__VA_ARGS__)
  45#define print_default(fmt, ...)                                                \
  46	__sample_print(fmt, sample_log_level & LL_DEFAULT, ##__VA_ARGS__)
  47#define __print_err(err, fmt, ...)                                             \
  48	({                                                                     \
  49		__sample_print(fmt, err > 0 || sample_log_level & LL_DEFAULT,  \
  50			       ##__VA_ARGS__);                                 \
  51		sample_err_exp = sample_err_exp ? true : err > 0;              \
  52	})
  53#define print_err(err, fmt, ...) __print_err(err, fmt, ##__VA_ARGS__)
  54
  55#define __COLUMN(x) "%'10" x " %-13s"
  56#define FMT_COLUMNf __COLUMN(".0f")
  57#define FMT_COLUMNd __COLUMN("d")
  58#define FMT_COLUMNl __COLUMN("llu")
  59#define RX(rx) rx, "rx/s"
  60#define PPS(pps) pps, "pkt/s"
  61#define DROP(drop) drop, "drop/s"
  62#define ERR(err) err, "error/s"
  63#define HITS(hits) hits, "hit/s"
  64#define XMIT(xmit) xmit, "xmit/s"
  65#define PASS(pass) pass, "pass/s"
  66#define REDIR(redir) redir, "redir/s"
  67#define NANOSEC_PER_SEC 1000000000 /* 10^9 */
  68
  69#define XDP_UNKNOWN (XDP_REDIRECT + 1)
  70#define XDP_ACTION_MAX (XDP_UNKNOWN + 1)
  71#define XDP_REDIRECT_ERR_MAX 7
  72
  73enum map_type {
  74	MAP_RX,
  75	MAP_REDIRECT_ERR,
  76	MAP_CPUMAP_ENQUEUE,
  77	MAP_CPUMAP_KTHREAD,
  78	MAP_EXCEPTION,
  79	MAP_DEVMAP_XMIT,
  80	MAP_DEVMAP_XMIT_MULTI,
  81	NUM_MAP,
  82};
  83
  84enum log_level {
  85	LL_DEFAULT = 1U << 0,
  86	LL_SIMPLE = 1U << 1,
  87	LL_DEBUG = 1U << 2,
  88};
  89
  90struct record {
  91	__u64 timestamp;
  92	struct datarec total;
  93	struct datarec *cpu;
  94};
  95
  96struct map_entry {
  97	struct hlist_node node;
  98	__u64 pair;
  99	struct record val;
 100};
 101
 102struct stats_record {
 103	struct record rx_cnt;
 104	struct record redir_err[XDP_REDIRECT_ERR_MAX];
 105	struct record kthread;
 106	struct record exception[XDP_ACTION_MAX];
 107	struct record devmap_xmit;
 108	DECLARE_HASHTABLE(xmit_map, 5);
 109	struct record enq[];
 110};
 111
 112struct sample_output {
 113	struct {
 114		__u64 rx;
 115		__u64 redir;
 116		__u64 drop;
 117		__u64 drop_xmit;
 118		__u64 err;
 119		__u64 xmit;
 120	} totals;
 121	struct {
 122		union {
 123			__u64 pps;
 124			__u64 num;
 125		};
 126		__u64 drop;
 127		__u64 err;
 128	} rx_cnt;
 129	struct {
 130		__u64 suc;
 131		__u64 err;
 132	} redir_cnt;
 133	struct {
 134		__u64 hits;
 135	} except_cnt;
 136	struct {
 137		__u64 pps;
 138		__u64 drop;
 139		__u64 err;
 140		double bavg;
 141	} xmit_cnt;
 142};
 143
 144struct xdp_desc {
 145	int ifindex;
 146	__u32 prog_id;
 147	int flags;
 148} sample_xdp_progs[32];
 149
 150struct datarec *sample_mmap[NUM_MAP];
 151struct bpf_map *sample_map[NUM_MAP];
 152size_t sample_map_count[NUM_MAP];
 153enum log_level sample_log_level;
 154struct sample_output sample_out;
 155unsigned long sample_interval;
 156bool sample_err_exp;
 157int sample_xdp_cnt;
 158int sample_n_cpus;
 159int sample_sig_fd;
 160int sample_mask;
 161
 162static const char *xdp_redirect_err_names[XDP_REDIRECT_ERR_MAX] = {
 163	/* Key=1 keeps unknown errors */
 164	"Success",
 165	"Unknown",
 166	"EINVAL",
 167	"ENETDOWN",
 168	"EMSGSIZE",
 169	"EOPNOTSUPP",
 170	"ENOSPC",
 171};
 172
 173/* Keyed from Unknown */
 174static const char *xdp_redirect_err_help[XDP_REDIRECT_ERR_MAX - 1] = {
 175	"Unknown error",
 176	"Invalid redirection",
 177	"Device being redirected to is down",
 178	"Packet length too large for device",
 179	"Operation not supported",
 180	"No space in ptr_ring of cpumap kthread",
 181};
 182
 183static const char *xdp_action_names[XDP_ACTION_MAX] = {
 184	[XDP_ABORTED]  = "XDP_ABORTED",
 185	[XDP_DROP]     = "XDP_DROP",
 186	[XDP_PASS]     = "XDP_PASS",
 187	[XDP_TX]       = "XDP_TX",
 188	[XDP_REDIRECT] = "XDP_REDIRECT",
 189	[XDP_UNKNOWN]  = "XDP_UNKNOWN",
 190};
 191
 192static __u64 gettime(void)
 193{
 194	struct timespec t;
 195	int res;
 196
 197	res = clock_gettime(CLOCK_MONOTONIC, &t);
 198	if (res < 0) {
 199		fprintf(stderr, "Error with gettimeofday! (%i)\n", res);
 200		return UINT64_MAX;
 201	}
 202	return (__u64)t.tv_sec * NANOSEC_PER_SEC + t.tv_nsec;
 203}
 204
 205static const char *action2str(int action)
 206{
 207	if (action < XDP_ACTION_MAX)
 208		return xdp_action_names[action];
 209	return NULL;
 210}
 211
 212static void sample_print_help(int mask)
 213{
 214	printf("Output format description\n\n"
 215	       "By default, redirect success statistics are disabled, use -s to enable.\n"
 216	       "The terse output mode is default, verbose mode can be activated using -v\n"
 217	       "Use SIGQUIT (Ctrl + \\) to switch the mode dynamically at runtime\n\n"
 218	       "Terse mode displays at most the following fields:\n"
 219	       "  rx/s        Number of packets received per second\n"
 220	       "  redir/s     Number of packets successfully redirected per second\n"
 221	       "  err,drop/s  Aggregated count of errors per second (including dropped packets)\n"
 222	       "  xmit/s      Number of packets transmitted on the output device per second\n\n"
 223	       "Output description for verbose mode:\n"
 224	       "  FIELD                 DESCRIPTION\n");
 225
 226	if (mask & SAMPLE_RX_CNT) {
 227		printf("  receive\t\tDisplays the number of packets received & errors encountered\n"
 228		       " \t\t\tWhenever an error or packet drop occurs, details of per CPU error\n"
 229		       " \t\t\tand drop statistics will be expanded inline in terse mode.\n"
 230		       " \t\t\t\tpkt/s     - Packets received per second\n"
 231		       " \t\t\t\tdrop/s    - Packets dropped per second\n"
 232		       " \t\t\t\terror/s   - Errors encountered per second\n\n");
 233	}
 234	if (mask & (SAMPLE_REDIRECT_CNT | SAMPLE_REDIRECT_ERR_CNT)) {
 235		printf("  redirect\t\tDisplays the number of packets successfully redirected\n"
 236		       "  \t\t\tErrors encountered are expanded under redirect_err field\n"
 237		       "  \t\t\tNote that passing -s to enable it has a per packet overhead\n"
 238		       "  \t\t\t\tredir/s   - Packets redirected successfully per second\n\n"
 239		       "  redirect_err\t\tDisplays the number of packets that failed redirection\n"
 240		       "  \t\t\tThe errno is expanded under this field with per CPU count\n"
 241		       "  \t\t\tThe recognized errors are:\n");
 242
 243		for (int i = 2; i < XDP_REDIRECT_ERR_MAX; i++)
 244			printf("\t\t\t  %s: %s\n", xdp_redirect_err_names[i],
 245			       xdp_redirect_err_help[i - 1]);
 246
 247		printf("  \n\t\t\t\terror/s   - Packets that failed redirection per second\n\n");
 248	}
 249
 250	if (mask & SAMPLE_CPUMAP_ENQUEUE_CNT) {
 251		printf("  enqueue to cpu N\tDisplays the number of packets enqueued to bulk queue of CPU N\n"
 252		       "  \t\t\tExpands to cpu:FROM->N to display enqueue stats for each CPU enqueuing to CPU N\n"
 253		       "  \t\t\tReceived packets can be associated with the CPU redirect program is enqueuing \n"
 254		       "  \t\t\tpackets to.\n"
 255		       "  \t\t\t\tpkt/s    - Packets enqueued per second from other CPU to CPU N\n"
 256		       "  \t\t\t\tdrop/s   - Packets dropped when trying to enqueue to CPU N\n"
 257		       "  \t\t\t\tbulk-avg - Average number of packets processed for each event\n\n");
 258	}
 259
 260	if (mask & SAMPLE_CPUMAP_KTHREAD_CNT) {
 261		printf("  kthread\t\tDisplays the number of packets processed in CPUMAP kthread for each CPU\n"
 262		       "  \t\t\tPackets consumed from ptr_ring in kthread, and its xdp_stats (after calling \n"
 263		       "  \t\t\tCPUMAP bpf prog) are expanded below this. xdp_stats are expanded as a total and\n"
 264		       "  \t\t\tthen per-CPU to associate it to each CPU's pinned CPUMAP kthread.\n"
 265		       "  \t\t\t\tpkt/s    - Packets consumed per second from ptr_ring\n"
 266		       "  \t\t\t\tdrop/s   - Packets dropped per second in kthread\n"
 267		       "  \t\t\t\tsched    - Number of times kthread called schedule()\n\n"
 268		       "  \t\t\txdp_stats (also expands to per-CPU counts)\n"
 269		       "  \t\t\t\tpass/s  - XDP_PASS count for CPUMAP program execution\n"
 270		       "  \t\t\t\tdrop/s  - XDP_DROP count for CPUMAP program execution\n"
 271		       "  \t\t\t\tredir/s - XDP_REDIRECT count for CPUMAP program execution\n\n");
 272	}
 273
 274	if (mask & SAMPLE_EXCEPTION_CNT) {
 275		printf("  xdp_exception\t\tDisplays xdp_exception tracepoint events\n"
 276		       "  \t\t\tThis can occur due to internal driver errors, unrecognized\n"
 277		       "  \t\t\tXDP actions and due to explicit user trigger by use of XDP_ABORTED\n"
 278		       "  \t\t\tEach action is expanded below this field with its count\n"
 279		       "  \t\t\t\thit/s     - Number of times the tracepoint was hit per second\n\n");
 280	}
 281
 282	if (mask & SAMPLE_DEVMAP_XMIT_CNT) {
 283		printf("  devmap_xmit\t\tDisplays devmap_xmit tracepoint events\n"
 284		       "  \t\t\tThis tracepoint is invoked for successful transmissions on output\n"
 285		       "  \t\t\tdevice but these statistics are not available for generic XDP mode,\n"
 286		       "  \t\t\thence they will be omitted from the output when using SKB mode\n"
 287		       "  \t\t\t\txmit/s    - Number of packets that were transmitted per second\n"
 288		       "  \t\t\t\tdrop/s    - Number of packets that failed transmissions per second\n"
 289		       "  \t\t\t\tdrv_err/s - Number of internal driver errors per second\n"
 290		       "  \t\t\t\tbulk-avg  - Average number of packets processed for each event\n\n");
 291	}
 292}
 293
 294void sample_usage(char *argv[], const struct option *long_options,
 295		  const char *doc, int mask, bool error)
 296{
 297	int i;
 298
 299	if (!error)
 300		sample_print_help(mask);
 301
 302	printf("\n%s\nOption for %s:\n", doc, argv[0]);
 303	for (i = 0; long_options[i].name != 0; i++) {
 304		printf(" --%-15s", long_options[i].name);
 305		if (long_options[i].flag != NULL)
 306			printf(" flag (internal value: %d)",
 307			       *long_options[i].flag);
 308		else
 309			printf("\t short-option: -%c", long_options[i].val);
 310		printf("\n");
 311	}
 312	printf("\n");
 313}
 314
 315static struct datarec *alloc_record_per_cpu(void)
 316{
 317	unsigned int nr_cpus = libbpf_num_possible_cpus();
 318	struct datarec *array;
 319
 320	array = calloc(nr_cpus, sizeof(*array));
 321	if (!array) {
 322		fprintf(stderr, "Failed to allocate memory (nr_cpus: %u)\n",
 323			nr_cpus);
 324		return NULL;
 325	}
 326	return array;
 327}
 328
 329static int map_entry_init(struct map_entry *e, __u64 pair)
 330{
 331	e->pair = pair;
 332	INIT_HLIST_NODE(&e->node);
 333	e->val.timestamp = gettime();
 334	e->val.cpu = alloc_record_per_cpu();
 335	if (!e->val.cpu)
 336		return -ENOMEM;
 337	return 0;
 338}
 339
 340static void map_collect_percpu(struct datarec *values, struct record *rec)
 341{
 342	/* For percpu maps, userspace gets a value per possible CPU */
 343	unsigned int nr_cpus = libbpf_num_possible_cpus();
 344	__u64 sum_xdp_redirect = 0;
 345	__u64 sum_processed = 0;
 346	__u64 sum_xdp_pass = 0;
 347	__u64 sum_xdp_drop = 0;
 348	__u64 sum_dropped = 0;
 349	__u64 sum_issue = 0;
 350	int i;
 351
 352	/* Get time as close as possible to reading map contents */
 353	rec->timestamp = gettime();
 354
 355	/* Record and sum values from each CPU */
 356	for (i = 0; i < nr_cpus; i++) {
 357		rec->cpu[i].processed = READ_ONCE(values[i].processed);
 358		rec->cpu[i].dropped = READ_ONCE(values[i].dropped);
 359		rec->cpu[i].issue = READ_ONCE(values[i].issue);
 360		rec->cpu[i].xdp_pass = READ_ONCE(values[i].xdp_pass);
 361		rec->cpu[i].xdp_drop = READ_ONCE(values[i].xdp_drop);
 362		rec->cpu[i].xdp_redirect = READ_ONCE(values[i].xdp_redirect);
 363
 364		sum_processed += rec->cpu[i].processed;
 365		sum_dropped += rec->cpu[i].dropped;
 366		sum_issue += rec->cpu[i].issue;
 367		sum_xdp_pass += rec->cpu[i].xdp_pass;
 368		sum_xdp_drop += rec->cpu[i].xdp_drop;
 369		sum_xdp_redirect += rec->cpu[i].xdp_redirect;
 370	}
 371
 372	rec->total.processed = sum_processed;
 373	rec->total.dropped = sum_dropped;
 374	rec->total.issue = sum_issue;
 375	rec->total.xdp_pass = sum_xdp_pass;
 376	rec->total.xdp_drop = sum_xdp_drop;
 377	rec->total.xdp_redirect = sum_xdp_redirect;
 378}
 379
 380static int map_collect_percpu_devmap(int map_fd, struct stats_record *rec)
 381{
 382	unsigned int nr_cpus = bpf_num_possible_cpus();
 383	__u32 batch, count = 32;
 384	struct datarec *values;
 385	bool init = false;
 386	__u64 *keys;
 387	int i, ret;
 388
 389	keys = calloc(count, sizeof(__u64));
 390	if (!keys)
 391		return -ENOMEM;
 392	values = calloc(count * nr_cpus, sizeof(struct datarec));
 393	if (!values) {
 394		free(keys);
 395		return -ENOMEM;
 396	}
 397
 398	for (;;) {
 399		bool exit = false;
 400
 401		ret = bpf_map_lookup_batch(map_fd, init ? &batch : NULL, &batch,
 402					   keys, values, &count, NULL);
 403		if (ret < 0 && errno != ENOENT)
 404			break;
 405		if (errno == ENOENT)
 406			exit = true;
 407
 408		init = true;
 409		for (i = 0; i < count; i++) {
 410			struct map_entry *e, *x = NULL;
 411			__u64 pair = keys[i];
 412			struct datarec *arr;
 413
 414			arr = &values[i * nr_cpus];
 415			hash_for_each_possible(rec->xmit_map, e, node, pair) {
 416				if (e->pair == pair) {
 417					x = e;
 418					break;
 419				}
 420			}
 421			if (!x) {
 422				x = calloc(1, sizeof(*x));
 423				if (!x)
 424					goto cleanup;
 425				if (map_entry_init(x, pair) < 0) {
 426					free(x);
 427					goto cleanup;
 428				}
 429				hash_add(rec->xmit_map, &x->node, pair);
 430			}
 431			map_collect_percpu(arr, &x->val);
 432		}
 433
 434		if (exit)
 435			break;
 436		count = 32;
 437	}
 438
 439	free(values);
 440	free(keys);
 441	return 0;
 442cleanup:
 443	free(values);
 444	free(keys);
 445	return -ENOMEM;
 446}
 447
 448static struct stats_record *alloc_stats_record(void)
 449{
 450	struct stats_record *rec;
 451	int i;
 452
 453	rec = calloc(1, sizeof(*rec) + sample_n_cpus * sizeof(struct record));
 454	if (!rec) {
 455		fprintf(stderr, "Failed to allocate memory\n");
 456		return NULL;
 457	}
 458
 459	if (sample_mask & SAMPLE_RX_CNT) {
 460		rec->rx_cnt.cpu = alloc_record_per_cpu();
 461		if (!rec->rx_cnt.cpu) {
 462			fprintf(stderr,
 463				"Failed to allocate rx_cnt per-CPU array\n");
 464			goto end_rec;
 465		}
 466	}
 467	if (sample_mask & (SAMPLE_REDIRECT_CNT | SAMPLE_REDIRECT_ERR_CNT)) {
 468		for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) {
 469			rec->redir_err[i].cpu = alloc_record_per_cpu();
 470			if (!rec->redir_err[i].cpu) {
 471				fprintf(stderr,
 472					"Failed to allocate redir_err per-CPU array for "
 473					"\"%s\" case\n",
 474					xdp_redirect_err_names[i]);
 475				while (i--)
 476					free(rec->redir_err[i].cpu);
 477				goto end_rx_cnt;
 478			}
 479		}
 480	}
 481	if (sample_mask & SAMPLE_CPUMAP_KTHREAD_CNT) {
 482		rec->kthread.cpu = alloc_record_per_cpu();
 483		if (!rec->kthread.cpu) {
 484			fprintf(stderr,
 485				"Failed to allocate kthread per-CPU array\n");
 486			goto end_redir;
 487		}
 488	}
 489	if (sample_mask & SAMPLE_EXCEPTION_CNT) {
 490		for (i = 0; i < XDP_ACTION_MAX; i++) {
 491			rec->exception[i].cpu = alloc_record_per_cpu();
 492			if (!rec->exception[i].cpu) {
 493				fprintf(stderr,
 494					"Failed to allocate exception per-CPU array for "
 495					"\"%s\" case\n",
 496					action2str(i));
 497				while (i--)
 498					free(rec->exception[i].cpu);
 499				goto end_kthread;
 500			}
 501		}
 502	}
 503	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT) {
 504		rec->devmap_xmit.cpu = alloc_record_per_cpu();
 505		if (!rec->devmap_xmit.cpu) {
 506			fprintf(stderr,
 507				"Failed to allocate devmap_xmit per-CPU array\n");
 508			goto end_exception;
 509		}
 510	}
 511	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
 512		hash_init(rec->xmit_map);
 513	if (sample_mask & SAMPLE_CPUMAP_ENQUEUE_CNT) {
 514		for (i = 0; i < sample_n_cpus; i++) {
 515			rec->enq[i].cpu = alloc_record_per_cpu();
 516			if (!rec->enq[i].cpu) {
 517				fprintf(stderr,
 518					"Failed to allocate enqueue per-CPU array for "
 519					"CPU %d\n",
 520					i);
 521				while (i--)
 522					free(rec->enq[i].cpu);
 523				goto end_devmap_xmit;
 524			}
 525		}
 526	}
 527
 528	return rec;
 529
 530end_devmap_xmit:
 531	free(rec->devmap_xmit.cpu);
 532end_exception:
 533	for (i = 0; i < XDP_ACTION_MAX; i++)
 534		free(rec->exception[i].cpu);
 535end_kthread:
 536	free(rec->kthread.cpu);
 537end_redir:
 538	for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++)
 539		free(rec->redir_err[i].cpu);
 540end_rx_cnt:
 541	free(rec->rx_cnt.cpu);
 542end_rec:
 543	free(rec);
 544	return NULL;
 545}
 546
 547static void free_stats_record(struct stats_record *r)
 548{
 549	struct hlist_node *tmp;
 550	struct map_entry *e;
 551	int i;
 552
 553	for (i = 0; i < sample_n_cpus; i++)
 554		free(r->enq[i].cpu);
 555	hash_for_each_safe(r->xmit_map, i, tmp, e, node) {
 556		hash_del(&e->node);
 557		free(e->val.cpu);
 558		free(e);
 559	}
 560	free(r->devmap_xmit.cpu);
 561	for (i = 0; i < XDP_ACTION_MAX; i++)
 562		free(r->exception[i].cpu);
 563	free(r->kthread.cpu);
 564	for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++)
 565		free(r->redir_err[i].cpu);
 566	free(r->rx_cnt.cpu);
 567	free(r);
 568}
 569
 570static double calc_period(struct record *r, struct record *p)
 571{
 572	double period_ = 0;
 573	__u64 period = 0;
 574
 575	period = r->timestamp - p->timestamp;
 576	if (period > 0)
 577		period_ = ((double)period / NANOSEC_PER_SEC);
 578
 579	return period_;
 580}
 581
 582static double sample_round(double val)
 583{
 584	if (val - floor(val) < 0.5)
 585		return floor(val);
 586	return ceil(val);
 587}
 588
 589static __u64 calc_pps(struct datarec *r, struct datarec *p, double period_)
 590{
 591	__u64 packets = 0;
 592	__u64 pps = 0;
 593
 594	if (period_ > 0) {
 595		packets = r->processed - p->processed;
 596		pps = sample_round(packets / period_);
 597	}
 598	return pps;
 599}
 600
 601static __u64 calc_drop_pps(struct datarec *r, struct datarec *p, double period_)
 602{
 603	__u64 packets = 0;
 604	__u64 pps = 0;
 605
 606	if (period_ > 0) {
 607		packets = r->dropped - p->dropped;
 608		pps = sample_round(packets / period_);
 609	}
 610	return pps;
 611}
 612
 613static __u64 calc_errs_pps(struct datarec *r, struct datarec *p, double period_)
 614{
 615	__u64 packets = 0;
 616	__u64 pps = 0;
 617
 618	if (period_ > 0) {
 619		packets = r->issue - p->issue;
 620		pps = sample_round(packets / period_);
 621	}
 622	return pps;
 623}
 624
 625static __u64 calc_info_pps(struct datarec *r, struct datarec *p, double period_)
 626{
 627	__u64 packets = 0;
 628	__u64 pps = 0;
 629
 630	if (period_ > 0) {
 631		packets = r->info - p->info;
 632		pps = sample_round(packets / period_);
 633	}
 634	return pps;
 635}
 636
 637static void calc_xdp_pps(struct datarec *r, struct datarec *p, double *xdp_pass,
 638			 double *xdp_drop, double *xdp_redirect, double period_)
 639{
 640	*xdp_pass = 0, *xdp_drop = 0, *xdp_redirect = 0;
 641	if (period_ > 0) {
 642		*xdp_redirect = (r->xdp_redirect - p->xdp_redirect) / period_;
 643		*xdp_pass = (r->xdp_pass - p->xdp_pass) / period_;
 644		*xdp_drop = (r->xdp_drop - p->xdp_drop) / period_;
 645	}
 646}
 647
 648static void stats_get_rx_cnt(struct stats_record *stats_rec,
 649			     struct stats_record *stats_prev,
 650			     unsigned int nr_cpus, struct sample_output *out)
 651{
 652	struct record *rec, *prev;
 653	double t, pps, drop, err;
 654	int i;
 655
 656	rec = &stats_rec->rx_cnt;
 657	prev = &stats_prev->rx_cnt;
 658	t = calc_period(rec, prev);
 659
 660	for (i = 0; i < nr_cpus; i++) {
 661		struct datarec *r = &rec->cpu[i];
 662		struct datarec *p = &prev->cpu[i];
 663		char str[64];
 664
 665		pps = calc_pps(r, p, t);
 666		drop = calc_drop_pps(r, p, t);
 667		err = calc_errs_pps(r, p, t);
 668		if (!pps && !drop && !err)
 669			continue;
 670
 671		snprintf(str, sizeof(str), "cpu:%d", i);
 672		print_default("    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 673			      "\n",
 674			      str, PPS(pps), DROP(drop), ERR(err));
 675	}
 676
 677	if (out) {
 678		pps = calc_pps(&rec->total, &prev->total, t);
 679		drop = calc_drop_pps(&rec->total, &prev->total, t);
 680		err = calc_errs_pps(&rec->total, &prev->total, t);
 681
 682		out->rx_cnt.pps = pps;
 683		out->rx_cnt.drop = drop;
 684		out->rx_cnt.err = err;
 685		out->totals.rx += pps;
 686		out->totals.drop += drop;
 687		out->totals.err += err;
 688	}
 689}
 690
 691static void stats_get_cpumap_enqueue(struct stats_record *stats_rec,
 692				     struct stats_record *stats_prev,
 693				     unsigned int nr_cpus)
 694{
 695	struct record *rec, *prev;
 696	double t, pps, drop, err;
 697	int i, to_cpu;
 698
 699	/* cpumap enqueue stats */
 700	for (to_cpu = 0; to_cpu < sample_n_cpus; to_cpu++) {
 701		rec = &stats_rec->enq[to_cpu];
 702		prev = &stats_prev->enq[to_cpu];
 703		t = calc_period(rec, prev);
 704
 705		pps = calc_pps(&rec->total, &prev->total, t);
 706		drop = calc_drop_pps(&rec->total, &prev->total, t);
 707		err = calc_errs_pps(&rec->total, &prev->total, t);
 708
 709		if (pps > 0 || drop > 0) {
 710			char str[64];
 711
 712			snprintf(str, sizeof(str), "enqueue to cpu %d", to_cpu);
 713
 714			if (err > 0)
 715				err = pps / err; /* calc average bulk size */
 716
 717			print_err(drop,
 718				  "  %-20s " FMT_COLUMNf FMT_COLUMNf __COLUMN(
 719					  ".2f") "\n",
 720				  str, PPS(pps), DROP(drop), err, "bulk-avg");
 721		}
 722
 723		for (i = 0; i < nr_cpus; i++) {
 724			struct datarec *r = &rec->cpu[i];
 725			struct datarec *p = &prev->cpu[i];
 726			char str[64];
 727
 728			pps = calc_pps(r, p, t);
 729			drop = calc_drop_pps(r, p, t);
 730			err = calc_errs_pps(r, p, t);
 731			if (!pps && !drop && !err)
 732				continue;
 733
 734			snprintf(str, sizeof(str), "cpu:%d->%d", i, to_cpu);
 735			if (err > 0)
 736				err = pps / err; /* calc average bulk size */
 737			print_default(
 738				"    %-18s " FMT_COLUMNf FMT_COLUMNf __COLUMN(
 739					".2f") "\n",
 740				str, PPS(pps), DROP(drop), err, "bulk-avg");
 741		}
 742	}
 743}
 744
 745static void stats_get_cpumap_remote(struct stats_record *stats_rec,
 746				    struct stats_record *stats_prev,
 747				    unsigned int nr_cpus)
 748{
 749	double xdp_pass, xdp_drop, xdp_redirect;
 750	struct record *rec, *prev;
 751	double t;
 752	int i;
 753
 754	rec = &stats_rec->kthread;
 755	prev = &stats_prev->kthread;
 756	t = calc_period(rec, prev);
 757
 758	calc_xdp_pps(&rec->total, &prev->total, &xdp_pass, &xdp_drop,
 759		     &xdp_redirect, t);
 760	if (xdp_pass || xdp_drop || xdp_redirect) {
 761		print_err(xdp_drop,
 762			  "    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf "\n",
 763			  "xdp_stats", PASS(xdp_pass), DROP(xdp_drop),
 764			  REDIR(xdp_redirect));
 765	}
 766
 767	for (i = 0; i < nr_cpus; i++) {
 768		struct datarec *r = &rec->cpu[i];
 769		struct datarec *p = &prev->cpu[i];
 770		char str[64];
 771
 772		calc_xdp_pps(r, p, &xdp_pass, &xdp_drop, &xdp_redirect, t);
 773		if (!xdp_pass && !xdp_drop && !xdp_redirect)
 774			continue;
 775
 776		snprintf(str, sizeof(str), "cpu:%d", i);
 777		print_default("      %-16s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 778			      "\n",
 779			      str, PASS(xdp_pass), DROP(xdp_drop),
 780			      REDIR(xdp_redirect));
 781	}
 782}
 783
 784static void stats_get_cpumap_kthread(struct stats_record *stats_rec,
 785				     struct stats_record *stats_prev,
 786				     unsigned int nr_cpus)
 787{
 788	struct record *rec, *prev;
 789	double t, pps, drop, err;
 790	int i;
 791
 792	rec = &stats_rec->kthread;
 793	prev = &stats_prev->kthread;
 794	t = calc_period(rec, prev);
 795
 796	pps = calc_pps(&rec->total, &prev->total, t);
 797	drop = calc_drop_pps(&rec->total, &prev->total, t);
 798	err = calc_errs_pps(&rec->total, &prev->total, t);
 799
 800	print_err(drop, "  %-20s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf "\n",
 801		  pps ? "kthread total" : "kthread", PPS(pps), DROP(drop), err,
 802		  "sched");
 803
 804	for (i = 0; i < nr_cpus; i++) {
 805		struct datarec *r = &rec->cpu[i];
 806		struct datarec *p = &prev->cpu[i];
 807		char str[64];
 808
 809		pps = calc_pps(r, p, t);
 810		drop = calc_drop_pps(r, p, t);
 811		err = calc_errs_pps(r, p, t);
 812		if (!pps && !drop && !err)
 813			continue;
 814
 815		snprintf(str, sizeof(str), "cpu:%d", i);
 816		print_default("    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 817			      "\n",
 818			      str, PPS(pps), DROP(drop), err, "sched");
 819	}
 820}
 821
 822static void stats_get_redirect_cnt(struct stats_record *stats_rec,
 823				   struct stats_record *stats_prev,
 824				   unsigned int nr_cpus,
 825				   struct sample_output *out)
 826{
 827	struct record *rec, *prev;
 828	double t, pps;
 829	int i;
 830
 831	rec = &stats_rec->redir_err[0];
 832	prev = &stats_prev->redir_err[0];
 833	t = calc_period(rec, prev);
 834	for (i = 0; i < nr_cpus; i++) {
 835		struct datarec *r = &rec->cpu[i];
 836		struct datarec *p = &prev->cpu[i];
 837		char str[64];
 838
 839		pps = calc_pps(r, p, t);
 840		if (!pps)
 841			continue;
 842
 843		snprintf(str, sizeof(str), "cpu:%d", i);
 844		print_default("    %-18s " FMT_COLUMNf "\n", str, REDIR(pps));
 845	}
 846
 847	if (out) {
 848		pps = calc_pps(&rec->total, &prev->total, t);
 849		out->redir_cnt.suc = pps;
 850		out->totals.redir += pps;
 851	}
 852}
 853
 854static void stats_get_redirect_err_cnt(struct stats_record *stats_rec,
 855				       struct stats_record *stats_prev,
 856				       unsigned int nr_cpus,
 857				       struct sample_output *out)
 858{
 859	struct record *rec, *prev;
 860	double t, drop, sum = 0;
 861	int rec_i, i;
 862
 863	for (rec_i = 1; rec_i < XDP_REDIRECT_ERR_MAX; rec_i++) {
 864		char str[64];
 865
 866		rec = &stats_rec->redir_err[rec_i];
 867		prev = &stats_prev->redir_err[rec_i];
 868		t = calc_period(rec, prev);
 869
 870		drop = calc_drop_pps(&rec->total, &prev->total, t);
 871		if (drop > 0 && !out) {
 872			snprintf(str, sizeof(str),
 873				 sample_log_level & LL_DEFAULT ? "%s total" :
 874								       "%s",
 875				 xdp_redirect_err_names[rec_i]);
 876			print_err(drop, "    %-18s " FMT_COLUMNf "\n", str,
 877				  ERR(drop));
 878		}
 879
 880		for (i = 0; i < nr_cpus; i++) {
 881			struct datarec *r = &rec->cpu[i];
 882			struct datarec *p = &prev->cpu[i];
 883			double drop;
 884
 885			drop = calc_drop_pps(r, p, t);
 886			if (!drop)
 887				continue;
 888
 889			snprintf(str, sizeof(str), "cpu:%d", i);
 890			print_default("       %-16s" FMT_COLUMNf "\n", str,
 891				      ERR(drop));
 892		}
 893
 894		sum += drop;
 895	}
 896
 897	if (out) {
 898		out->redir_cnt.err = sum;
 899		out->totals.err += sum;
 900	}
 901}
 902
 903static void stats_get_exception_cnt(struct stats_record *stats_rec,
 904				    struct stats_record *stats_prev,
 905				    unsigned int nr_cpus,
 906				    struct sample_output *out)
 907{
 908	double t, drop, sum = 0;
 909	struct record *rec, *prev;
 910	int rec_i, i;
 911
 912	for (rec_i = 0; rec_i < XDP_ACTION_MAX; rec_i++) {
 913		rec = &stats_rec->exception[rec_i];
 914		prev = &stats_prev->exception[rec_i];
 915		t = calc_period(rec, prev);
 916
 917		drop = calc_drop_pps(&rec->total, &prev->total, t);
 918		/* Fold out errors after heading */
 919		sum += drop;
 920
 921		if (drop > 0 && !out) {
 922			print_always("    %-18s " FMT_COLUMNf "\n",
 923				     action2str(rec_i), ERR(drop));
 924
 925			for (i = 0; i < nr_cpus; i++) {
 926				struct datarec *r = &rec->cpu[i];
 927				struct datarec *p = &prev->cpu[i];
 928				char str[64];
 929				double drop;
 930
 931				drop = calc_drop_pps(r, p, t);
 932				if (!drop)
 933					continue;
 934
 935				snprintf(str, sizeof(str), "cpu:%d", i);
 936				print_default("       %-16s" FMT_COLUMNf "\n",
 937					      str, ERR(drop));
 938			}
 939		}
 940	}
 941
 942	if (out) {
 943		out->except_cnt.hits = sum;
 944		out->totals.err += sum;
 945	}
 946}
 947
 948static void stats_get_devmap_xmit(struct stats_record *stats_rec,
 949				  struct stats_record *stats_prev,
 950				  unsigned int nr_cpus,
 951				  struct sample_output *out)
 952{
 953	double pps, drop, info, err;
 954	struct record *rec, *prev;
 955	double t;
 956	int i;
 957
 958	rec = &stats_rec->devmap_xmit;
 959	prev = &stats_prev->devmap_xmit;
 960	t = calc_period(rec, prev);
 961	for (i = 0; i < nr_cpus; i++) {
 962		struct datarec *r = &rec->cpu[i];
 963		struct datarec *p = &prev->cpu[i];
 964		char str[64];
 965
 966		pps = calc_pps(r, p, t);
 967		drop = calc_drop_pps(r, p, t);
 968		err = calc_errs_pps(r, p, t);
 969
 970		if (!pps && !drop && !err)
 971			continue;
 972
 973		snprintf(str, sizeof(str), "cpu:%d", i);
 974		info = calc_info_pps(r, p, t);
 975		if (info > 0)
 976			info = (pps + drop) / info; /* calc avg bulk */
 977		print_default("     %-18s" FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 978				      __COLUMN(".2f") "\n",
 979			      str, XMIT(pps), DROP(drop), err, "drv_err/s",
 980			      info, "bulk-avg");
 981	}
 982	if (out) {
 983		pps = calc_pps(&rec->total, &prev->total, t);
 984		drop = calc_drop_pps(&rec->total, &prev->total, t);
 985		info = calc_info_pps(&rec->total, &prev->total, t);
 986		if (info > 0)
 987			info = (pps + drop) / info; /* calc avg bulk */
 988		err = calc_errs_pps(&rec->total, &prev->total, t);
 989
 990		out->xmit_cnt.pps = pps;
 991		out->xmit_cnt.drop = drop;
 992		out->xmit_cnt.bavg = info;
 993		out->xmit_cnt.err = err;
 994		out->totals.xmit += pps;
 995		out->totals.drop_xmit += drop;
 996		out->totals.err += err;
 997	}
 998}
 999
1000static void stats_get_devmap_xmit_multi(struct stats_record *stats_rec,
1001					struct stats_record *stats_prev,
1002					unsigned int nr_cpus,
1003					struct sample_output *out,
1004					bool xmit_total)
1005{
1006	double pps, drop, info, err;
1007	struct map_entry *entry;
1008	struct record *r, *p;
1009	double t;
1010	int bkt;
1011
1012	hash_for_each(stats_rec->xmit_map, bkt, entry, node) {
1013		struct map_entry *e, *x = NULL;
1014		char ifname_from[IFNAMSIZ];
1015		char ifname_to[IFNAMSIZ];
1016		const char *fstr, *tstr;
1017		unsigned long prev_time;
1018		struct record beg = {};
1019		__u32 from_idx, to_idx;
1020		char str[128];
1021		__u64 pair;
1022		int i;
1023
1024		prev_time = sample_interval * NANOSEC_PER_SEC;
1025
1026		pair = entry->pair;
1027		from_idx = pair >> 32;
1028		to_idx = pair & 0xFFFFFFFF;
1029
1030		r = &entry->val;
1031		beg.timestamp = r->timestamp - prev_time;
1032
1033		/* Find matching entry from stats_prev map */
1034		hash_for_each_possible(stats_prev->xmit_map, e, node, pair) {
1035			if (e->pair == pair) {
1036				x = e;
1037				break;
1038			}
1039		}
1040		if (x)
1041			p = &x->val;
1042		else
1043			p = &beg;
1044		t = calc_period(r, p);
1045		pps = calc_pps(&r->total, &p->total, t);
1046		drop = calc_drop_pps(&r->total, &p->total, t);
1047		info = calc_info_pps(&r->total, &p->total, t);
1048		if (info > 0)
1049			info = (pps + drop) / info; /* calc avg bulk */
1050		err = calc_errs_pps(&r->total, &p->total, t);
1051
1052		if (out) {
1053			/* We are responsible for filling out totals */
1054			out->totals.xmit += pps;
1055			out->totals.drop_xmit += drop;
1056			out->totals.err += err;
1057			continue;
1058		}
1059
1060		fstr = tstr = NULL;
1061		if (if_indextoname(from_idx, ifname_from))
1062			fstr = ifname_from;
1063		if (if_indextoname(to_idx, ifname_to))
1064			tstr = ifname_to;
1065
1066		snprintf(str, sizeof(str), "xmit %s->%s", fstr ?: "?",
1067			 tstr ?: "?");
1068		/* Skip idle streams of redirection */
1069		if (pps || drop || err) {
1070			print_err(drop,
1071				  "  %-20s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
1072				  __COLUMN(".2f") "\n", str, XMIT(pps), DROP(drop),
1073				  err, "drv_err/s", info, "bulk-avg");
1074		}
1075
1076		for (i = 0; i < nr_cpus; i++) {
1077			struct datarec *rc = &r->cpu[i];
1078			struct datarec *pc, p_beg = {};
1079			char str[64];
1080
1081			pc = p == &beg ? &p_beg : &p->cpu[i];
1082
1083			pps = calc_pps(rc, pc, t);
1084			drop = calc_drop_pps(rc, pc, t);
1085			err = calc_errs_pps(rc, pc, t);
1086
1087			if (!pps && !drop && !err)
1088				continue;
1089
1090			snprintf(str, sizeof(str), "cpu:%d", i);
1091			info = calc_info_pps(rc, pc, t);
1092			if (info > 0)
1093				info = (pps + drop) / info; /* calc avg bulk */
1094
1095			print_default("     %-18s" FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
1096				      __COLUMN(".2f") "\n", str, XMIT(pps),
1097				      DROP(drop), err, "drv_err/s", info, "bulk-avg");
1098		}
1099	}
1100}
1101
1102static void stats_print(const char *prefix, int mask, struct stats_record *r,
1103			struct stats_record *p, struct sample_output *out)
1104{
1105	int nr_cpus = libbpf_num_possible_cpus();
1106	const char *str;
1107
1108	print_always("%-23s", prefix ?: "Summary");
1109	if (mask & SAMPLE_RX_CNT)
1110		print_always(FMT_COLUMNl, RX(out->totals.rx));
1111	if (mask & SAMPLE_REDIRECT_CNT)
1112		print_always(FMT_COLUMNl, REDIR(out->totals.redir));
1113	printf(FMT_COLUMNl,
1114	       out->totals.err + out->totals.drop + out->totals.drop_xmit,
1115	       "err,drop/s");
1116	if (mask & SAMPLE_DEVMAP_XMIT_CNT ||
1117	    mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
1118		printf(FMT_COLUMNl, XMIT(out->totals.xmit));
1119	printf("\n");
1120
1121	if (mask & SAMPLE_RX_CNT) {
1122		str = (sample_log_level & LL_DEFAULT) && out->rx_cnt.pps ?
1123				    "receive total" :
1124				    "receive";
1125		print_err((out->rx_cnt.err || out->rx_cnt.drop),
1126			  "  %-20s " FMT_COLUMNl FMT_COLUMNl FMT_COLUMNl "\n",
1127			  str, PPS(out->rx_cnt.pps), DROP(out->rx_cnt.drop),
1128			  ERR(out->rx_cnt.err));
1129
1130		stats_get_rx_cnt(r, p, nr_cpus, NULL);
1131	}
1132
1133	if (mask & SAMPLE_CPUMAP_ENQUEUE_CNT)
1134		stats_get_cpumap_enqueue(r, p, nr_cpus);
1135
1136	if (mask & SAMPLE_CPUMAP_KTHREAD_CNT) {
1137		stats_get_cpumap_kthread(r, p, nr_cpus);
1138		stats_get_cpumap_remote(r, p, nr_cpus);
1139	}
1140
1141	if (mask & SAMPLE_REDIRECT_CNT) {
1142		str = out->redir_cnt.suc ? "redirect total" : "redirect";
1143		print_default("  %-20s " FMT_COLUMNl "\n", str,
1144			      REDIR(out->redir_cnt.suc));
1145
1146		stats_get_redirect_cnt(r, p, nr_cpus, NULL);
1147	}
1148
1149	if (mask & SAMPLE_REDIRECT_ERR_CNT) {
1150		str = (sample_log_level & LL_DEFAULT) && out->redir_cnt.err ?
1151				    "redirect_err total" :
1152				    "redirect_err";
1153		print_err(out->redir_cnt.err, "  %-20s " FMT_COLUMNl "\n", str,
1154			  ERR(out->redir_cnt.err));
1155
1156		stats_get_redirect_err_cnt(r, p, nr_cpus, NULL);
1157	}
1158
1159	if (mask & SAMPLE_EXCEPTION_CNT) {
1160		str = out->except_cnt.hits ? "xdp_exception total" :
1161						   "xdp_exception";
1162
1163		print_err(out->except_cnt.hits, "  %-20s " FMT_COLUMNl "\n", str,
1164			  HITS(out->except_cnt.hits));
1165
1166		stats_get_exception_cnt(r, p, nr_cpus, NULL);
1167	}
1168
1169	if (mask & SAMPLE_DEVMAP_XMIT_CNT) {
1170		str = (sample_log_level & LL_DEFAULT) && out->xmit_cnt.pps ?
1171				    "devmap_xmit total" :
1172				    "devmap_xmit";
1173
1174		print_err(out->xmit_cnt.err || out->xmit_cnt.drop,
1175			  "  %-20s " FMT_COLUMNl FMT_COLUMNl FMT_COLUMNl
1176				  __COLUMN(".2f") "\n",
1177			  str, XMIT(out->xmit_cnt.pps),
1178			  DROP(out->xmit_cnt.drop), out->xmit_cnt.err,
1179			  "drv_err/s", out->xmit_cnt.bavg, "bulk-avg");
1180
1181		stats_get_devmap_xmit(r, p, nr_cpus, NULL);
1182	}
1183
1184	if (mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
1185		stats_get_devmap_xmit_multi(r, p, nr_cpus, NULL,
1186					    mask & SAMPLE_DEVMAP_XMIT_CNT);
1187
1188	if (sample_log_level & LL_DEFAULT ||
1189	    ((sample_log_level & LL_SIMPLE) && sample_err_exp)) {
1190		sample_err_exp = false;
1191		printf("\n");
1192	}
1193}
1194
1195int sample_setup_maps(struct bpf_map **maps)
1196{
1197	sample_n_cpus = libbpf_num_possible_cpus();
1198
1199	for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) {
1200		sample_map[i] = maps[i];
1201
1202		switch (i) {
1203		case MAP_RX:
1204		case MAP_CPUMAP_KTHREAD:
1205		case MAP_DEVMAP_XMIT:
1206			sample_map_count[i] = sample_n_cpus;
1207			break;
1208		case MAP_REDIRECT_ERR:
1209			sample_map_count[i] =
1210				XDP_REDIRECT_ERR_MAX * sample_n_cpus;
1211			break;
1212		case MAP_EXCEPTION:
1213			sample_map_count[i] = XDP_ACTION_MAX * sample_n_cpus;
1214		case MAP_CPUMAP_ENQUEUE:
1215			sample_map_count[i] = sample_n_cpus * sample_n_cpus;
1216			break;
1217		default:
1218			return -EINVAL;
1219		}
1220		if (bpf_map__set_max_entries(sample_map[i], sample_map_count[i]) < 0)
1221			return -errno;
1222	}
1223	sample_map[MAP_DEVMAP_XMIT_MULTI] = maps[MAP_DEVMAP_XMIT_MULTI];
1224	return 0;
1225}
1226
1227static int sample_setup_maps_mappings(void)
1228{
1229	for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) {
1230		size_t size = sample_map_count[i] * sizeof(struct datarec);
1231
1232		sample_mmap[i] = mmap(NULL, size, PROT_READ | PROT_WRITE,
1233				      MAP_SHARED, bpf_map__fd(sample_map[i]), 0);
1234		if (sample_mmap[i] == MAP_FAILED)
1235			return -errno;
1236	}
1237	return 0;
1238}
1239
1240int __sample_init(int mask)
1241{
1242	sigset_t st;
1243
1244	sigemptyset(&st);
1245	sigaddset(&st, SIGQUIT);
1246	sigaddset(&st, SIGINT);
1247	sigaddset(&st, SIGTERM);
1248
1249	if (sigprocmask(SIG_BLOCK, &st, NULL) < 0)
1250		return -errno;
1251
1252	sample_sig_fd = signalfd(-1, &st, SFD_CLOEXEC | SFD_NONBLOCK);
1253	if (sample_sig_fd < 0)
1254		return -errno;
1255
1256	sample_mask = mask;
1257
1258	return sample_setup_maps_mappings();
1259}
1260
1261static int __sample_remove_xdp(int ifindex, __u32 prog_id, int xdp_flags)
1262{
1263	__u32 cur_prog_id = 0;
1264	int ret;
1265
1266	if (prog_id) {
1267		ret = bpf_xdp_query_id(ifindex, xdp_flags, &cur_prog_id);
1268		if (ret < 0)
1269			return -errno;
1270
1271		if (prog_id != cur_prog_id) {
1272			print_always(
1273				"Program on ifindex %d does not match installed "
1274				"program, skipping unload\n",
1275				ifindex);
1276			return -ENOENT;
1277		}
1278	}
1279
1280	return bpf_xdp_detach(ifindex, xdp_flags, NULL);
1281}
1282
1283int sample_install_xdp(struct bpf_program *xdp_prog, int ifindex, bool generic,
1284		       bool force)
1285{
1286	int ret, xdp_flags = 0;
1287	__u32 prog_id = 0;
1288
1289	if (sample_xdp_cnt == 32) {
1290		fprintf(stderr,
1291			"Total limit for installed XDP programs in a sample reached\n");
1292		return -ENOTSUP;
1293	}
1294
1295	xdp_flags |= !force ? XDP_FLAGS_UPDATE_IF_NOEXIST : 0;
1296	xdp_flags |= generic ? XDP_FLAGS_SKB_MODE : XDP_FLAGS_DRV_MODE;
1297	ret = bpf_xdp_attach(ifindex, bpf_program__fd(xdp_prog), xdp_flags, NULL);
1298	if (ret < 0) {
1299		ret = -errno;
1300		fprintf(stderr,
1301			"Failed to install program \"%s\" on ifindex %d, mode = %s, "
1302			"force = %s: %s\n",
1303			bpf_program__name(xdp_prog), ifindex,
1304			generic ? "skb" : "native", force ? "true" : "false",
1305			strerror(-ret));
1306		return ret;
1307	}
1308
1309	ret = bpf_xdp_query_id(ifindex, xdp_flags, &prog_id);
1310	if (ret < 0) {
1311		ret = -errno;
1312		fprintf(stderr,
1313			"Failed to get XDP program id for ifindex %d, removing program: %s\n",
1314			ifindex, strerror(errno));
1315		__sample_remove_xdp(ifindex, 0, xdp_flags);
1316		return ret;
1317	}
1318	sample_xdp_progs[sample_xdp_cnt++] =
1319		(struct xdp_desc){ ifindex, prog_id, xdp_flags };
1320
1321	return 0;
1322}
1323
1324static void sample_summary_print(void)
1325{
1326	double num = sample_out.rx_cnt.num;
1327
1328	if (sample_out.totals.rx) {
1329		double pkts = sample_out.totals.rx;
1330
1331		print_always("  Packets received    : %'-10llu\n",
1332			     sample_out.totals.rx);
1333		print_always("  Average packets/s   : %'-10.0f\n",
1334			     sample_round(pkts / num));
1335	}
1336	if (sample_out.totals.redir) {
1337		double pkts = sample_out.totals.redir;
1338
1339		print_always("  Packets redirected  : %'-10llu\n",
1340			     sample_out.totals.redir);
1341		print_always("  Average redir/s     : %'-10.0f\n",
1342			     sample_round(pkts / num));
1343	}
1344	if (sample_out.totals.drop)
1345		print_always("  Rx dropped          : %'-10llu\n",
1346			     sample_out.totals.drop);
1347	if (sample_out.totals.drop_xmit)
1348		print_always("  Tx dropped          : %'-10llu\n",
1349			     sample_out.totals.drop_xmit);
1350	if (sample_out.totals.err)
1351		print_always("  Errors recorded     : %'-10llu\n",
1352			     sample_out.totals.err);
1353	if (sample_out.totals.xmit) {
1354		double pkts = sample_out.totals.xmit;
1355
1356		print_always("  Packets transmitted : %'-10llu\n",
1357			     sample_out.totals.xmit);
1358		print_always("  Average transmit/s  : %'-10.0f\n",
1359			     sample_round(pkts / num));
1360	}
1361}
1362
1363void sample_exit(int status)
1364{
1365	size_t size;
1366
1367	for (int i = 0; i < NUM_MAP; i++) {
1368		size = sample_map_count[i] * sizeof(**sample_mmap);
1369		munmap(sample_mmap[i], size);
1370	}
1371	while (sample_xdp_cnt--) {
1372		int i = sample_xdp_cnt, ifindex, xdp_flags;
1373		__u32 prog_id;
1374
1375		prog_id = sample_xdp_progs[i].prog_id;
1376		ifindex = sample_xdp_progs[i].ifindex;
1377		xdp_flags = sample_xdp_progs[i].flags;
1378
1379		__sample_remove_xdp(ifindex, prog_id, xdp_flags);
1380	}
1381	sample_summary_print();
1382	close(sample_sig_fd);
1383	exit(status);
1384}
1385
1386static int sample_stats_collect(struct stats_record *rec)
1387{
1388	int i;
1389
1390	if (sample_mask & SAMPLE_RX_CNT)
1391		map_collect_percpu(sample_mmap[MAP_RX], &rec->rx_cnt);
1392
1393	if (sample_mask & SAMPLE_REDIRECT_CNT)
1394		map_collect_percpu(sample_mmap[MAP_REDIRECT_ERR], &rec->redir_err[0]);
1395
1396	if (sample_mask & SAMPLE_REDIRECT_ERR_CNT) {
1397		for (i = 1; i < XDP_REDIRECT_ERR_MAX; i++)
1398			map_collect_percpu(&sample_mmap[MAP_REDIRECT_ERR][i * sample_n_cpus],
1399					   &rec->redir_err[i]);
1400	}
1401
1402	if (sample_mask & SAMPLE_CPUMAP_ENQUEUE_CNT)
1403		for (i = 0; i < sample_n_cpus; i++)
1404			map_collect_percpu(&sample_mmap[MAP_CPUMAP_ENQUEUE][i * sample_n_cpus],
1405					   &rec->enq[i]);
1406
1407	if (sample_mask & SAMPLE_CPUMAP_KTHREAD_CNT)
1408		map_collect_percpu(sample_mmap[MAP_CPUMAP_KTHREAD],
1409				   &rec->kthread);
1410
1411	if (sample_mask & SAMPLE_EXCEPTION_CNT)
1412		for (i = 0; i < XDP_ACTION_MAX; i++)
1413			map_collect_percpu(&sample_mmap[MAP_EXCEPTION][i * sample_n_cpus],
1414					   &rec->exception[i]);
1415
1416	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT)
1417		map_collect_percpu(sample_mmap[MAP_DEVMAP_XMIT], &rec->devmap_xmit);
1418
1419	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) {
1420		if (map_collect_percpu_devmap(bpf_map__fd(sample_map[MAP_DEVMAP_XMIT_MULTI]), rec) < 0)
1421			return -EINVAL;
1422	}
1423	return 0;
1424}
1425
1426static void sample_summary_update(struct sample_output *out)
1427{
1428	sample_out.totals.rx += out->totals.rx;
1429	sample_out.totals.redir += out->totals.redir;
1430	sample_out.totals.drop += out->totals.drop;
1431	sample_out.totals.drop_xmit += out->totals.drop_xmit;
1432	sample_out.totals.err += out->totals.err;
1433	sample_out.totals.xmit += out->totals.xmit;
1434	sample_out.rx_cnt.num++;
1435}
1436
1437static void sample_stats_print(int mask, struct stats_record *cur,
1438			       struct stats_record *prev, char *prog_name)
1439{
1440	struct sample_output out = {};
1441
1442	if (mask & SAMPLE_RX_CNT)
1443		stats_get_rx_cnt(cur, prev, 0, &out);
1444	if (mask & SAMPLE_REDIRECT_CNT)
1445		stats_get_redirect_cnt(cur, prev, 0, &out);
1446	if (mask & SAMPLE_REDIRECT_ERR_CNT)
1447		stats_get_redirect_err_cnt(cur, prev, 0, &out);
1448	if (mask & SAMPLE_EXCEPTION_CNT)
1449		stats_get_exception_cnt(cur, prev, 0, &out);
1450	if (mask & SAMPLE_DEVMAP_XMIT_CNT)
1451		stats_get_devmap_xmit(cur, prev, 0, &out);
1452	else if (mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
1453		stats_get_devmap_xmit_multi(cur, prev, 0, &out,
1454					    mask & SAMPLE_DEVMAP_XMIT_CNT);
1455	sample_summary_update(&out);
1456
1457	stats_print(prog_name, mask, cur, prev, &out);
1458}
1459
1460void sample_switch_mode(void)
1461{
1462	sample_log_level ^= LL_DEBUG - 1;
1463}
1464
1465static int sample_signal_cb(void)
1466{
1467	struct signalfd_siginfo si;
1468	int r;
1469
1470	r = read(sample_sig_fd, &si, sizeof(si));
1471	if (r < 0)
1472		return -errno;
1473
1474	switch (si.ssi_signo) {
1475	case SIGQUIT:
1476		sample_switch_mode();
1477		printf("\n");
1478		break;
1479	default:
1480		printf("\n");
1481		return 1;
1482	}
1483
1484	return 0;
1485}
1486
1487/* Pointer swap trick */
1488static void swap(struct stats_record **a, struct stats_record **b)
1489{
1490	struct stats_record *tmp;
1491
1492	tmp = *a;
1493	*a = *b;
1494	*b = tmp;
1495}
1496
1497static int sample_timer_cb(int timerfd, struct stats_record **rec,
1498			   struct stats_record **prev)
1499{
1500	char line[64] = "Summary";
1501	int ret;
1502	__u64 t;
1503
1504	ret = read(timerfd, &t, sizeof(t));
1505	if (ret < 0)
1506		return -errno;
1507
1508	swap(prev, rec);
1509	ret = sample_stats_collect(*rec);
1510	if (ret < 0)
1511		return ret;
1512
1513	if (sample_xdp_cnt == 2 && !(sample_mask & SAMPLE_SKIP_HEADING)) {
1514		char fi[IFNAMSIZ];
1515		char to[IFNAMSIZ];
1516		const char *f, *t;
1517
1518		f = t = NULL;
1519		if (if_indextoname(sample_xdp_progs[0].ifindex, fi))
1520			f = fi;
1521		if (if_indextoname(sample_xdp_progs[1].ifindex, to))
1522			t = to;
1523
1524		snprintf(line, sizeof(line), "%s->%s", f ?: "?", t ?: "?");
1525	}
1526
1527	sample_stats_print(sample_mask, *rec, *prev, line);
1528	return 0;
1529}
1530
1531int sample_run(int interval, void (*post_cb)(void *), void *ctx)
1532{
1533	struct timespec ts = { interval, 0 };
1534	struct itimerspec its = { ts, ts };
1535	struct stats_record *rec, *prev;
1536	struct pollfd pfd[2] = {};
1537	int timerfd, ret;
1538
1539	if (!interval) {
1540		fprintf(stderr, "Incorrect interval 0\n");
1541		return -EINVAL;
1542	}
1543	sample_interval = interval;
1544	/* Pretty print numbers */
1545	setlocale(LC_NUMERIC, "en_US.UTF-8");
1546
1547	timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_CLOEXEC | TFD_NONBLOCK);
1548	if (timerfd < 0)
1549		return -errno;
1550	timerfd_settime(timerfd, 0, &its, NULL);
1551
1552	pfd[0].fd = sample_sig_fd;
1553	pfd[0].events = POLLIN;
1554
1555	pfd[1].fd = timerfd;
1556	pfd[1].events = POLLIN;
1557
1558	ret = -ENOMEM;
1559	rec = alloc_stats_record();
1560	if (!rec)
1561		goto end;
1562	prev = alloc_stats_record();
1563	if (!prev)
1564		goto end_rec;
1565
1566	ret = sample_stats_collect(rec);
1567	if (ret < 0)
1568		goto end_rec_prev;
1569
1570	for (;;) {
1571		ret = poll(pfd, 2, -1);
1572		if (ret < 0) {
1573			if (errno == EINTR)
1574				continue;
1575			else
1576				break;
1577		}
1578
1579		if (pfd[0].revents & POLLIN)
1580			ret = sample_signal_cb();
1581		else if (pfd[1].revents & POLLIN)
1582			ret = sample_timer_cb(timerfd, &rec, &prev);
1583
1584		if (ret)
1585			break;
1586
1587		if (post_cb)
1588			post_cb(ctx);
1589	}
1590
1591end_rec_prev:
1592	free_stats_record(prev);
1593end_rec:
1594	free_stats_record(rec);
1595end:
1596	close(timerfd);
1597
1598	return ret;
1599}
1600
1601const char *get_driver_name(int ifindex)
1602{
1603	struct ethtool_drvinfo drv = {};
1604	char ifname[IF_NAMESIZE];
1605	static char drvname[32];
1606	struct ifreq ifr = {};
1607	int fd, r = 0;
1608
1609	fd = socket(AF_INET, SOCK_DGRAM, 0);
1610	if (fd < 0)
1611		return "[error]";
1612
1613	if (!if_indextoname(ifindex, ifname))
1614		goto end;
1615
1616	drv.cmd = ETHTOOL_GDRVINFO;
1617	safe_strncpy(ifr.ifr_name, ifname, sizeof(ifr.ifr_name));
1618	ifr.ifr_data = (void *)&drv;
1619
1620	r = ioctl(fd, SIOCETHTOOL, &ifr);
1621	if (r)
1622		goto end;
1623
1624	safe_strncpy(drvname, drv.driver, sizeof(drvname));
1625
1626	close(fd);
1627	return drvname;
1628
1629end:
1630	r = errno;
1631	close(fd);
1632	return r == EOPNOTSUPP ? "loopback" : "[error]";
1633}
1634
1635int get_mac_addr(int ifindex, void *mac_addr)
1636{
1637	char ifname[IF_NAMESIZE];
1638	struct ifreq ifr = {};
1639	int fd, r;
1640
1641	fd = socket(AF_INET, SOCK_DGRAM, 0);
1642	if (fd < 0)
1643		return -errno;
1644
1645	if (!if_indextoname(ifindex, ifname)) {
1646		r = -errno;
1647		goto end;
1648	}
1649
1650	safe_strncpy(ifr.ifr_name, ifname, sizeof(ifr.ifr_name));
1651
1652	r = ioctl(fd, SIOCGIFHWADDR, &ifr);
1653	if (r) {
1654		r = -errno;
1655		goto end;
1656	}
1657
1658	memcpy(mac_addr, ifr.ifr_hwaddr.sa_data, 6 * sizeof(char));
1659
1660end:
1661	close(fd);
1662	return r;
1663}
1664
1665__attribute__((constructor)) static void sample_ctor(void)
1666{
1667	if (libbpf_set_strict_mode(LIBBPF_STRICT_ALL) < 0) {
1668		fprintf(stderr, "Failed to set libbpf strict mode: %s\n",
1669			strerror(errno));
1670		/* Just exit, nothing to cleanup right now */
1671		exit(EXIT_FAIL_BPF);
1672	}
1673}