forked from mirrors/linux
		
	 b25acdafd3
			
		
	
	
		b25acdafd3
		
	
	
	
	
		
			
			We have switched to memcg-based memory accouting and thus the rlimit is not needed any more. LIBBPF_STRICT_AUTO_RLIMIT_MEMLOCK was introduced in libbpf for backward compatibility, so we can use it instead now. This patch also removes the useless header sys/resource.h from many files in samples/bpf. Signed-off-by: Yafang Shao <laoar.shao@gmail.com> Signed-off-by: Andrii Nakryiko <andrii@kernel.org> Link: https://lore.kernel.org/bpf/20220409125958.92629-2-laoar.shao@gmail.com
		
			
				
	
	
		
			1673 lines
		
	
	
	
		
			42 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1673 lines
		
	
	
	
		
			42 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-only
 | |
| #define _GNU_SOURCE
 | |
| 
 | |
| #include <arpa/inet.h>
 | |
| #include <bpf/bpf.h>
 | |
| #include <bpf/libbpf.h>
 | |
| #include <errno.h>
 | |
| #include <fcntl.h>
 | |
| #include <getopt.h>
 | |
| #include <linux/ethtool.h>
 | |
| #include <linux/hashtable.h>
 | |
| #include <linux/if_link.h>
 | |
| #include <linux/jhash.h>
 | |
| #include <linux/limits.h>
 | |
| #include <linux/list.h>
 | |
| #include <linux/sockios.h>
 | |
| #include <locale.h>
 | |
| #include <math.h>
 | |
| #include <net/if.h>
 | |
| #include <poll.h>
 | |
| #include <signal.h>
 | |
| #include <stdbool.h>
 | |
| #include <stdio.h>
 | |
| #include <stdlib.h>
 | |
| #include <string.h>
 | |
| #include <sys/ioctl.h>
 | |
| #include <sys/mman.h>
 | |
| #include <sys/signalfd.h>
 | |
| #include <sys/sysinfo.h>
 | |
| #include <sys/timerfd.h>
 | |
| #include <sys/utsname.h>
 | |
| #include <time.h>
 | |
| #include <unistd.h>
 | |
| 
 | |
| #include "bpf_util.h"
 | |
| #include "xdp_sample_user.h"
 | |
| 
 | |
| #define __sample_print(fmt, cond, ...)                                         \
 | |
| 	({                                                                     \
 | |
| 		if (cond)                                                      \
 | |
| 			printf(fmt, ##__VA_ARGS__);                            \
 | |
| 	})
 | |
| 
 | |
| #define print_always(fmt, ...) __sample_print(fmt, 1, ##__VA_ARGS__)
 | |
| #define print_default(fmt, ...)                                                \
 | |
| 	__sample_print(fmt, sample_log_level & LL_DEFAULT, ##__VA_ARGS__)
 | |
| #define __print_err(err, fmt, ...)                                             \
 | |
| 	({                                                                     \
 | |
| 		__sample_print(fmt, err > 0 || sample_log_level & LL_DEFAULT,  \
 | |
| 			       ##__VA_ARGS__);                                 \
 | |
| 		sample_err_exp = sample_err_exp ? true : err > 0;              \
 | |
| 	})
 | |
| #define print_err(err, fmt, ...) __print_err(err, fmt, ##__VA_ARGS__)
 | |
| 
 | |
| #define __COLUMN(x) "%'10" x " %-13s"
 | |
| #define FMT_COLUMNf __COLUMN(".0f")
 | |
| #define FMT_COLUMNd __COLUMN("d")
 | |
| #define FMT_COLUMNl __COLUMN("llu")
 | |
| #define RX(rx) rx, "rx/s"
 | |
| #define PPS(pps) pps, "pkt/s"
 | |
| #define DROP(drop) drop, "drop/s"
 | |
| #define ERR(err) err, "error/s"
 | |
| #define HITS(hits) hits, "hit/s"
 | |
| #define XMIT(xmit) xmit, "xmit/s"
 | |
| #define PASS(pass) pass, "pass/s"
 | |
| #define REDIR(redir) redir, "redir/s"
 | |
| #define NANOSEC_PER_SEC 1000000000 /* 10^9 */
 | |
| 
 | |
| #define XDP_UNKNOWN (XDP_REDIRECT + 1)
 | |
| #define XDP_ACTION_MAX (XDP_UNKNOWN + 1)
 | |
| #define XDP_REDIRECT_ERR_MAX 7
 | |
| 
 | |
| enum map_type {
 | |
| 	MAP_RX,
 | |
| 	MAP_REDIRECT_ERR,
 | |
| 	MAP_CPUMAP_ENQUEUE,
 | |
| 	MAP_CPUMAP_KTHREAD,
 | |
| 	MAP_EXCEPTION,
 | |
| 	MAP_DEVMAP_XMIT,
 | |
| 	MAP_DEVMAP_XMIT_MULTI,
 | |
| 	NUM_MAP,
 | |
| };
 | |
| 
 | |
| enum log_level {
 | |
| 	LL_DEFAULT = 1U << 0,
 | |
| 	LL_SIMPLE = 1U << 1,
 | |
| 	LL_DEBUG = 1U << 2,
 | |
| };
 | |
| 
 | |
| struct record {
 | |
| 	__u64 timestamp;
 | |
| 	struct datarec total;
 | |
| 	struct datarec *cpu;
 | |
| };
 | |
| 
 | |
| struct map_entry {
 | |
| 	struct hlist_node node;
 | |
| 	__u64 pair;
 | |
| 	struct record val;
 | |
| };
 | |
| 
 | |
| struct stats_record {
 | |
| 	struct record rx_cnt;
 | |
| 	struct record redir_err[XDP_REDIRECT_ERR_MAX];
 | |
| 	struct record kthread;
 | |
| 	struct record exception[XDP_ACTION_MAX];
 | |
| 	struct record devmap_xmit;
 | |
| 	DECLARE_HASHTABLE(xmit_map, 5);
 | |
| 	struct record enq[];
 | |
| };
 | |
| 
 | |
| struct sample_output {
 | |
| 	struct {
 | |
| 		__u64 rx;
 | |
| 		__u64 redir;
 | |
| 		__u64 drop;
 | |
| 		__u64 drop_xmit;
 | |
| 		__u64 err;
 | |
| 		__u64 xmit;
 | |
| 	} totals;
 | |
| 	struct {
 | |
| 		union {
 | |
| 			__u64 pps;
 | |
| 			__u64 num;
 | |
| 		};
 | |
| 		__u64 drop;
 | |
| 		__u64 err;
 | |
| 	} rx_cnt;
 | |
| 	struct {
 | |
| 		__u64 suc;
 | |
| 		__u64 err;
 | |
| 	} redir_cnt;
 | |
| 	struct {
 | |
| 		__u64 hits;
 | |
| 	} except_cnt;
 | |
| 	struct {
 | |
| 		__u64 pps;
 | |
| 		__u64 drop;
 | |
| 		__u64 err;
 | |
| 		double bavg;
 | |
| 	} xmit_cnt;
 | |
| };
 | |
| 
 | |
| struct xdp_desc {
 | |
| 	int ifindex;
 | |
| 	__u32 prog_id;
 | |
| 	int flags;
 | |
| } sample_xdp_progs[32];
 | |
| 
 | |
| struct datarec *sample_mmap[NUM_MAP];
 | |
| struct bpf_map *sample_map[NUM_MAP];
 | |
| size_t sample_map_count[NUM_MAP];
 | |
| enum log_level sample_log_level;
 | |
| struct sample_output sample_out;
 | |
| unsigned long sample_interval;
 | |
| bool sample_err_exp;
 | |
| int sample_xdp_cnt;
 | |
| int sample_n_cpus;
 | |
| int sample_sig_fd;
 | |
| int sample_mask;
 | |
| 
 | |
| static const char *xdp_redirect_err_names[XDP_REDIRECT_ERR_MAX] = {
 | |
| 	/* Key=1 keeps unknown errors */
 | |
| 	"Success",
 | |
| 	"Unknown",
 | |
| 	"EINVAL",
 | |
| 	"ENETDOWN",
 | |
| 	"EMSGSIZE",
 | |
| 	"EOPNOTSUPP",
 | |
| 	"ENOSPC",
 | |
| };
 | |
| 
 | |
| /* Keyed from Unknown */
 | |
| static const char *xdp_redirect_err_help[XDP_REDIRECT_ERR_MAX - 1] = {
 | |
| 	"Unknown error",
 | |
| 	"Invalid redirection",
 | |
| 	"Device being redirected to is down",
 | |
| 	"Packet length too large for device",
 | |
| 	"Operation not supported",
 | |
| 	"No space in ptr_ring of cpumap kthread",
 | |
| };
 | |
| 
 | |
| static const char *xdp_action_names[XDP_ACTION_MAX] = {
 | |
| 	[XDP_ABORTED]  = "XDP_ABORTED",
 | |
| 	[XDP_DROP]     = "XDP_DROP",
 | |
| 	[XDP_PASS]     = "XDP_PASS",
 | |
| 	[XDP_TX]       = "XDP_TX",
 | |
| 	[XDP_REDIRECT] = "XDP_REDIRECT",
 | |
| 	[XDP_UNKNOWN]  = "XDP_UNKNOWN",
 | |
| };
 | |
| 
 | |
| static __u64 gettime(void)
 | |
| {
 | |
| 	struct timespec t;
 | |
| 	int res;
 | |
| 
 | |
| 	res = clock_gettime(CLOCK_MONOTONIC, &t);
 | |
| 	if (res < 0) {
 | |
| 		fprintf(stderr, "Error with gettimeofday! (%i)\n", res);
 | |
| 		return UINT64_MAX;
 | |
| 	}
 | |
| 	return (__u64)t.tv_sec * NANOSEC_PER_SEC + t.tv_nsec;
 | |
| }
 | |
| 
 | |
| static const char *action2str(int action)
 | |
| {
 | |
| 	if (action < XDP_ACTION_MAX)
 | |
| 		return xdp_action_names[action];
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| static void sample_print_help(int mask)
 | |
| {
 | |
| 	printf("Output format description\n\n"
 | |
| 	       "By default, redirect success statistics are disabled, use -s to enable.\n"
 | |
| 	       "The terse output mode is default, verbose mode can be activated using -v\n"
 | |
| 	       "Use SIGQUIT (Ctrl + \\) to switch the mode dynamically at runtime\n\n"
 | |
| 	       "Terse mode displays at most the following fields:\n"
 | |
| 	       "  rx/s        Number of packets received per second\n"
 | |
| 	       "  redir/s     Number of packets successfully redirected per second\n"
 | |
| 	       "  err,drop/s  Aggregated count of errors per second (including dropped packets)\n"
 | |
| 	       "  xmit/s      Number of packets transmitted on the output device per second\n\n"
 | |
| 	       "Output description for verbose mode:\n"
 | |
| 	       "  FIELD                 DESCRIPTION\n");
 | |
| 
 | |
| 	if (mask & SAMPLE_RX_CNT) {
 | |
| 		printf("  receive\t\tDisplays the number of packets received & errors encountered\n"
 | |
| 		       " \t\t\tWhenever an error or packet drop occurs, details of per CPU error\n"
 | |
| 		       " \t\t\tand drop statistics will be expanded inline in terse mode.\n"
 | |
| 		       " \t\t\t\tpkt/s     - Packets received per second\n"
 | |
| 		       " \t\t\t\tdrop/s    - Packets dropped per second\n"
 | |
| 		       " \t\t\t\terror/s   - Errors encountered per second\n\n");
 | |
| 	}
 | |
| 	if (mask & (SAMPLE_REDIRECT_CNT | SAMPLE_REDIRECT_ERR_CNT)) {
 | |
| 		printf("  redirect\t\tDisplays the number of packets successfully redirected\n"
 | |
| 		       "  \t\t\tErrors encountered are expanded under redirect_err field\n"
 | |
| 		       "  \t\t\tNote that passing -s to enable it has a per packet overhead\n"
 | |
| 		       "  \t\t\t\tredir/s   - Packets redirected successfully per second\n\n"
 | |
| 		       "  redirect_err\t\tDisplays the number of packets that failed redirection\n"
 | |
| 		       "  \t\t\tThe errno is expanded under this field with per CPU count\n"
 | |
| 		       "  \t\t\tThe recognized errors are:\n");
 | |
| 
 | |
| 		for (int i = 2; i < XDP_REDIRECT_ERR_MAX; i++)
 | |
| 			printf("\t\t\t  %s: %s\n", xdp_redirect_err_names[i],
 | |
| 			       xdp_redirect_err_help[i - 1]);
 | |
| 
 | |
| 		printf("  \n\t\t\t\terror/s   - Packets that failed redirection per second\n\n");
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_CPUMAP_ENQUEUE_CNT) {
 | |
| 		printf("  enqueue to cpu N\tDisplays the number of packets enqueued to bulk queue of CPU N\n"
 | |
| 		       "  \t\t\tExpands to cpu:FROM->N to display enqueue stats for each CPU enqueuing to CPU N\n"
 | |
| 		       "  \t\t\tReceived packets can be associated with the CPU redirect program is enqueuing \n"
 | |
| 		       "  \t\t\tpackets to.\n"
 | |
| 		       "  \t\t\t\tpkt/s    - Packets enqueued per second from other CPU to CPU N\n"
 | |
| 		       "  \t\t\t\tdrop/s   - Packets dropped when trying to enqueue to CPU N\n"
 | |
| 		       "  \t\t\t\tbulk-avg - Average number of packets processed for each event\n\n");
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_CPUMAP_KTHREAD_CNT) {
 | |
| 		printf("  kthread\t\tDisplays the number of packets processed in CPUMAP kthread for each CPU\n"
 | |
| 		       "  \t\t\tPackets consumed from ptr_ring in kthread, and its xdp_stats (after calling \n"
 | |
| 		       "  \t\t\tCPUMAP bpf prog) are expanded below this. xdp_stats are expanded as a total and\n"
 | |
| 		       "  \t\t\tthen per-CPU to associate it to each CPU's pinned CPUMAP kthread.\n"
 | |
| 		       "  \t\t\t\tpkt/s    - Packets consumed per second from ptr_ring\n"
 | |
| 		       "  \t\t\t\tdrop/s   - Packets dropped per second in kthread\n"
 | |
| 		       "  \t\t\t\tsched    - Number of times kthread called schedule()\n\n"
 | |
| 		       "  \t\t\txdp_stats (also expands to per-CPU counts)\n"
 | |
| 		       "  \t\t\t\tpass/s  - XDP_PASS count for CPUMAP program execution\n"
 | |
| 		       "  \t\t\t\tdrop/s  - XDP_DROP count for CPUMAP program execution\n"
 | |
| 		       "  \t\t\t\tredir/s - XDP_REDIRECT count for CPUMAP program execution\n\n");
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_EXCEPTION_CNT) {
 | |
| 		printf("  xdp_exception\t\tDisplays xdp_exception tracepoint events\n"
 | |
| 		       "  \t\t\tThis can occur due to internal driver errors, unrecognized\n"
 | |
| 		       "  \t\t\tXDP actions and due to explicit user trigger by use of XDP_ABORTED\n"
 | |
| 		       "  \t\t\tEach action is expanded below this field with its count\n"
 | |
| 		       "  \t\t\t\thit/s     - Number of times the tracepoint was hit per second\n\n");
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_DEVMAP_XMIT_CNT) {
 | |
| 		printf("  devmap_xmit\t\tDisplays devmap_xmit tracepoint events\n"
 | |
| 		       "  \t\t\tThis tracepoint is invoked for successful transmissions on output\n"
 | |
| 		       "  \t\t\tdevice but these statistics are not available for generic XDP mode,\n"
 | |
| 		       "  \t\t\thence they will be omitted from the output when using SKB mode\n"
 | |
| 		       "  \t\t\t\txmit/s    - Number of packets that were transmitted per second\n"
 | |
| 		       "  \t\t\t\tdrop/s    - Number of packets that failed transmissions per second\n"
 | |
| 		       "  \t\t\t\tdrv_err/s - Number of internal driver errors per second\n"
 | |
| 		       "  \t\t\t\tbulk-avg  - Average number of packets processed for each event\n\n");
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void sample_usage(char *argv[], const struct option *long_options,
 | |
| 		  const char *doc, int mask, bool error)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	if (!error)
 | |
| 		sample_print_help(mask);
 | |
| 
 | |
| 	printf("\n%s\nOption for %s:\n", doc, argv[0]);
 | |
| 	for (i = 0; long_options[i].name != 0; i++) {
 | |
| 		printf(" --%-15s", long_options[i].name);
 | |
| 		if (long_options[i].flag != NULL)
 | |
| 			printf(" flag (internal value: %d)",
 | |
| 			       *long_options[i].flag);
 | |
| 		else
 | |
| 			printf("\t short-option: -%c", long_options[i].val);
 | |
| 		printf("\n");
 | |
| 	}
 | |
| 	printf("\n");
 | |
| }
 | |
| 
 | |
| static struct datarec *alloc_record_per_cpu(void)
 | |
| {
 | |
| 	unsigned int nr_cpus = libbpf_num_possible_cpus();
 | |
| 	struct datarec *array;
 | |
| 
 | |
| 	array = calloc(nr_cpus, sizeof(*array));
 | |
| 	if (!array) {
 | |
| 		fprintf(stderr, "Failed to allocate memory (nr_cpus: %u)\n",
 | |
| 			nr_cpus);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 	return array;
 | |
| }
 | |
| 
 | |
| static int map_entry_init(struct map_entry *e, __u64 pair)
 | |
| {
 | |
| 	e->pair = pair;
 | |
| 	INIT_HLIST_NODE(&e->node);
 | |
| 	e->val.timestamp = gettime();
 | |
| 	e->val.cpu = alloc_record_per_cpu();
 | |
| 	if (!e->val.cpu)
 | |
| 		return -ENOMEM;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void map_collect_percpu(struct datarec *values, struct record *rec)
 | |
| {
 | |
| 	/* For percpu maps, userspace gets a value per possible CPU */
 | |
| 	unsigned int nr_cpus = libbpf_num_possible_cpus();
 | |
| 	__u64 sum_xdp_redirect = 0;
 | |
| 	__u64 sum_processed = 0;
 | |
| 	__u64 sum_xdp_pass = 0;
 | |
| 	__u64 sum_xdp_drop = 0;
 | |
| 	__u64 sum_dropped = 0;
 | |
| 	__u64 sum_issue = 0;
 | |
| 	int i;
 | |
| 
 | |
| 	/* Get time as close as possible to reading map contents */
 | |
| 	rec->timestamp = gettime();
 | |
| 
 | |
| 	/* Record and sum values from each CPU */
 | |
| 	for (i = 0; i < nr_cpus; i++) {
 | |
| 		rec->cpu[i].processed = READ_ONCE(values[i].processed);
 | |
| 		rec->cpu[i].dropped = READ_ONCE(values[i].dropped);
 | |
| 		rec->cpu[i].issue = READ_ONCE(values[i].issue);
 | |
| 		rec->cpu[i].xdp_pass = READ_ONCE(values[i].xdp_pass);
 | |
| 		rec->cpu[i].xdp_drop = READ_ONCE(values[i].xdp_drop);
 | |
| 		rec->cpu[i].xdp_redirect = READ_ONCE(values[i].xdp_redirect);
 | |
| 
 | |
| 		sum_processed += rec->cpu[i].processed;
 | |
| 		sum_dropped += rec->cpu[i].dropped;
 | |
| 		sum_issue += rec->cpu[i].issue;
 | |
| 		sum_xdp_pass += rec->cpu[i].xdp_pass;
 | |
| 		sum_xdp_drop += rec->cpu[i].xdp_drop;
 | |
| 		sum_xdp_redirect += rec->cpu[i].xdp_redirect;
 | |
| 	}
 | |
| 
 | |
| 	rec->total.processed = sum_processed;
 | |
| 	rec->total.dropped = sum_dropped;
 | |
| 	rec->total.issue = sum_issue;
 | |
| 	rec->total.xdp_pass = sum_xdp_pass;
 | |
| 	rec->total.xdp_drop = sum_xdp_drop;
 | |
| 	rec->total.xdp_redirect = sum_xdp_redirect;
 | |
| }
 | |
| 
 | |
| static int map_collect_percpu_devmap(int map_fd, struct stats_record *rec)
 | |
| {
 | |
| 	unsigned int nr_cpus = bpf_num_possible_cpus();
 | |
| 	__u32 batch, count = 32;
 | |
| 	struct datarec *values;
 | |
| 	bool init = false;
 | |
| 	__u64 *keys;
 | |
| 	int i, ret;
 | |
| 
 | |
| 	keys = calloc(count, sizeof(__u64));
 | |
| 	if (!keys)
 | |
| 		return -ENOMEM;
 | |
| 	values = calloc(count * nr_cpus, sizeof(struct datarec));
 | |
| 	if (!values) {
 | |
| 		free(keys);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	for (;;) {
 | |
| 		bool exit = false;
 | |
| 
 | |
| 		ret = bpf_map_lookup_batch(map_fd, init ? &batch : NULL, &batch,
 | |
| 					   keys, values, &count, NULL);
 | |
| 		if (ret < 0 && errno != ENOENT)
 | |
| 			break;
 | |
| 		if (errno == ENOENT)
 | |
| 			exit = true;
 | |
| 
 | |
| 		init = true;
 | |
| 		for (i = 0; i < count; i++) {
 | |
| 			struct map_entry *e, *x = NULL;
 | |
| 			__u64 pair = keys[i];
 | |
| 			struct datarec *arr;
 | |
| 
 | |
| 			arr = &values[i * nr_cpus];
 | |
| 			hash_for_each_possible(rec->xmit_map, e, node, pair) {
 | |
| 				if (e->pair == pair) {
 | |
| 					x = e;
 | |
| 					break;
 | |
| 				}
 | |
| 			}
 | |
| 			if (!x) {
 | |
| 				x = calloc(1, sizeof(*x));
 | |
| 				if (!x)
 | |
| 					goto cleanup;
 | |
| 				if (map_entry_init(x, pair) < 0) {
 | |
| 					free(x);
 | |
| 					goto cleanup;
 | |
| 				}
 | |
| 				hash_add(rec->xmit_map, &x->node, pair);
 | |
| 			}
 | |
| 			map_collect_percpu(arr, &x->val);
 | |
| 		}
 | |
| 
 | |
| 		if (exit)
 | |
| 			break;
 | |
| 		count = 32;
 | |
| 	}
 | |
| 
 | |
| 	free(values);
 | |
| 	free(keys);
 | |
| 	return 0;
 | |
| cleanup:
 | |
| 	free(values);
 | |
| 	free(keys);
 | |
| 	return -ENOMEM;
 | |
| }
 | |
| 
 | |
| static struct stats_record *alloc_stats_record(void)
 | |
| {
 | |
| 	struct stats_record *rec;
 | |
| 	int i;
 | |
| 
 | |
| 	rec = calloc(1, sizeof(*rec) + sample_n_cpus * sizeof(struct record));
 | |
| 	if (!rec) {
 | |
| 		fprintf(stderr, "Failed to allocate memory\n");
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_RX_CNT) {
 | |
| 		rec->rx_cnt.cpu = alloc_record_per_cpu();
 | |
| 		if (!rec->rx_cnt.cpu) {
 | |
| 			fprintf(stderr,
 | |
| 				"Failed to allocate rx_cnt per-CPU array\n");
 | |
| 			goto end_rec;
 | |
| 		}
 | |
| 	}
 | |
| 	if (sample_mask & (SAMPLE_REDIRECT_CNT | SAMPLE_REDIRECT_ERR_CNT)) {
 | |
| 		for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) {
 | |
| 			rec->redir_err[i].cpu = alloc_record_per_cpu();
 | |
| 			if (!rec->redir_err[i].cpu) {
 | |
| 				fprintf(stderr,
 | |
| 					"Failed to allocate redir_err per-CPU array for "
 | |
| 					"\"%s\" case\n",
 | |
| 					xdp_redirect_err_names[i]);
 | |
| 				while (i--)
 | |
| 					free(rec->redir_err[i].cpu);
 | |
| 				goto end_rx_cnt;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	if (sample_mask & SAMPLE_CPUMAP_KTHREAD_CNT) {
 | |
| 		rec->kthread.cpu = alloc_record_per_cpu();
 | |
| 		if (!rec->kthread.cpu) {
 | |
| 			fprintf(stderr,
 | |
| 				"Failed to allocate kthread per-CPU array\n");
 | |
| 			goto end_redir;
 | |
| 		}
 | |
| 	}
 | |
| 	if (sample_mask & SAMPLE_EXCEPTION_CNT) {
 | |
| 		for (i = 0; i < XDP_ACTION_MAX; i++) {
 | |
| 			rec->exception[i].cpu = alloc_record_per_cpu();
 | |
| 			if (!rec->exception[i].cpu) {
 | |
| 				fprintf(stderr,
 | |
| 					"Failed to allocate exception per-CPU array for "
 | |
| 					"\"%s\" case\n",
 | |
| 					action2str(i));
 | |
| 				while (i--)
 | |
| 					free(rec->exception[i].cpu);
 | |
| 				goto end_kthread;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT) {
 | |
| 		rec->devmap_xmit.cpu = alloc_record_per_cpu();
 | |
| 		if (!rec->devmap_xmit.cpu) {
 | |
| 			fprintf(stderr,
 | |
| 				"Failed to allocate devmap_xmit per-CPU array\n");
 | |
| 			goto end_exception;
 | |
| 		}
 | |
| 	}
 | |
| 	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
 | |
| 		hash_init(rec->xmit_map);
 | |
| 	if (sample_mask & SAMPLE_CPUMAP_ENQUEUE_CNT) {
 | |
| 		for (i = 0; i < sample_n_cpus; i++) {
 | |
| 			rec->enq[i].cpu = alloc_record_per_cpu();
 | |
| 			if (!rec->enq[i].cpu) {
 | |
| 				fprintf(stderr,
 | |
| 					"Failed to allocate enqueue per-CPU array for "
 | |
| 					"CPU %d\n",
 | |
| 					i);
 | |
| 				while (i--)
 | |
| 					free(rec->enq[i].cpu);
 | |
| 				goto end_devmap_xmit;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return rec;
 | |
| 
 | |
| end_devmap_xmit:
 | |
| 	free(rec->devmap_xmit.cpu);
 | |
| end_exception:
 | |
| 	for (i = 0; i < XDP_ACTION_MAX; i++)
 | |
| 		free(rec->exception[i].cpu);
 | |
| end_kthread:
 | |
| 	free(rec->kthread.cpu);
 | |
| end_redir:
 | |
| 	for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++)
 | |
| 		free(rec->redir_err[i].cpu);
 | |
| end_rx_cnt:
 | |
| 	free(rec->rx_cnt.cpu);
 | |
| end_rec:
 | |
| 	free(rec);
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| static void free_stats_record(struct stats_record *r)
 | |
| {
 | |
| 	struct hlist_node *tmp;
 | |
| 	struct map_entry *e;
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = 0; i < sample_n_cpus; i++)
 | |
| 		free(r->enq[i].cpu);
 | |
| 	hash_for_each_safe(r->xmit_map, i, tmp, e, node) {
 | |
| 		hash_del(&e->node);
 | |
| 		free(e->val.cpu);
 | |
| 		free(e);
 | |
| 	}
 | |
| 	free(r->devmap_xmit.cpu);
 | |
| 	for (i = 0; i < XDP_ACTION_MAX; i++)
 | |
| 		free(r->exception[i].cpu);
 | |
| 	free(r->kthread.cpu);
 | |
| 	for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++)
 | |
| 		free(r->redir_err[i].cpu);
 | |
| 	free(r->rx_cnt.cpu);
 | |
| 	free(r);
 | |
| }
 | |
| 
 | |
| static double calc_period(struct record *r, struct record *p)
 | |
| {
 | |
| 	double period_ = 0;
 | |
| 	__u64 period = 0;
 | |
| 
 | |
| 	period = r->timestamp - p->timestamp;
 | |
| 	if (period > 0)
 | |
| 		period_ = ((double)period / NANOSEC_PER_SEC);
 | |
| 
 | |
| 	return period_;
 | |
| }
 | |
| 
 | |
| static double sample_round(double val)
 | |
| {
 | |
| 	if (val - floor(val) < 0.5)
 | |
| 		return floor(val);
 | |
| 	return ceil(val);
 | |
| }
 | |
| 
 | |
| static __u64 calc_pps(struct datarec *r, struct datarec *p, double period_)
 | |
| {
 | |
| 	__u64 packets = 0;
 | |
| 	__u64 pps = 0;
 | |
| 
 | |
| 	if (period_ > 0) {
 | |
| 		packets = r->processed - p->processed;
 | |
| 		pps = sample_round(packets / period_);
 | |
| 	}
 | |
| 	return pps;
 | |
| }
 | |
| 
 | |
| static __u64 calc_drop_pps(struct datarec *r, struct datarec *p, double period_)
 | |
| {
 | |
| 	__u64 packets = 0;
 | |
| 	__u64 pps = 0;
 | |
| 
 | |
| 	if (period_ > 0) {
 | |
| 		packets = r->dropped - p->dropped;
 | |
| 		pps = sample_round(packets / period_);
 | |
| 	}
 | |
| 	return pps;
 | |
| }
 | |
| 
 | |
| static __u64 calc_errs_pps(struct datarec *r, struct datarec *p, double period_)
 | |
| {
 | |
| 	__u64 packets = 0;
 | |
| 	__u64 pps = 0;
 | |
| 
 | |
| 	if (period_ > 0) {
 | |
| 		packets = r->issue - p->issue;
 | |
| 		pps = sample_round(packets / period_);
 | |
| 	}
 | |
| 	return pps;
 | |
| }
 | |
| 
 | |
| static __u64 calc_info_pps(struct datarec *r, struct datarec *p, double period_)
 | |
| {
 | |
| 	__u64 packets = 0;
 | |
| 	__u64 pps = 0;
 | |
| 
 | |
| 	if (period_ > 0) {
 | |
| 		packets = r->info - p->info;
 | |
| 		pps = sample_round(packets / period_);
 | |
| 	}
 | |
| 	return pps;
 | |
| }
 | |
| 
 | |
| static void calc_xdp_pps(struct datarec *r, struct datarec *p, double *xdp_pass,
 | |
| 			 double *xdp_drop, double *xdp_redirect, double period_)
 | |
| {
 | |
| 	*xdp_pass = 0, *xdp_drop = 0, *xdp_redirect = 0;
 | |
| 	if (period_ > 0) {
 | |
| 		*xdp_redirect = (r->xdp_redirect - p->xdp_redirect) / period_;
 | |
| 		*xdp_pass = (r->xdp_pass - p->xdp_pass) / period_;
 | |
| 		*xdp_drop = (r->xdp_drop - p->xdp_drop) / period_;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_rx_cnt(struct stats_record *stats_rec,
 | |
| 			     struct stats_record *stats_prev,
 | |
| 			     unsigned int nr_cpus, struct sample_output *out)
 | |
| {
 | |
| 	struct record *rec, *prev;
 | |
| 	double t, pps, drop, err;
 | |
| 	int i;
 | |
| 
 | |
| 	rec = &stats_rec->rx_cnt;
 | |
| 	prev = &stats_prev->rx_cnt;
 | |
| 	t = calc_period(rec, prev);
 | |
| 
 | |
| 	for (i = 0; i < nr_cpus; i++) {
 | |
| 		struct datarec *r = &rec->cpu[i];
 | |
| 		struct datarec *p = &prev->cpu[i];
 | |
| 		char str[64];
 | |
| 
 | |
| 		pps = calc_pps(r, p, t);
 | |
| 		drop = calc_drop_pps(r, p, t);
 | |
| 		err = calc_errs_pps(r, p, t);
 | |
| 		if (!pps && !drop && !err)
 | |
| 			continue;
 | |
| 
 | |
| 		snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 		print_default("    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 | |
| 			      "\n",
 | |
| 			      str, PPS(pps), DROP(drop), ERR(err));
 | |
| 	}
 | |
| 
 | |
| 	if (out) {
 | |
| 		pps = calc_pps(&rec->total, &prev->total, t);
 | |
| 		drop = calc_drop_pps(&rec->total, &prev->total, t);
 | |
| 		err = calc_errs_pps(&rec->total, &prev->total, t);
 | |
| 
 | |
| 		out->rx_cnt.pps = pps;
 | |
| 		out->rx_cnt.drop = drop;
 | |
| 		out->rx_cnt.err = err;
 | |
| 		out->totals.rx += pps;
 | |
| 		out->totals.drop += drop;
 | |
| 		out->totals.err += err;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_cpumap_enqueue(struct stats_record *stats_rec,
 | |
| 				     struct stats_record *stats_prev,
 | |
| 				     unsigned int nr_cpus)
 | |
| {
 | |
| 	struct record *rec, *prev;
 | |
| 	double t, pps, drop, err;
 | |
| 	int i, to_cpu;
 | |
| 
 | |
| 	/* cpumap enqueue stats */
 | |
| 	for (to_cpu = 0; to_cpu < sample_n_cpus; to_cpu++) {
 | |
| 		rec = &stats_rec->enq[to_cpu];
 | |
| 		prev = &stats_prev->enq[to_cpu];
 | |
| 		t = calc_period(rec, prev);
 | |
| 
 | |
| 		pps = calc_pps(&rec->total, &prev->total, t);
 | |
| 		drop = calc_drop_pps(&rec->total, &prev->total, t);
 | |
| 		err = calc_errs_pps(&rec->total, &prev->total, t);
 | |
| 
 | |
| 		if (pps > 0 || drop > 0) {
 | |
| 			char str[64];
 | |
| 
 | |
| 			snprintf(str, sizeof(str), "enqueue to cpu %d", to_cpu);
 | |
| 
 | |
| 			if (err > 0)
 | |
| 				err = pps / err; /* calc average bulk size */
 | |
| 
 | |
| 			print_err(drop,
 | |
| 				  "  %-20s " FMT_COLUMNf FMT_COLUMNf __COLUMN(
 | |
| 					  ".2f") "\n",
 | |
| 				  str, PPS(pps), DROP(drop), err, "bulk-avg");
 | |
| 		}
 | |
| 
 | |
| 		for (i = 0; i < nr_cpus; i++) {
 | |
| 			struct datarec *r = &rec->cpu[i];
 | |
| 			struct datarec *p = &prev->cpu[i];
 | |
| 			char str[64];
 | |
| 
 | |
| 			pps = calc_pps(r, p, t);
 | |
| 			drop = calc_drop_pps(r, p, t);
 | |
| 			err = calc_errs_pps(r, p, t);
 | |
| 			if (!pps && !drop && !err)
 | |
| 				continue;
 | |
| 
 | |
| 			snprintf(str, sizeof(str), "cpu:%d->%d", i, to_cpu);
 | |
| 			if (err > 0)
 | |
| 				err = pps / err; /* calc average bulk size */
 | |
| 			print_default(
 | |
| 				"    %-18s " FMT_COLUMNf FMT_COLUMNf __COLUMN(
 | |
| 					".2f") "\n",
 | |
| 				str, PPS(pps), DROP(drop), err, "bulk-avg");
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_cpumap_remote(struct stats_record *stats_rec,
 | |
| 				    struct stats_record *stats_prev,
 | |
| 				    unsigned int nr_cpus)
 | |
| {
 | |
| 	double xdp_pass, xdp_drop, xdp_redirect;
 | |
| 	struct record *rec, *prev;
 | |
| 	double t;
 | |
| 	int i;
 | |
| 
 | |
| 	rec = &stats_rec->kthread;
 | |
| 	prev = &stats_prev->kthread;
 | |
| 	t = calc_period(rec, prev);
 | |
| 
 | |
| 	calc_xdp_pps(&rec->total, &prev->total, &xdp_pass, &xdp_drop,
 | |
| 		     &xdp_redirect, t);
 | |
| 	if (xdp_pass || xdp_drop || xdp_redirect) {
 | |
| 		print_err(xdp_drop,
 | |
| 			  "    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf "\n",
 | |
| 			  "xdp_stats", PASS(xdp_pass), DROP(xdp_drop),
 | |
| 			  REDIR(xdp_redirect));
 | |
| 	}
 | |
| 
 | |
| 	for (i = 0; i < nr_cpus; i++) {
 | |
| 		struct datarec *r = &rec->cpu[i];
 | |
| 		struct datarec *p = &prev->cpu[i];
 | |
| 		char str[64];
 | |
| 
 | |
| 		calc_xdp_pps(r, p, &xdp_pass, &xdp_drop, &xdp_redirect, t);
 | |
| 		if (!xdp_pass && !xdp_drop && !xdp_redirect)
 | |
| 			continue;
 | |
| 
 | |
| 		snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 		print_default("      %-16s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 | |
| 			      "\n",
 | |
| 			      str, PASS(xdp_pass), DROP(xdp_drop),
 | |
| 			      REDIR(xdp_redirect));
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_cpumap_kthread(struct stats_record *stats_rec,
 | |
| 				     struct stats_record *stats_prev,
 | |
| 				     unsigned int nr_cpus)
 | |
| {
 | |
| 	struct record *rec, *prev;
 | |
| 	double t, pps, drop, err;
 | |
| 	int i;
 | |
| 
 | |
| 	rec = &stats_rec->kthread;
 | |
| 	prev = &stats_prev->kthread;
 | |
| 	t = calc_period(rec, prev);
 | |
| 
 | |
| 	pps = calc_pps(&rec->total, &prev->total, t);
 | |
| 	drop = calc_drop_pps(&rec->total, &prev->total, t);
 | |
| 	err = calc_errs_pps(&rec->total, &prev->total, t);
 | |
| 
 | |
| 	print_err(drop, "  %-20s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf "\n",
 | |
| 		  pps ? "kthread total" : "kthread", PPS(pps), DROP(drop), err,
 | |
| 		  "sched");
 | |
| 
 | |
| 	for (i = 0; i < nr_cpus; i++) {
 | |
| 		struct datarec *r = &rec->cpu[i];
 | |
| 		struct datarec *p = &prev->cpu[i];
 | |
| 		char str[64];
 | |
| 
 | |
| 		pps = calc_pps(r, p, t);
 | |
| 		drop = calc_drop_pps(r, p, t);
 | |
| 		err = calc_errs_pps(r, p, t);
 | |
| 		if (!pps && !drop && !err)
 | |
| 			continue;
 | |
| 
 | |
| 		snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 		print_default("    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 | |
| 			      "\n",
 | |
| 			      str, PPS(pps), DROP(drop), err, "sched");
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_redirect_cnt(struct stats_record *stats_rec,
 | |
| 				   struct stats_record *stats_prev,
 | |
| 				   unsigned int nr_cpus,
 | |
| 				   struct sample_output *out)
 | |
| {
 | |
| 	struct record *rec, *prev;
 | |
| 	double t, pps;
 | |
| 	int i;
 | |
| 
 | |
| 	rec = &stats_rec->redir_err[0];
 | |
| 	prev = &stats_prev->redir_err[0];
 | |
| 	t = calc_period(rec, prev);
 | |
| 	for (i = 0; i < nr_cpus; i++) {
 | |
| 		struct datarec *r = &rec->cpu[i];
 | |
| 		struct datarec *p = &prev->cpu[i];
 | |
| 		char str[64];
 | |
| 
 | |
| 		pps = calc_pps(r, p, t);
 | |
| 		if (!pps)
 | |
| 			continue;
 | |
| 
 | |
| 		snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 		print_default("    %-18s " FMT_COLUMNf "\n", str, REDIR(pps));
 | |
| 	}
 | |
| 
 | |
| 	if (out) {
 | |
| 		pps = calc_pps(&rec->total, &prev->total, t);
 | |
| 		out->redir_cnt.suc = pps;
 | |
| 		out->totals.redir += pps;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_redirect_err_cnt(struct stats_record *stats_rec,
 | |
| 				       struct stats_record *stats_prev,
 | |
| 				       unsigned int nr_cpus,
 | |
| 				       struct sample_output *out)
 | |
| {
 | |
| 	struct record *rec, *prev;
 | |
| 	double t, drop, sum = 0;
 | |
| 	int rec_i, i;
 | |
| 
 | |
| 	for (rec_i = 1; rec_i < XDP_REDIRECT_ERR_MAX; rec_i++) {
 | |
| 		char str[64];
 | |
| 
 | |
| 		rec = &stats_rec->redir_err[rec_i];
 | |
| 		prev = &stats_prev->redir_err[rec_i];
 | |
| 		t = calc_period(rec, prev);
 | |
| 
 | |
| 		drop = calc_drop_pps(&rec->total, &prev->total, t);
 | |
| 		if (drop > 0 && !out) {
 | |
| 			snprintf(str, sizeof(str),
 | |
| 				 sample_log_level & LL_DEFAULT ? "%s total" :
 | |
| 								       "%s",
 | |
| 				 xdp_redirect_err_names[rec_i]);
 | |
| 			print_err(drop, "    %-18s " FMT_COLUMNf "\n", str,
 | |
| 				  ERR(drop));
 | |
| 		}
 | |
| 
 | |
| 		for (i = 0; i < nr_cpus; i++) {
 | |
| 			struct datarec *r = &rec->cpu[i];
 | |
| 			struct datarec *p = &prev->cpu[i];
 | |
| 			double drop;
 | |
| 
 | |
| 			drop = calc_drop_pps(r, p, t);
 | |
| 			if (!drop)
 | |
| 				continue;
 | |
| 
 | |
| 			snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 			print_default("       %-16s" FMT_COLUMNf "\n", str,
 | |
| 				      ERR(drop));
 | |
| 		}
 | |
| 
 | |
| 		sum += drop;
 | |
| 	}
 | |
| 
 | |
| 	if (out) {
 | |
| 		out->redir_cnt.err = sum;
 | |
| 		out->totals.err += sum;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_exception_cnt(struct stats_record *stats_rec,
 | |
| 				    struct stats_record *stats_prev,
 | |
| 				    unsigned int nr_cpus,
 | |
| 				    struct sample_output *out)
 | |
| {
 | |
| 	double t, drop, sum = 0;
 | |
| 	struct record *rec, *prev;
 | |
| 	int rec_i, i;
 | |
| 
 | |
| 	for (rec_i = 0; rec_i < XDP_ACTION_MAX; rec_i++) {
 | |
| 		rec = &stats_rec->exception[rec_i];
 | |
| 		prev = &stats_prev->exception[rec_i];
 | |
| 		t = calc_period(rec, prev);
 | |
| 
 | |
| 		drop = calc_drop_pps(&rec->total, &prev->total, t);
 | |
| 		/* Fold out errors after heading */
 | |
| 		sum += drop;
 | |
| 
 | |
| 		if (drop > 0 && !out) {
 | |
| 			print_always("    %-18s " FMT_COLUMNf "\n",
 | |
| 				     action2str(rec_i), ERR(drop));
 | |
| 
 | |
| 			for (i = 0; i < nr_cpus; i++) {
 | |
| 				struct datarec *r = &rec->cpu[i];
 | |
| 				struct datarec *p = &prev->cpu[i];
 | |
| 				char str[64];
 | |
| 				double drop;
 | |
| 
 | |
| 				drop = calc_drop_pps(r, p, t);
 | |
| 				if (!drop)
 | |
| 					continue;
 | |
| 
 | |
| 				snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 				print_default("       %-16s" FMT_COLUMNf "\n",
 | |
| 					      str, ERR(drop));
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (out) {
 | |
| 		out->except_cnt.hits = sum;
 | |
| 		out->totals.err += sum;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_devmap_xmit(struct stats_record *stats_rec,
 | |
| 				  struct stats_record *stats_prev,
 | |
| 				  unsigned int nr_cpus,
 | |
| 				  struct sample_output *out)
 | |
| {
 | |
| 	double pps, drop, info, err;
 | |
| 	struct record *rec, *prev;
 | |
| 	double t;
 | |
| 	int i;
 | |
| 
 | |
| 	rec = &stats_rec->devmap_xmit;
 | |
| 	prev = &stats_prev->devmap_xmit;
 | |
| 	t = calc_period(rec, prev);
 | |
| 	for (i = 0; i < nr_cpus; i++) {
 | |
| 		struct datarec *r = &rec->cpu[i];
 | |
| 		struct datarec *p = &prev->cpu[i];
 | |
| 		char str[64];
 | |
| 
 | |
| 		pps = calc_pps(r, p, t);
 | |
| 		drop = calc_drop_pps(r, p, t);
 | |
| 		err = calc_errs_pps(r, p, t);
 | |
| 
 | |
| 		if (!pps && !drop && !err)
 | |
| 			continue;
 | |
| 
 | |
| 		snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 		info = calc_info_pps(r, p, t);
 | |
| 		if (info > 0)
 | |
| 			info = (pps + drop) / info; /* calc avg bulk */
 | |
| 		print_default("     %-18s" FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 | |
| 				      __COLUMN(".2f") "\n",
 | |
| 			      str, XMIT(pps), DROP(drop), err, "drv_err/s",
 | |
| 			      info, "bulk-avg");
 | |
| 	}
 | |
| 	if (out) {
 | |
| 		pps = calc_pps(&rec->total, &prev->total, t);
 | |
| 		drop = calc_drop_pps(&rec->total, &prev->total, t);
 | |
| 		info = calc_info_pps(&rec->total, &prev->total, t);
 | |
| 		if (info > 0)
 | |
| 			info = (pps + drop) / info; /* calc avg bulk */
 | |
| 		err = calc_errs_pps(&rec->total, &prev->total, t);
 | |
| 
 | |
| 		out->xmit_cnt.pps = pps;
 | |
| 		out->xmit_cnt.drop = drop;
 | |
| 		out->xmit_cnt.bavg = info;
 | |
| 		out->xmit_cnt.err = err;
 | |
| 		out->totals.xmit += pps;
 | |
| 		out->totals.drop_xmit += drop;
 | |
| 		out->totals.err += err;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_get_devmap_xmit_multi(struct stats_record *stats_rec,
 | |
| 					struct stats_record *stats_prev,
 | |
| 					unsigned int nr_cpus,
 | |
| 					struct sample_output *out,
 | |
| 					bool xmit_total)
 | |
| {
 | |
| 	double pps, drop, info, err;
 | |
| 	struct map_entry *entry;
 | |
| 	struct record *r, *p;
 | |
| 	double t;
 | |
| 	int bkt;
 | |
| 
 | |
| 	hash_for_each(stats_rec->xmit_map, bkt, entry, node) {
 | |
| 		struct map_entry *e, *x = NULL;
 | |
| 		char ifname_from[IFNAMSIZ];
 | |
| 		char ifname_to[IFNAMSIZ];
 | |
| 		const char *fstr, *tstr;
 | |
| 		unsigned long prev_time;
 | |
| 		struct record beg = {};
 | |
| 		__u32 from_idx, to_idx;
 | |
| 		char str[128];
 | |
| 		__u64 pair;
 | |
| 		int i;
 | |
| 
 | |
| 		prev_time = sample_interval * NANOSEC_PER_SEC;
 | |
| 
 | |
| 		pair = entry->pair;
 | |
| 		from_idx = pair >> 32;
 | |
| 		to_idx = pair & 0xFFFFFFFF;
 | |
| 
 | |
| 		r = &entry->val;
 | |
| 		beg.timestamp = r->timestamp - prev_time;
 | |
| 
 | |
| 		/* Find matching entry from stats_prev map */
 | |
| 		hash_for_each_possible(stats_prev->xmit_map, e, node, pair) {
 | |
| 			if (e->pair == pair) {
 | |
| 				x = e;
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 		if (x)
 | |
| 			p = &x->val;
 | |
| 		else
 | |
| 			p = &beg;
 | |
| 		t = calc_period(r, p);
 | |
| 		pps = calc_pps(&r->total, &p->total, t);
 | |
| 		drop = calc_drop_pps(&r->total, &p->total, t);
 | |
| 		info = calc_info_pps(&r->total, &p->total, t);
 | |
| 		if (info > 0)
 | |
| 			info = (pps + drop) / info; /* calc avg bulk */
 | |
| 		err = calc_errs_pps(&r->total, &p->total, t);
 | |
| 
 | |
| 		if (out) {
 | |
| 			/* We are responsible for filling out totals */
 | |
| 			out->totals.xmit += pps;
 | |
| 			out->totals.drop_xmit += drop;
 | |
| 			out->totals.err += err;
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		fstr = tstr = NULL;
 | |
| 		if (if_indextoname(from_idx, ifname_from))
 | |
| 			fstr = ifname_from;
 | |
| 		if (if_indextoname(to_idx, ifname_to))
 | |
| 			tstr = ifname_to;
 | |
| 
 | |
| 		snprintf(str, sizeof(str), "xmit %s->%s", fstr ?: "?",
 | |
| 			 tstr ?: "?");
 | |
| 		/* Skip idle streams of redirection */
 | |
| 		if (pps || drop || err) {
 | |
| 			print_err(drop,
 | |
| 				  "  %-20s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 | |
| 				  __COLUMN(".2f") "\n", str, XMIT(pps), DROP(drop),
 | |
| 				  err, "drv_err/s", info, "bulk-avg");
 | |
| 		}
 | |
| 
 | |
| 		for (i = 0; i < nr_cpus; i++) {
 | |
| 			struct datarec *rc = &r->cpu[i];
 | |
| 			struct datarec *pc, p_beg = {};
 | |
| 			char str[64];
 | |
| 
 | |
| 			pc = p == &beg ? &p_beg : &p->cpu[i];
 | |
| 
 | |
| 			pps = calc_pps(rc, pc, t);
 | |
| 			drop = calc_drop_pps(rc, pc, t);
 | |
| 			err = calc_errs_pps(rc, pc, t);
 | |
| 
 | |
| 			if (!pps && !drop && !err)
 | |
| 				continue;
 | |
| 
 | |
| 			snprintf(str, sizeof(str), "cpu:%d", i);
 | |
| 			info = calc_info_pps(rc, pc, t);
 | |
| 			if (info > 0)
 | |
| 				info = (pps + drop) / info; /* calc avg bulk */
 | |
| 
 | |
| 			print_default("     %-18s" FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf
 | |
| 				      __COLUMN(".2f") "\n", str, XMIT(pps),
 | |
| 				      DROP(drop), err, "drv_err/s", info, "bulk-avg");
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void stats_print(const char *prefix, int mask, struct stats_record *r,
 | |
| 			struct stats_record *p, struct sample_output *out)
 | |
| {
 | |
| 	int nr_cpus = libbpf_num_possible_cpus();
 | |
| 	const char *str;
 | |
| 
 | |
| 	print_always("%-23s", prefix ?: "Summary");
 | |
| 	if (mask & SAMPLE_RX_CNT)
 | |
| 		print_always(FMT_COLUMNl, RX(out->totals.rx));
 | |
| 	if (mask & SAMPLE_REDIRECT_CNT)
 | |
| 		print_always(FMT_COLUMNl, REDIR(out->totals.redir));
 | |
| 	printf(FMT_COLUMNl,
 | |
| 	       out->totals.err + out->totals.drop + out->totals.drop_xmit,
 | |
| 	       "err,drop/s");
 | |
| 	if (mask & SAMPLE_DEVMAP_XMIT_CNT ||
 | |
| 	    mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
 | |
| 		printf(FMT_COLUMNl, XMIT(out->totals.xmit));
 | |
| 	printf("\n");
 | |
| 
 | |
| 	if (mask & SAMPLE_RX_CNT) {
 | |
| 		str = (sample_log_level & LL_DEFAULT) && out->rx_cnt.pps ?
 | |
| 				    "receive total" :
 | |
| 				    "receive";
 | |
| 		print_err((out->rx_cnt.err || out->rx_cnt.drop),
 | |
| 			  "  %-20s " FMT_COLUMNl FMT_COLUMNl FMT_COLUMNl "\n",
 | |
| 			  str, PPS(out->rx_cnt.pps), DROP(out->rx_cnt.drop),
 | |
| 			  ERR(out->rx_cnt.err));
 | |
| 
 | |
| 		stats_get_rx_cnt(r, p, nr_cpus, NULL);
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_CPUMAP_ENQUEUE_CNT)
 | |
| 		stats_get_cpumap_enqueue(r, p, nr_cpus);
 | |
| 
 | |
| 	if (mask & SAMPLE_CPUMAP_KTHREAD_CNT) {
 | |
| 		stats_get_cpumap_kthread(r, p, nr_cpus);
 | |
| 		stats_get_cpumap_remote(r, p, nr_cpus);
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_REDIRECT_CNT) {
 | |
| 		str = out->redir_cnt.suc ? "redirect total" : "redirect";
 | |
| 		print_default("  %-20s " FMT_COLUMNl "\n", str,
 | |
| 			      REDIR(out->redir_cnt.suc));
 | |
| 
 | |
| 		stats_get_redirect_cnt(r, p, nr_cpus, NULL);
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_REDIRECT_ERR_CNT) {
 | |
| 		str = (sample_log_level & LL_DEFAULT) && out->redir_cnt.err ?
 | |
| 				    "redirect_err total" :
 | |
| 				    "redirect_err";
 | |
| 		print_err(out->redir_cnt.err, "  %-20s " FMT_COLUMNl "\n", str,
 | |
| 			  ERR(out->redir_cnt.err));
 | |
| 
 | |
| 		stats_get_redirect_err_cnt(r, p, nr_cpus, NULL);
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_EXCEPTION_CNT) {
 | |
| 		str = out->except_cnt.hits ? "xdp_exception total" :
 | |
| 						   "xdp_exception";
 | |
| 
 | |
| 		print_err(out->except_cnt.hits, "  %-20s " FMT_COLUMNl "\n", str,
 | |
| 			  HITS(out->except_cnt.hits));
 | |
| 
 | |
| 		stats_get_exception_cnt(r, p, nr_cpus, NULL);
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_DEVMAP_XMIT_CNT) {
 | |
| 		str = (sample_log_level & LL_DEFAULT) && out->xmit_cnt.pps ?
 | |
| 				    "devmap_xmit total" :
 | |
| 				    "devmap_xmit";
 | |
| 
 | |
| 		print_err(out->xmit_cnt.err || out->xmit_cnt.drop,
 | |
| 			  "  %-20s " FMT_COLUMNl FMT_COLUMNl FMT_COLUMNl
 | |
| 				  __COLUMN(".2f") "\n",
 | |
| 			  str, XMIT(out->xmit_cnt.pps),
 | |
| 			  DROP(out->xmit_cnt.drop), out->xmit_cnt.err,
 | |
| 			  "drv_err/s", out->xmit_cnt.bavg, "bulk-avg");
 | |
| 
 | |
| 		stats_get_devmap_xmit(r, p, nr_cpus, NULL);
 | |
| 	}
 | |
| 
 | |
| 	if (mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
 | |
| 		stats_get_devmap_xmit_multi(r, p, nr_cpus, NULL,
 | |
| 					    mask & SAMPLE_DEVMAP_XMIT_CNT);
 | |
| 
 | |
| 	if (sample_log_level & LL_DEFAULT ||
 | |
| 	    ((sample_log_level & LL_SIMPLE) && sample_err_exp)) {
 | |
| 		sample_err_exp = false;
 | |
| 		printf("\n");
 | |
| 	}
 | |
| }
 | |
| 
 | |
| int sample_setup_maps(struct bpf_map **maps)
 | |
| {
 | |
| 	sample_n_cpus = libbpf_num_possible_cpus();
 | |
| 
 | |
| 	for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) {
 | |
| 		sample_map[i] = maps[i];
 | |
| 
 | |
| 		switch (i) {
 | |
| 		case MAP_RX:
 | |
| 		case MAP_CPUMAP_KTHREAD:
 | |
| 		case MAP_DEVMAP_XMIT:
 | |
| 			sample_map_count[i] = sample_n_cpus;
 | |
| 			break;
 | |
| 		case MAP_REDIRECT_ERR:
 | |
| 			sample_map_count[i] =
 | |
| 				XDP_REDIRECT_ERR_MAX * sample_n_cpus;
 | |
| 			break;
 | |
| 		case MAP_EXCEPTION:
 | |
| 			sample_map_count[i] = XDP_ACTION_MAX * sample_n_cpus;
 | |
| 		case MAP_CPUMAP_ENQUEUE:
 | |
| 			sample_map_count[i] = sample_n_cpus * sample_n_cpus;
 | |
| 			break;
 | |
| 		default:
 | |
| 			return -EINVAL;
 | |
| 		}
 | |
| 		if (bpf_map__set_max_entries(sample_map[i], sample_map_count[i]) < 0)
 | |
| 			return -errno;
 | |
| 	}
 | |
| 	sample_map[MAP_DEVMAP_XMIT_MULTI] = maps[MAP_DEVMAP_XMIT_MULTI];
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int sample_setup_maps_mappings(void)
 | |
| {
 | |
| 	for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) {
 | |
| 		size_t size = sample_map_count[i] * sizeof(struct datarec);
 | |
| 
 | |
| 		sample_mmap[i] = mmap(NULL, size, PROT_READ | PROT_WRITE,
 | |
| 				      MAP_SHARED, bpf_map__fd(sample_map[i]), 0);
 | |
| 		if (sample_mmap[i] == MAP_FAILED)
 | |
| 			return -errno;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int __sample_init(int mask)
 | |
| {
 | |
| 	sigset_t st;
 | |
| 
 | |
| 	sigemptyset(&st);
 | |
| 	sigaddset(&st, SIGQUIT);
 | |
| 	sigaddset(&st, SIGINT);
 | |
| 	sigaddset(&st, SIGTERM);
 | |
| 
 | |
| 	if (sigprocmask(SIG_BLOCK, &st, NULL) < 0)
 | |
| 		return -errno;
 | |
| 
 | |
| 	sample_sig_fd = signalfd(-1, &st, SFD_CLOEXEC | SFD_NONBLOCK);
 | |
| 	if (sample_sig_fd < 0)
 | |
| 		return -errno;
 | |
| 
 | |
| 	sample_mask = mask;
 | |
| 
 | |
| 	return sample_setup_maps_mappings();
 | |
| }
 | |
| 
 | |
| static int __sample_remove_xdp(int ifindex, __u32 prog_id, int xdp_flags)
 | |
| {
 | |
| 	__u32 cur_prog_id = 0;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (prog_id) {
 | |
| 		ret = bpf_xdp_query_id(ifindex, xdp_flags, &cur_prog_id);
 | |
| 		if (ret < 0)
 | |
| 			return -errno;
 | |
| 
 | |
| 		if (prog_id != cur_prog_id) {
 | |
| 			print_always(
 | |
| 				"Program on ifindex %d does not match installed "
 | |
| 				"program, skipping unload\n",
 | |
| 				ifindex);
 | |
| 			return -ENOENT;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return bpf_xdp_detach(ifindex, xdp_flags, NULL);
 | |
| }
 | |
| 
 | |
| int sample_install_xdp(struct bpf_program *xdp_prog, int ifindex, bool generic,
 | |
| 		       bool force)
 | |
| {
 | |
| 	int ret, xdp_flags = 0;
 | |
| 	__u32 prog_id = 0;
 | |
| 
 | |
| 	if (sample_xdp_cnt == 32) {
 | |
| 		fprintf(stderr,
 | |
| 			"Total limit for installed XDP programs in a sample reached\n");
 | |
| 		return -ENOTSUP;
 | |
| 	}
 | |
| 
 | |
| 	xdp_flags |= !force ? XDP_FLAGS_UPDATE_IF_NOEXIST : 0;
 | |
| 	xdp_flags |= generic ? XDP_FLAGS_SKB_MODE : XDP_FLAGS_DRV_MODE;
 | |
| 	ret = bpf_xdp_attach(ifindex, bpf_program__fd(xdp_prog), xdp_flags, NULL);
 | |
| 	if (ret < 0) {
 | |
| 		ret = -errno;
 | |
| 		fprintf(stderr,
 | |
| 			"Failed to install program \"%s\" on ifindex %d, mode = %s, "
 | |
| 			"force = %s: %s\n",
 | |
| 			bpf_program__name(xdp_prog), ifindex,
 | |
| 			generic ? "skb" : "native", force ? "true" : "false",
 | |
| 			strerror(-ret));
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	ret = bpf_xdp_query_id(ifindex, xdp_flags, &prog_id);
 | |
| 	if (ret < 0) {
 | |
| 		ret = -errno;
 | |
| 		fprintf(stderr,
 | |
| 			"Failed to get XDP program id for ifindex %d, removing program: %s\n",
 | |
| 			ifindex, strerror(errno));
 | |
| 		__sample_remove_xdp(ifindex, 0, xdp_flags);
 | |
| 		return ret;
 | |
| 	}
 | |
| 	sample_xdp_progs[sample_xdp_cnt++] =
 | |
| 		(struct xdp_desc){ ifindex, prog_id, xdp_flags };
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void sample_summary_print(void)
 | |
| {
 | |
| 	double num = sample_out.rx_cnt.num;
 | |
| 
 | |
| 	if (sample_out.totals.rx) {
 | |
| 		double pkts = sample_out.totals.rx;
 | |
| 
 | |
| 		print_always("  Packets received    : %'-10llu\n",
 | |
| 			     sample_out.totals.rx);
 | |
| 		print_always("  Average packets/s   : %'-10.0f\n",
 | |
| 			     sample_round(pkts / num));
 | |
| 	}
 | |
| 	if (sample_out.totals.redir) {
 | |
| 		double pkts = sample_out.totals.redir;
 | |
| 
 | |
| 		print_always("  Packets redirected  : %'-10llu\n",
 | |
| 			     sample_out.totals.redir);
 | |
| 		print_always("  Average redir/s     : %'-10.0f\n",
 | |
| 			     sample_round(pkts / num));
 | |
| 	}
 | |
| 	if (sample_out.totals.drop)
 | |
| 		print_always("  Rx dropped          : %'-10llu\n",
 | |
| 			     sample_out.totals.drop);
 | |
| 	if (sample_out.totals.drop_xmit)
 | |
| 		print_always("  Tx dropped          : %'-10llu\n",
 | |
| 			     sample_out.totals.drop_xmit);
 | |
| 	if (sample_out.totals.err)
 | |
| 		print_always("  Errors recorded     : %'-10llu\n",
 | |
| 			     sample_out.totals.err);
 | |
| 	if (sample_out.totals.xmit) {
 | |
| 		double pkts = sample_out.totals.xmit;
 | |
| 
 | |
| 		print_always("  Packets transmitted : %'-10llu\n",
 | |
| 			     sample_out.totals.xmit);
 | |
| 		print_always("  Average transmit/s  : %'-10.0f\n",
 | |
| 			     sample_round(pkts / num));
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void sample_exit(int status)
 | |
| {
 | |
| 	size_t size;
 | |
| 
 | |
| 	for (int i = 0; i < NUM_MAP; i++) {
 | |
| 		size = sample_map_count[i] * sizeof(**sample_mmap);
 | |
| 		munmap(sample_mmap[i], size);
 | |
| 	}
 | |
| 	while (sample_xdp_cnt--) {
 | |
| 		int i = sample_xdp_cnt, ifindex, xdp_flags;
 | |
| 		__u32 prog_id;
 | |
| 
 | |
| 		prog_id = sample_xdp_progs[i].prog_id;
 | |
| 		ifindex = sample_xdp_progs[i].ifindex;
 | |
| 		xdp_flags = sample_xdp_progs[i].flags;
 | |
| 
 | |
| 		__sample_remove_xdp(ifindex, prog_id, xdp_flags);
 | |
| 	}
 | |
| 	sample_summary_print();
 | |
| 	close(sample_sig_fd);
 | |
| 	exit(status);
 | |
| }
 | |
| 
 | |
| static int sample_stats_collect(struct stats_record *rec)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_RX_CNT)
 | |
| 		map_collect_percpu(sample_mmap[MAP_RX], &rec->rx_cnt);
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_REDIRECT_CNT)
 | |
| 		map_collect_percpu(sample_mmap[MAP_REDIRECT_ERR], &rec->redir_err[0]);
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_REDIRECT_ERR_CNT) {
 | |
| 		for (i = 1; i < XDP_REDIRECT_ERR_MAX; i++)
 | |
| 			map_collect_percpu(&sample_mmap[MAP_REDIRECT_ERR][i * sample_n_cpus],
 | |
| 					   &rec->redir_err[i]);
 | |
| 	}
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_CPUMAP_ENQUEUE_CNT)
 | |
| 		for (i = 0; i < sample_n_cpus; i++)
 | |
| 			map_collect_percpu(&sample_mmap[MAP_CPUMAP_ENQUEUE][i * sample_n_cpus],
 | |
| 					   &rec->enq[i]);
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_CPUMAP_KTHREAD_CNT)
 | |
| 		map_collect_percpu(sample_mmap[MAP_CPUMAP_KTHREAD],
 | |
| 				   &rec->kthread);
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_EXCEPTION_CNT)
 | |
| 		for (i = 0; i < XDP_ACTION_MAX; i++)
 | |
| 			map_collect_percpu(&sample_mmap[MAP_EXCEPTION][i * sample_n_cpus],
 | |
| 					   &rec->exception[i]);
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT)
 | |
| 		map_collect_percpu(sample_mmap[MAP_DEVMAP_XMIT], &rec->devmap_xmit);
 | |
| 
 | |
| 	if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) {
 | |
| 		if (map_collect_percpu_devmap(bpf_map__fd(sample_map[MAP_DEVMAP_XMIT_MULTI]), rec) < 0)
 | |
| 			return -EINVAL;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void sample_summary_update(struct sample_output *out)
 | |
| {
 | |
| 	sample_out.totals.rx += out->totals.rx;
 | |
| 	sample_out.totals.redir += out->totals.redir;
 | |
| 	sample_out.totals.drop += out->totals.drop;
 | |
| 	sample_out.totals.drop_xmit += out->totals.drop_xmit;
 | |
| 	sample_out.totals.err += out->totals.err;
 | |
| 	sample_out.totals.xmit += out->totals.xmit;
 | |
| 	sample_out.rx_cnt.num++;
 | |
| }
 | |
| 
 | |
| static void sample_stats_print(int mask, struct stats_record *cur,
 | |
| 			       struct stats_record *prev, char *prog_name)
 | |
| {
 | |
| 	struct sample_output out = {};
 | |
| 
 | |
| 	if (mask & SAMPLE_RX_CNT)
 | |
| 		stats_get_rx_cnt(cur, prev, 0, &out);
 | |
| 	if (mask & SAMPLE_REDIRECT_CNT)
 | |
| 		stats_get_redirect_cnt(cur, prev, 0, &out);
 | |
| 	if (mask & SAMPLE_REDIRECT_ERR_CNT)
 | |
| 		stats_get_redirect_err_cnt(cur, prev, 0, &out);
 | |
| 	if (mask & SAMPLE_EXCEPTION_CNT)
 | |
| 		stats_get_exception_cnt(cur, prev, 0, &out);
 | |
| 	if (mask & SAMPLE_DEVMAP_XMIT_CNT)
 | |
| 		stats_get_devmap_xmit(cur, prev, 0, &out);
 | |
| 	else if (mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI)
 | |
| 		stats_get_devmap_xmit_multi(cur, prev, 0, &out,
 | |
| 					    mask & SAMPLE_DEVMAP_XMIT_CNT);
 | |
| 	sample_summary_update(&out);
 | |
| 
 | |
| 	stats_print(prog_name, mask, cur, prev, &out);
 | |
| }
 | |
| 
 | |
| void sample_switch_mode(void)
 | |
| {
 | |
| 	sample_log_level ^= LL_DEBUG - 1;
 | |
| }
 | |
| 
 | |
| static int sample_signal_cb(void)
 | |
| {
 | |
| 	struct signalfd_siginfo si;
 | |
| 	int r;
 | |
| 
 | |
| 	r = read(sample_sig_fd, &si, sizeof(si));
 | |
| 	if (r < 0)
 | |
| 		return -errno;
 | |
| 
 | |
| 	switch (si.ssi_signo) {
 | |
| 	case SIGQUIT:
 | |
| 		sample_switch_mode();
 | |
| 		printf("\n");
 | |
| 		break;
 | |
| 	default:
 | |
| 		printf("\n");
 | |
| 		return 1;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Pointer swap trick */
 | |
| static void swap(struct stats_record **a, struct stats_record **b)
 | |
| {
 | |
| 	struct stats_record *tmp;
 | |
| 
 | |
| 	tmp = *a;
 | |
| 	*a = *b;
 | |
| 	*b = tmp;
 | |
| }
 | |
| 
 | |
| static int sample_timer_cb(int timerfd, struct stats_record **rec,
 | |
| 			   struct stats_record **prev)
 | |
| {
 | |
| 	char line[64] = "Summary";
 | |
| 	int ret;
 | |
| 	__u64 t;
 | |
| 
 | |
| 	ret = read(timerfd, &t, sizeof(t));
 | |
| 	if (ret < 0)
 | |
| 		return -errno;
 | |
| 
 | |
| 	swap(prev, rec);
 | |
| 	ret = sample_stats_collect(*rec);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (sample_xdp_cnt == 2 && !(sample_mask & SAMPLE_SKIP_HEADING)) {
 | |
| 		char fi[IFNAMSIZ];
 | |
| 		char to[IFNAMSIZ];
 | |
| 		const char *f, *t;
 | |
| 
 | |
| 		f = t = NULL;
 | |
| 		if (if_indextoname(sample_xdp_progs[0].ifindex, fi))
 | |
| 			f = fi;
 | |
| 		if (if_indextoname(sample_xdp_progs[1].ifindex, to))
 | |
| 			t = to;
 | |
| 
 | |
| 		snprintf(line, sizeof(line), "%s->%s", f ?: "?", t ?: "?");
 | |
| 	}
 | |
| 
 | |
| 	sample_stats_print(sample_mask, *rec, *prev, line);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int sample_run(int interval, void (*post_cb)(void *), void *ctx)
 | |
| {
 | |
| 	struct timespec ts = { interval, 0 };
 | |
| 	struct itimerspec its = { ts, ts };
 | |
| 	struct stats_record *rec, *prev;
 | |
| 	struct pollfd pfd[2] = {};
 | |
| 	int timerfd, ret;
 | |
| 
 | |
| 	if (!interval) {
 | |
| 		fprintf(stderr, "Incorrect interval 0\n");
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	sample_interval = interval;
 | |
| 	/* Pretty print numbers */
 | |
| 	setlocale(LC_NUMERIC, "en_US.UTF-8");
 | |
| 
 | |
| 	timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_CLOEXEC | TFD_NONBLOCK);
 | |
| 	if (timerfd < 0)
 | |
| 		return -errno;
 | |
| 	timerfd_settime(timerfd, 0, &its, NULL);
 | |
| 
 | |
| 	pfd[0].fd = sample_sig_fd;
 | |
| 	pfd[0].events = POLLIN;
 | |
| 
 | |
| 	pfd[1].fd = timerfd;
 | |
| 	pfd[1].events = POLLIN;
 | |
| 
 | |
| 	ret = -ENOMEM;
 | |
| 	rec = alloc_stats_record();
 | |
| 	if (!rec)
 | |
| 		goto end;
 | |
| 	prev = alloc_stats_record();
 | |
| 	if (!prev)
 | |
| 		goto end_rec;
 | |
| 
 | |
| 	ret = sample_stats_collect(rec);
 | |
| 	if (ret < 0)
 | |
| 		goto end_rec_prev;
 | |
| 
 | |
| 	for (;;) {
 | |
| 		ret = poll(pfd, 2, -1);
 | |
| 		if (ret < 0) {
 | |
| 			if (errno == EINTR)
 | |
| 				continue;
 | |
| 			else
 | |
| 				break;
 | |
| 		}
 | |
| 
 | |
| 		if (pfd[0].revents & POLLIN)
 | |
| 			ret = sample_signal_cb();
 | |
| 		else if (pfd[1].revents & POLLIN)
 | |
| 			ret = sample_timer_cb(timerfd, &rec, &prev);
 | |
| 
 | |
| 		if (ret)
 | |
| 			break;
 | |
| 
 | |
| 		if (post_cb)
 | |
| 			post_cb(ctx);
 | |
| 	}
 | |
| 
 | |
| end_rec_prev:
 | |
| 	free_stats_record(prev);
 | |
| end_rec:
 | |
| 	free_stats_record(rec);
 | |
| end:
 | |
| 	close(timerfd);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| const char *get_driver_name(int ifindex)
 | |
| {
 | |
| 	struct ethtool_drvinfo drv = {};
 | |
| 	char ifname[IF_NAMESIZE];
 | |
| 	static char drvname[32];
 | |
| 	struct ifreq ifr = {};
 | |
| 	int fd, r = 0;
 | |
| 
 | |
| 	fd = socket(AF_INET, SOCK_DGRAM, 0);
 | |
| 	if (fd < 0)
 | |
| 		return "[error]";
 | |
| 
 | |
| 	if (!if_indextoname(ifindex, ifname))
 | |
| 		goto end;
 | |
| 
 | |
| 	drv.cmd = ETHTOOL_GDRVINFO;
 | |
| 	safe_strncpy(ifr.ifr_name, ifname, sizeof(ifr.ifr_name));
 | |
| 	ifr.ifr_data = (void *)&drv;
 | |
| 
 | |
| 	r = ioctl(fd, SIOCETHTOOL, &ifr);
 | |
| 	if (r)
 | |
| 		goto end;
 | |
| 
 | |
| 	safe_strncpy(drvname, drv.driver, sizeof(drvname));
 | |
| 
 | |
| 	close(fd);
 | |
| 	return drvname;
 | |
| 
 | |
| end:
 | |
| 	r = errno;
 | |
| 	close(fd);
 | |
| 	return r == EOPNOTSUPP ? "loopback" : "[error]";
 | |
| }
 | |
| 
 | |
| int get_mac_addr(int ifindex, void *mac_addr)
 | |
| {
 | |
| 	char ifname[IF_NAMESIZE];
 | |
| 	struct ifreq ifr = {};
 | |
| 	int fd, r;
 | |
| 
 | |
| 	fd = socket(AF_INET, SOCK_DGRAM, 0);
 | |
| 	if (fd < 0)
 | |
| 		return -errno;
 | |
| 
 | |
| 	if (!if_indextoname(ifindex, ifname)) {
 | |
| 		r = -errno;
 | |
| 		goto end;
 | |
| 	}
 | |
| 
 | |
| 	safe_strncpy(ifr.ifr_name, ifname, sizeof(ifr.ifr_name));
 | |
| 
 | |
| 	r = ioctl(fd, SIOCGIFHWADDR, &ifr);
 | |
| 	if (r) {
 | |
| 		r = -errno;
 | |
| 		goto end;
 | |
| 	}
 | |
| 
 | |
| 	memcpy(mac_addr, ifr.ifr_hwaddr.sa_data, 6 * sizeof(char));
 | |
| 
 | |
| end:
 | |
| 	close(fd);
 | |
| 	return r;
 | |
| }
 | |
| 
 | |
| __attribute__((constructor)) static void sample_ctor(void)
 | |
| {
 | |
| 	if (libbpf_set_strict_mode(LIBBPF_STRICT_ALL) < 0) {
 | |
| 		fprintf(stderr, "Failed to set libbpf strict mode: %s\n",
 | |
| 			strerror(errno));
 | |
| 		/* Just exit, nothing to cleanup right now */
 | |
| 		exit(EXIT_FAIL_BPF);
 | |
| 	}
 | |
| }
 |