forked from mirrors/linux
		
	 5f116212f4
			
		
	
	
		5f116212f4
		
	
	
	
	
		
			
			This adds support for the devmap_xmit tracepoint, and its multi device variant that can be used to obtain streams for each individual net_device to net_device redirection. This is useful for decomposing total xmit stats in xdp_monitor. Signed-off-by: Kumar Kartikeya Dwivedi <memxor@gmail.com> Signed-off-by: Alexei Starovoitov <ast@kernel.org> Link: https://lore.kernel.org/bpf/20210821002010.845777-11-memxor@gmail.com
		
			
				
	
	
		
			266 lines
		
	
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			266 lines
		
	
	
	
		
			7.1 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*  GPLv2, Copyright(c) 2017 Jesper Dangaard Brouer, Red Hat, Inc. */
 | |
| #include "xdp_sample.bpf.h"
 | |
| 
 | |
| #include <bpf/bpf_tracing.h>
 | |
| #include <bpf/bpf_core_read.h>
 | |
| #include <bpf/bpf_helpers.h>
 | |
| 
 | |
| array_map rx_cnt SEC(".maps");
 | |
| array_map redir_err_cnt SEC(".maps");
 | |
| array_map cpumap_enqueue_cnt SEC(".maps");
 | |
| array_map cpumap_kthread_cnt SEC(".maps");
 | |
| array_map exception_cnt SEC(".maps");
 | |
| array_map devmap_xmit_cnt SEC(".maps");
 | |
| 
 | |
| struct {
 | |
| 	__uint(type, BPF_MAP_TYPE_PERCPU_HASH);
 | |
| 	__uint(max_entries, 32 * 32);
 | |
| 	__type(key, u64);
 | |
| 	__type(value, struct datarec);
 | |
| } devmap_xmit_cnt_multi SEC(".maps");
 | |
| 
 | |
| const volatile int nr_cpus = 0;
 | |
| 
 | |
| /* These can be set before loading so that redundant comparisons can be DCE'd by
 | |
|  * the verifier, and only actual matches are tried after loading tp_btf program.
 | |
|  * This allows sample to filter tracepoint stats based on net_device.
 | |
|  */
 | |
| const volatile int from_match[32] = {};
 | |
| const volatile int to_match[32] = {};
 | |
| 
 | |
| int cpumap_map_id = 0;
 | |
| 
 | |
| /* Find if b is part of set a, but if a is empty set then evaluate to true */
 | |
| #define IN_SET(a, b)                                                 \
 | |
| 	({                                                           \
 | |
| 		bool __res = !(a)[0];                                \
 | |
| 		for (int i = 0; i < ARRAY_SIZE(a) && (a)[i]; i++) { \
 | |
| 			__res = (a)[i] == (b);                       \
 | |
| 			if (__res)                                   \
 | |
| 				break;                               \
 | |
| 		}                                                    \
 | |
| 		__res;                                               \
 | |
| 	})
 | |
| 
 | |
| static __always_inline __u32 xdp_get_err_key(int err)
 | |
| {
 | |
| 	switch (err) {
 | |
| 	case 0:
 | |
| 		return 0;
 | |
| 	case -EINVAL:
 | |
| 		return 2;
 | |
| 	case -ENETDOWN:
 | |
| 		return 3;
 | |
| 	case -EMSGSIZE:
 | |
| 		return 4;
 | |
| 	case -EOPNOTSUPP:
 | |
| 		return 5;
 | |
| 	case -ENOSPC:
 | |
| 		return 6;
 | |
| 	default:
 | |
| 		return 1;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static __always_inline int xdp_redirect_collect_stat(int from, int err)
 | |
| {
 | |
| 	u32 cpu = bpf_get_smp_processor_id();
 | |
| 	u32 key = XDP_REDIRECT_ERROR;
 | |
| 	struct datarec *rec;
 | |
| 	u32 idx;
 | |
| 
 | |
| 	if (!IN_SET(from_match, from))
 | |
| 		return 0;
 | |
| 
 | |
| 	key = xdp_get_err_key(err);
 | |
| 
 | |
| 	idx = key * nr_cpus + cpu;
 | |
| 	rec = bpf_map_lookup_elem(&redir_err_cnt, &idx);
 | |
| 	if (!rec)
 | |
| 		return 0;
 | |
| 	if (key)
 | |
| 		NO_TEAR_INC(rec->dropped);
 | |
| 	else
 | |
| 		NO_TEAR_INC(rec->processed);
 | |
| 	return 0; /* Indicate event was filtered (no further processing)*/
 | |
| 	/*
 | |
| 	 * Returning 1 here would allow e.g. a perf-record tracepoint
 | |
| 	 * to see and record these events, but it doesn't work well
 | |
| 	 * in-practice as stopping perf-record also unload this
 | |
| 	 * bpf_prog.  Plus, there is additional overhead of doing so.
 | |
| 	 */
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_redirect_err")
 | |
| int BPF_PROG(tp_xdp_redirect_err, const struct net_device *dev,
 | |
| 	     const struct bpf_prog *xdp, const void *tgt, int err,
 | |
| 	     const struct bpf_map *map, u32 index)
 | |
| {
 | |
| 	return xdp_redirect_collect_stat(dev->ifindex, err);
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_redirect_map_err")
 | |
| int BPF_PROG(tp_xdp_redirect_map_err, const struct net_device *dev,
 | |
| 	     const struct bpf_prog *xdp, const void *tgt, int err,
 | |
| 	     const struct bpf_map *map, u32 index)
 | |
| {
 | |
| 	return xdp_redirect_collect_stat(dev->ifindex, err);
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_redirect")
 | |
| int BPF_PROG(tp_xdp_redirect, const struct net_device *dev,
 | |
| 	     const struct bpf_prog *xdp, const void *tgt, int err,
 | |
| 	     const struct bpf_map *map, u32 index)
 | |
| {
 | |
| 	return xdp_redirect_collect_stat(dev->ifindex, err);
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_redirect_map")
 | |
| int BPF_PROG(tp_xdp_redirect_map, const struct net_device *dev,
 | |
| 	     const struct bpf_prog *xdp, const void *tgt, int err,
 | |
| 	     const struct bpf_map *map, u32 index)
 | |
| {
 | |
| 	return xdp_redirect_collect_stat(dev->ifindex, err);
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_cpumap_enqueue")
 | |
| int BPF_PROG(tp_xdp_cpumap_enqueue, int map_id, unsigned int processed,
 | |
| 	     unsigned int drops, int to_cpu)
 | |
| {
 | |
| 	u32 cpu = bpf_get_smp_processor_id();
 | |
| 	struct datarec *rec;
 | |
| 	u32 idx;
 | |
| 
 | |
| 	if (cpumap_map_id && cpumap_map_id != map_id)
 | |
| 		return 0;
 | |
| 
 | |
| 	idx = to_cpu * nr_cpus + cpu;
 | |
| 	rec = bpf_map_lookup_elem(&cpumap_enqueue_cnt, &idx);
 | |
| 	if (!rec)
 | |
| 		return 0;
 | |
| 	NO_TEAR_ADD(rec->processed, processed);
 | |
| 	NO_TEAR_ADD(rec->dropped, drops);
 | |
| 	/* Record bulk events, then userspace can calc average bulk size */
 | |
| 	if (processed > 0)
 | |
| 		NO_TEAR_INC(rec->issue);
 | |
| 	/* Inception: It's possible to detect overload situations, via
 | |
| 	 * this tracepoint.  This can be used for creating a feedback
 | |
| 	 * loop to XDP, which can take appropriate actions to mitigate
 | |
| 	 * this overload situation.
 | |
| 	 */
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_cpumap_kthread")
 | |
| int BPF_PROG(tp_xdp_cpumap_kthread, int map_id, unsigned int processed,
 | |
| 	     unsigned int drops, int sched, struct xdp_cpumap_stats *xdp_stats)
 | |
| {
 | |
| 	struct datarec *rec;
 | |
| 	u32 cpu;
 | |
| 
 | |
| 	if (cpumap_map_id && cpumap_map_id != map_id)
 | |
| 		return 0;
 | |
| 
 | |
| 	cpu = bpf_get_smp_processor_id();
 | |
| 	rec = bpf_map_lookup_elem(&cpumap_kthread_cnt, &cpu);
 | |
| 	if (!rec)
 | |
| 		return 0;
 | |
| 	NO_TEAR_ADD(rec->processed, processed);
 | |
| 	NO_TEAR_ADD(rec->dropped, drops);
 | |
| 	NO_TEAR_ADD(rec->xdp_pass, xdp_stats->pass);
 | |
| 	NO_TEAR_ADD(rec->xdp_drop, xdp_stats->drop);
 | |
| 	NO_TEAR_ADD(rec->xdp_redirect, xdp_stats->redirect);
 | |
| 	/* Count times kthread yielded CPU via schedule call */
 | |
| 	if (sched)
 | |
| 		NO_TEAR_INC(rec->issue);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_exception")
 | |
| int BPF_PROG(tp_xdp_exception, const struct net_device *dev,
 | |
| 	     const struct bpf_prog *xdp, u32 act)
 | |
| {
 | |
| 	u32 cpu = bpf_get_smp_processor_id();
 | |
| 	struct datarec *rec;
 | |
| 	u32 key = act, idx;
 | |
| 
 | |
| 	if (!IN_SET(from_match, dev->ifindex))
 | |
| 		return 0;
 | |
| 	if (!IN_SET(to_match, dev->ifindex))
 | |
| 		return 0;
 | |
| 
 | |
| 	if (key > XDP_REDIRECT)
 | |
| 		key = XDP_REDIRECT + 1;
 | |
| 
 | |
| 	idx = key * nr_cpus + cpu;
 | |
| 	rec = bpf_map_lookup_elem(&exception_cnt, &idx);
 | |
| 	if (!rec)
 | |
| 		return 0;
 | |
| 	NO_TEAR_INC(rec->dropped);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_devmap_xmit")
 | |
| int BPF_PROG(tp_xdp_devmap_xmit, const struct net_device *from_dev,
 | |
| 	     const struct net_device *to_dev, int sent, int drops, int err)
 | |
| {
 | |
| 	struct datarec *rec;
 | |
| 	int idx_in, idx_out;
 | |
| 	u32 cpu;
 | |
| 
 | |
| 	idx_in = from_dev->ifindex;
 | |
| 	idx_out = to_dev->ifindex;
 | |
| 
 | |
| 	if (!IN_SET(from_match, idx_in))
 | |
| 		return 0;
 | |
| 	if (!IN_SET(to_match, idx_out))
 | |
| 		return 0;
 | |
| 
 | |
| 	cpu = bpf_get_smp_processor_id();
 | |
| 	rec = bpf_map_lookup_elem(&devmap_xmit_cnt, &cpu);
 | |
| 	if (!rec)
 | |
| 		return 0;
 | |
| 	NO_TEAR_ADD(rec->processed, sent);
 | |
| 	NO_TEAR_ADD(rec->dropped, drops);
 | |
| 	/* Record bulk events, then userspace can calc average bulk size */
 | |
| 	NO_TEAR_INC(rec->info);
 | |
| 	/* Record error cases, where no frame were sent */
 | |
| 	/* Catch API error of drv ndo_xdp_xmit sent more than count */
 | |
| 	if (err || drops < 0)
 | |
| 		NO_TEAR_INC(rec->issue);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| SEC("tp_btf/xdp_devmap_xmit")
 | |
| int BPF_PROG(tp_xdp_devmap_xmit_multi, const struct net_device *from_dev,
 | |
| 	     const struct net_device *to_dev, int sent, int drops, int err)
 | |
| {
 | |
| 	struct datarec empty = {};
 | |
| 	struct datarec *rec;
 | |
| 	int idx_in, idx_out;
 | |
| 	u64 idx;
 | |
| 
 | |
| 	idx_in = from_dev->ifindex;
 | |
| 	idx_out = to_dev->ifindex;
 | |
| 	idx = idx_in;
 | |
| 	idx = idx << 32 | idx_out;
 | |
| 
 | |
| 	if (!IN_SET(from_match, idx_in))
 | |
| 		return 0;
 | |
| 	if (!IN_SET(to_match, idx_out))
 | |
| 		return 0;
 | |
| 
 | |
| 	bpf_map_update_elem(&devmap_xmit_cnt_multi, &idx, &empty, BPF_NOEXIST);
 | |
| 	rec = bpf_map_lookup_elem(&devmap_xmit_cnt_multi, &idx);
 | |
| 	if (!rec)
 | |
| 		return 0;
 | |
| 
 | |
| 	NO_TEAR_ADD(rec->processed, sent);
 | |
| 	NO_TEAR_ADD(rec->dropped, drops);
 | |
| 	NO_TEAR_INC(rec->info);
 | |
| 	if (err || drops < 0)
 | |
| 		NO_TEAR_INC(rec->issue);
 | |
| 	return 0;
 | |
| }
 |