mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	bpf: cpumap use ptr_ring_consume_batched
Move ptr_ring dequeue outside loop, that allocate SKBs and calls network stack, as these operations that can take some time. The ptr_ring is a communication channel between CPUs, where we want to reduce/limit any cacheline bouncing. Do a concentrated bulk dequeue via ptr_ring_consume_batched, to shorten the period and times the remote cacheline in ptr_ring is read Batch size 8 is both to (1) limit BH-disable period, and (2) consume one cacheline on 64-bit archs. After reducing the BH-disable section further then we can consider changing this, while still thinking about L1 cacheline size being active. Signed-off-by: Jesper Dangaard Brouer <brouer@redhat.com> Acked-by: Song Liu <songliubraving@fb.com> Signed-off-by: Alexei Starovoitov <ast@kernel.org>
This commit is contained in:
		
							parent
							
								
									00967e84f7
								
							
						
					
					
						commit
						77361825bb
					
				
					 1 changed files with 11 additions and 10 deletions
				
			
		| 
						 | 
					@ -240,6 +240,8 @@ static void put_cpu_map_entry(struct bpf_cpu_map_entry *rcpu)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					#define CPUMAP_BATCH 8
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int cpu_map_kthread_run(void *data)
 | 
					static int cpu_map_kthread_run(void *data)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct bpf_cpu_map_entry *rcpu = data;
 | 
						struct bpf_cpu_map_entry *rcpu = data;
 | 
				
			||||||
| 
						 | 
					@ -252,8 +254,9 @@ static int cpu_map_kthread_run(void *data)
 | 
				
			||||||
	 * kthread_stop signal until queue is empty.
 | 
						 * kthread_stop signal until queue is empty.
 | 
				
			||||||
	 */
 | 
						 */
 | 
				
			||||||
	while (!kthread_should_stop() || !__ptr_ring_empty(rcpu->queue)) {
 | 
						while (!kthread_should_stop() || !__ptr_ring_empty(rcpu->queue)) {
 | 
				
			||||||
		unsigned int processed = 0, drops = 0, sched = 0;
 | 
							unsigned int drops = 0, sched = 0;
 | 
				
			||||||
		struct xdp_frame *xdpf;
 | 
							void *frames[CPUMAP_BATCH];
 | 
				
			||||||
 | 
							int i, n;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* Release CPU reschedule checks */
 | 
							/* Release CPU reschedule checks */
 | 
				
			||||||
		if (__ptr_ring_empty(rcpu->queue)) {
 | 
							if (__ptr_ring_empty(rcpu->queue)) {
 | 
				
			||||||
| 
						 | 
					@ -269,14 +272,16 @@ static int cpu_map_kthread_run(void *data)
 | 
				
			||||||
			sched = cond_resched();
 | 
								sched = cond_resched();
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		/* Process packets in rcpu->queue */
 | 
					 | 
				
			||||||
		local_bh_disable();
 | 
					 | 
				
			||||||
		/*
 | 
							/*
 | 
				
			||||||
		 * The bpf_cpu_map_entry is single consumer, with this
 | 
							 * The bpf_cpu_map_entry is single consumer, with this
 | 
				
			||||||
		 * kthread CPU pinned. Lockless access to ptr_ring
 | 
							 * kthread CPU pinned. Lockless access to ptr_ring
 | 
				
			||||||
		 * consume side valid as no-resize allowed of queue.
 | 
							 * consume side valid as no-resize allowed of queue.
 | 
				
			||||||
		 */
 | 
							 */
 | 
				
			||||||
		while ((xdpf = __ptr_ring_consume(rcpu->queue))) {
 | 
							n = ptr_ring_consume_batched(rcpu->queue, frames, CPUMAP_BATCH);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							local_bh_disable();
 | 
				
			||||||
 | 
							for (i = 0; i < n; i++) {
 | 
				
			||||||
 | 
								struct xdp_frame *xdpf = frames[i];
 | 
				
			||||||
			struct sk_buff *skb;
 | 
								struct sk_buff *skb;
 | 
				
			||||||
			int ret;
 | 
								int ret;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -290,13 +295,9 @@ static int cpu_map_kthread_run(void *data)
 | 
				
			||||||
			ret = netif_receive_skb_core(skb);
 | 
								ret = netif_receive_skb_core(skb);
 | 
				
			||||||
			if (ret == NET_RX_DROP)
 | 
								if (ret == NET_RX_DROP)
 | 
				
			||||||
				drops++;
 | 
									drops++;
 | 
				
			||||||
 | 
					 | 
				
			||||||
			/* Limit BH-disable period */
 | 
					 | 
				
			||||||
			if (++processed == 8)
 | 
					 | 
				
			||||||
				break;
 | 
					 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		/* Feedback loop via tracepoint */
 | 
							/* Feedback loop via tracepoint */
 | 
				
			||||||
		trace_xdp_cpumap_kthread(rcpu->map_id, processed, drops, sched);
 | 
							trace_xdp_cpumap_kthread(rcpu->map_id, n, drops, sched);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		local_bh_enable(); /* resched point, may call do_softirq() */
 | 
							local_bh_enable(); /* resched point, may call do_softirq() */
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue