mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 02:30:34 +02:00 
			
		
		
		
	udp: add batching to udp_rmem_release()
If udp_recvmsg() constantly releases sk_rmem_alloc for every read packet, it gives opportunity for producers to immediately grab spinlocks and desperatly try adding another packet, causing false sharing. We can add a simple heuristic to give the signal by batches of ~25 % of the queue capacity. This patch considerably increases performance under flood by about 50 %, since the thread draining the queue is no longer slowed by false sharing. Signed-off-by: Eric Dumazet <edumazet@google.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
		
							parent
							
								
									c84d949057
								
							
						
					
					
						commit
						6b229cf77d
					
				
					 2 changed files with 15 additions and 0 deletions
				
			
		| 
						 | 
				
			
			@ -79,6 +79,9 @@ struct udp_sock {
 | 
			
		|||
	int			(*gro_complete)(struct sock *sk,
 | 
			
		||||
						struct sk_buff *skb,
 | 
			
		||||
						int nhoff);
 | 
			
		||||
 | 
			
		||||
	/* This field is dirtied by udp_recvmsg() */
 | 
			
		||||
	int		forward_deficit;
 | 
			
		||||
};
 | 
			
		||||
 | 
			
		||||
static inline struct udp_sock *udp_sk(const struct sock *sk)
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -1177,8 +1177,20 @@ int udp_sendpage(struct sock *sk, struct page *page, int offset,
 | 
			
		|||
/* fully reclaim rmem/fwd memory allocated for skb */
 | 
			
		||||
static void udp_rmem_release(struct sock *sk, int size, int partial)
 | 
			
		||||
{
 | 
			
		||||
	struct udp_sock *up = udp_sk(sk);
 | 
			
		||||
	int amt;
 | 
			
		||||
 | 
			
		||||
	if (likely(partial)) {
 | 
			
		||||
		up->forward_deficit += size;
 | 
			
		||||
		size = up->forward_deficit;
 | 
			
		||||
		if (size < (sk->sk_rcvbuf >> 2) &&
 | 
			
		||||
		    !skb_queue_empty(&sk->sk_receive_queue))
 | 
			
		||||
			return;
 | 
			
		||||
	} else {
 | 
			
		||||
		size += up->forward_deficit;
 | 
			
		||||
	}
 | 
			
		||||
	up->forward_deficit = 0;
 | 
			
		||||
 | 
			
		||||
	atomic_sub(size, &sk->sk_rmem_alloc);
 | 
			
		||||
	sk->sk_forward_alloc += size;
 | 
			
		||||
	amt = (sk->sk_forward_alloc - partial) & ~(SK_MEM_QUANTUM - 1);
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue