forked from mirrors/linux
		
	xen/netfront: handle compound page fragments on transmit
An SKB paged fragment can consist of a compound page with order > 0.
However the netchannel protocol deals only in PAGE_SIZE frames.
Handle this in xennet_make_frags by iterating over the frames which
make up the page.
This is the netfront equivalent to 6a8ed462f1 for netback.
Signed-off-by: Ian Campbell <ian.campbell@citrix.com>
Cc: netdev@vger.kernel.org
Cc: xen-devel@lists.xen.org
Cc: Eric Dumazet <edumazet@google.com>
Cc: Konrad Rzeszutek Wilk <konrad@kernel.org>
Cc: ANNIE LI <annie.li@oracle.com>
Cc: Sander Eikelenboom <linux@eikelenboom.it>
Cc: Stefan Bader <stefan.bader@canonical.com>
Acked-by: Eric Dumazet <edumazet@google.com>
Acked-by: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
			
			
This commit is contained in:
		
							parent
							
								
									f30a944392
								
							
						
					
					
						commit
						f36c374782
					
				
					 1 changed files with 77 additions and 21 deletions
				
			
		| 
						 | 
					@ -452,29 +452,85 @@ static void xennet_make_frags(struct sk_buff *skb, struct net_device *dev,
 | 
				
			||||||
	/* Grant backend access to each skb fragment page. */
 | 
						/* Grant backend access to each skb fragment page. */
 | 
				
			||||||
	for (i = 0; i < frags; i++) {
 | 
						for (i = 0; i < frags; i++) {
 | 
				
			||||||
		skb_frag_t *frag = skb_shinfo(skb)->frags + i;
 | 
							skb_frag_t *frag = skb_shinfo(skb)->frags + i;
 | 
				
			||||||
 | 
							struct page *page = skb_frag_page(frag);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							len = skb_frag_size(frag);
 | 
				
			||||||
 | 
							offset = frag->page_offset;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							/* Data must not cross a page boundary. */
 | 
				
			||||||
 | 
							BUG_ON(len + offset > PAGE_SIZE<<compound_order(page));
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							/* Skip unused frames from start of page */
 | 
				
			||||||
 | 
							page += offset >> PAGE_SHIFT;
 | 
				
			||||||
 | 
							offset &= ~PAGE_MASK;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							while (len > 0) {
 | 
				
			||||||
 | 
								unsigned long bytes;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
								BUG_ON(offset >= PAGE_SIZE);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
								bytes = PAGE_SIZE - offset;
 | 
				
			||||||
 | 
								if (bytes > len)
 | 
				
			||||||
 | 
									bytes = len;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			tx->flags |= XEN_NETTXF_more_data;
 | 
								tx->flags |= XEN_NETTXF_more_data;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		id = get_id_from_freelist(&np->tx_skb_freelist, np->tx_skbs);
 | 
								id = get_id_from_freelist(&np->tx_skb_freelist,
 | 
				
			||||||
 | 
											  np->tx_skbs);
 | 
				
			||||||
			np->tx_skbs[id].skb = skb_get(skb);
 | 
								np->tx_skbs[id].skb = skb_get(skb);
 | 
				
			||||||
			tx = RING_GET_REQUEST(&np->tx, prod++);
 | 
								tx = RING_GET_REQUEST(&np->tx, prod++);
 | 
				
			||||||
			tx->id = id;
 | 
								tx->id = id;
 | 
				
			||||||
			ref = gnttab_claim_grant_reference(&np->gref_tx_head);
 | 
								ref = gnttab_claim_grant_reference(&np->gref_tx_head);
 | 
				
			||||||
			BUG_ON((signed short)ref < 0);
 | 
								BUG_ON((signed short)ref < 0);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		mfn = pfn_to_mfn(page_to_pfn(skb_frag_page(frag)));
 | 
								mfn = pfn_to_mfn(page_to_pfn(page));
 | 
				
			||||||
		gnttab_grant_foreign_access_ref(ref, np->xbdev->otherend_id,
 | 
								gnttab_grant_foreign_access_ref(ref,
 | 
				
			||||||
 | 
												np->xbdev->otherend_id,
 | 
				
			||||||
							mfn, GNTMAP_readonly);
 | 
												mfn, GNTMAP_readonly);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			tx->gref = np->grant_tx_ref[id] = ref;
 | 
								tx->gref = np->grant_tx_ref[id] = ref;
 | 
				
			||||||
		tx->offset = frag->page_offset;
 | 
								tx->offset = offset;
 | 
				
			||||||
		tx->size = skb_frag_size(frag);
 | 
								tx->size = bytes;
 | 
				
			||||||
			tx->flags = 0;
 | 
								tx->flags = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
								offset += bytes;
 | 
				
			||||||
 | 
								len -= bytes;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
								/* Next frame */
 | 
				
			||||||
 | 
								if (offset == PAGE_SIZE && len) {
 | 
				
			||||||
 | 
									BUG_ON(!PageCompound(page));
 | 
				
			||||||
 | 
									page++;
 | 
				
			||||||
 | 
									offset = 0;
 | 
				
			||||||
 | 
								}
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	np->tx.req_prod_pvt = prod;
 | 
						np->tx.req_prod_pvt = prod;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					/*
 | 
				
			||||||
 | 
					 * Count how many ring slots are required to send the frags of this
 | 
				
			||||||
 | 
					 * skb. Each frag might be a compound page.
 | 
				
			||||||
 | 
					 */
 | 
				
			||||||
 | 
					static int xennet_count_skb_frag_slots(struct sk_buff *skb)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						int i, frags = skb_shinfo(skb)->nr_frags;
 | 
				
			||||||
 | 
						int pages = 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						for (i = 0; i < frags; i++) {
 | 
				
			||||||
 | 
							skb_frag_t *frag = skb_shinfo(skb)->frags + i;
 | 
				
			||||||
 | 
							unsigned long size = skb_frag_size(frag);
 | 
				
			||||||
 | 
							unsigned long offset = frag->page_offset;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							/* Skip unused frames from start of page */
 | 
				
			||||||
 | 
							offset &= ~PAGE_MASK;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							pages += PFN_UP(offset + size);
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						return pages;
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev)
 | 
					static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	unsigned short id;
 | 
						unsigned short id;
 | 
				
			||||||
| 
						 | 
					@ -487,23 +543,23 @@ static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev)
 | 
				
			||||||
	grant_ref_t ref;
 | 
						grant_ref_t ref;
 | 
				
			||||||
	unsigned long mfn;
 | 
						unsigned long mfn;
 | 
				
			||||||
	int notify;
 | 
						int notify;
 | 
				
			||||||
	int frags = skb_shinfo(skb)->nr_frags;
 | 
						int slots;
 | 
				
			||||||
	unsigned int offset = offset_in_page(data);
 | 
						unsigned int offset = offset_in_page(data);
 | 
				
			||||||
	unsigned int len = skb_headlen(skb);
 | 
						unsigned int len = skb_headlen(skb);
 | 
				
			||||||
	unsigned long flags;
 | 
						unsigned long flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	frags += DIV_ROUND_UP(offset + len, PAGE_SIZE);
 | 
						slots = DIV_ROUND_UP(offset + len, PAGE_SIZE) +
 | 
				
			||||||
	if (unlikely(frags > MAX_SKB_FRAGS + 1)) {
 | 
							xennet_count_skb_frag_slots(skb);
 | 
				
			||||||
		printk(KERN_ALERT "xennet: skb rides the rocket: %d frags\n",
 | 
						if (unlikely(slots > MAX_SKB_FRAGS + 1)) {
 | 
				
			||||||
		       frags);
 | 
							net_alert_ratelimited(
 | 
				
			||||||
		dump_stack();
 | 
								"xennet: skb rides the rocket: %d slots\n", slots);
 | 
				
			||||||
		goto drop;
 | 
							goto drop;
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	spin_lock_irqsave(&np->tx_lock, flags);
 | 
						spin_lock_irqsave(&np->tx_lock, flags);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (unlikely(!netif_carrier_ok(dev) ||
 | 
						if (unlikely(!netif_carrier_ok(dev) ||
 | 
				
			||||||
		     (frags > 1 && !xennet_can_sg(dev)) ||
 | 
							     (slots > 1 && !xennet_can_sg(dev)) ||
 | 
				
			||||||
		     netif_needs_gso(skb, netif_skb_features(skb)))) {
 | 
							     netif_needs_gso(skb, netif_skb_features(skb)))) {
 | 
				
			||||||
		spin_unlock_irqrestore(&np->tx_lock, flags);
 | 
							spin_unlock_irqrestore(&np->tx_lock, flags);
 | 
				
			||||||
		goto drop;
 | 
							goto drop;
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue