forked from mirrors/linux
		
	ipvlan: Scrub skb before crossing the namespace boundry
The earlier patch c3aaa06d5a (ipvlan: scrub skb before routing
in L3 mode.) did this but only for TX path in L3 mode. This
patch extends it for both the modes for TX/RX path.
Signed-off-by: Mahesh Bandewar <maheshb@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
			
			
This commit is contained in:
		
							parent
							
								
									16eab5598c
								
							
						
					
					
						commit
						b93dd49c1a
					
				
					 1 changed files with 25 additions and 14 deletions
				
			
		|  | @ -254,6 +254,18 @@ void ipvlan_process_multicast(struct work_struct *work) | ||||||
| 	} | 	} | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | static void ipvlan_skb_crossing_ns(struct sk_buff *skb, struct net_device *dev) | ||||||
|  | { | ||||||
|  | 	bool xnet = true; | ||||||
|  | 
 | ||||||
|  | 	if (dev) | ||||||
|  | 		xnet = !net_eq(dev_net(skb->dev), dev_net(dev)); | ||||||
|  | 
 | ||||||
|  | 	skb_scrub_packet(skb, xnet); | ||||||
|  | 	if (dev) | ||||||
|  | 		skb->dev = dev; | ||||||
|  | } | ||||||
|  | 
 | ||||||
| static int ipvlan_rcv_frame(struct ipvl_addr *addr, struct sk_buff **pskb, | static int ipvlan_rcv_frame(struct ipvl_addr *addr, struct sk_buff **pskb, | ||||||
| 			    bool local) | 			    bool local) | ||||||
| { | { | ||||||
|  | @ -280,7 +292,7 @@ static int ipvlan_rcv_frame(struct ipvl_addr *addr, struct sk_buff **pskb, | ||||||
| 
 | 
 | ||||||
| 		*pskb = skb; | 		*pskb = skb; | ||||||
| 	} | 	} | ||||||
| 	skb->dev = dev; | 	ipvlan_skb_crossing_ns(skb, dev); | ||||||
| 
 | 
 | ||||||
| 	if (local) { | 	if (local) { | ||||||
| 		skb->pkt_type = PACKET_HOST; | 		skb->pkt_type = PACKET_HOST; | ||||||
|  | @ -347,7 +359,7 @@ static struct ipvl_addr *ipvlan_addr_lookup(struct ipvl_port *port, | ||||||
| 	return addr; | 	return addr; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int ipvlan_process_v4_outbound(struct sk_buff *skb, bool xnet) | static int ipvlan_process_v4_outbound(struct sk_buff *skb) | ||||||
| { | { | ||||||
| 	const struct iphdr *ip4h = ip_hdr(skb); | 	const struct iphdr *ip4h = ip_hdr(skb); | ||||||
| 	struct net_device *dev = skb->dev; | 	struct net_device *dev = skb->dev; | ||||||
|  | @ -370,7 +382,6 @@ static int ipvlan_process_v4_outbound(struct sk_buff *skb, bool xnet) | ||||||
| 		ip_rt_put(rt); | 		ip_rt_put(rt); | ||||||
| 		goto err; | 		goto err; | ||||||
| 	} | 	} | ||||||
| 	skb_scrub_packet(skb, xnet); |  | ||||||
| 	skb_dst_set(skb, &rt->dst); | 	skb_dst_set(skb, &rt->dst); | ||||||
| 	err = ip_local_out(net, skb->sk, skb); | 	err = ip_local_out(net, skb->sk, skb); | ||||||
| 	if (unlikely(net_xmit_eval(err))) | 	if (unlikely(net_xmit_eval(err))) | ||||||
|  | @ -385,7 +396,7 @@ static int ipvlan_process_v4_outbound(struct sk_buff *skb, bool xnet) | ||||||
| 	return ret; | 	return ret; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int ipvlan_process_v6_outbound(struct sk_buff *skb, bool xnet) | static int ipvlan_process_v6_outbound(struct sk_buff *skb) | ||||||
| { | { | ||||||
| 	const struct ipv6hdr *ip6h = ipv6_hdr(skb); | 	const struct ipv6hdr *ip6h = ipv6_hdr(skb); | ||||||
| 	struct net_device *dev = skb->dev; | 	struct net_device *dev = skb->dev; | ||||||
|  | @ -408,7 +419,6 @@ static int ipvlan_process_v6_outbound(struct sk_buff *skb, bool xnet) | ||||||
| 		dst_release(dst); | 		dst_release(dst); | ||||||
| 		goto err; | 		goto err; | ||||||
| 	} | 	} | ||||||
| 	skb_scrub_packet(skb, xnet); |  | ||||||
| 	skb_dst_set(skb, dst); | 	skb_dst_set(skb, dst); | ||||||
| 	err = ip6_local_out(net, skb->sk, skb); | 	err = ip6_local_out(net, skb->sk, skb); | ||||||
| 	if (unlikely(net_xmit_eval(err))) | 	if (unlikely(net_xmit_eval(err))) | ||||||
|  | @ -423,7 +433,7 @@ static int ipvlan_process_v6_outbound(struct sk_buff *skb, bool xnet) | ||||||
| 	return ret; | 	return ret; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int ipvlan_process_outbound(struct sk_buff *skb, bool xnet) | static int ipvlan_process_outbound(struct sk_buff *skb) | ||||||
| { | { | ||||||
| 	struct ethhdr *ethh = eth_hdr(skb); | 	struct ethhdr *ethh = eth_hdr(skb); | ||||||
| 	int ret = NET_XMIT_DROP; | 	int ret = NET_XMIT_DROP; | ||||||
|  | @ -447,9 +457,9 @@ static int ipvlan_process_outbound(struct sk_buff *skb, bool xnet) | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	if (skb->protocol == htons(ETH_P_IPV6)) | 	if (skb->protocol == htons(ETH_P_IPV6)) | ||||||
| 		ret = ipvlan_process_v6_outbound(skb, xnet); | 		ret = ipvlan_process_v6_outbound(skb); | ||||||
| 	else if (skb->protocol == htons(ETH_P_IP)) | 	else if (skb->protocol == htons(ETH_P_IP)) | ||||||
| 		ret = ipvlan_process_v4_outbound(skb, xnet); | 		ret = ipvlan_process_v4_outbound(skb); | ||||||
| 	else { | 	else { | ||||||
| 		pr_warn_ratelimited("Dropped outbound packet type=%x\n", | 		pr_warn_ratelimited("Dropped outbound packet type=%x\n", | ||||||
| 				    ntohs(skb->protocol)); | 				    ntohs(skb->protocol)); | ||||||
|  | @ -485,7 +495,6 @@ static int ipvlan_xmit_mode_l3(struct sk_buff *skb, struct net_device *dev) | ||||||
| 	void *lyr3h; | 	void *lyr3h; | ||||||
| 	struct ipvl_addr *addr; | 	struct ipvl_addr *addr; | ||||||
| 	int addr_type; | 	int addr_type; | ||||||
| 	bool xnet; |  | ||||||
| 
 | 
 | ||||||
| 	lyr3h = ipvlan_get_L3_hdr(skb, &addr_type); | 	lyr3h = ipvlan_get_L3_hdr(skb, &addr_type); | ||||||
| 	if (!lyr3h) | 	if (!lyr3h) | ||||||
|  | @ -496,9 +505,8 @@ static int ipvlan_xmit_mode_l3(struct sk_buff *skb, struct net_device *dev) | ||||||
| 		return ipvlan_rcv_frame(addr, &skb, true); | 		return ipvlan_rcv_frame(addr, &skb, true); | ||||||
| 
 | 
 | ||||||
| out: | out: | ||||||
| 	xnet = !net_eq(dev_net(skb->dev), dev_net(ipvlan->phy_dev)); | 	ipvlan_skb_crossing_ns(skb, ipvlan->phy_dev); | ||||||
| 	skb->dev = ipvlan->phy_dev; | 	return ipvlan_process_outbound(skb); | ||||||
| 	return ipvlan_process_outbound(skb, xnet); |  | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int ipvlan_xmit_mode_l2(struct sk_buff *skb, struct net_device *dev) | static int ipvlan_xmit_mode_l2(struct sk_buff *skb, struct net_device *dev) | ||||||
|  | @ -528,11 +536,12 @@ static int ipvlan_xmit_mode_l2(struct sk_buff *skb, struct net_device *dev) | ||||||
| 		return dev_forward_skb(ipvlan->phy_dev, skb); | 		return dev_forward_skb(ipvlan->phy_dev, skb); | ||||||
| 
 | 
 | ||||||
| 	} else if (is_multicast_ether_addr(eth->h_dest)) { | 	} else if (is_multicast_ether_addr(eth->h_dest)) { | ||||||
|  | 		ipvlan_skb_crossing_ns(skb, NULL); | ||||||
| 		ipvlan_multicast_enqueue(ipvlan->port, skb); | 		ipvlan_multicast_enqueue(ipvlan->port, skb); | ||||||
| 		return NET_XMIT_SUCCESS; | 		return NET_XMIT_SUCCESS; | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	skb->dev = ipvlan->phy_dev; | 	ipvlan_skb_crossing_ns(skb, ipvlan->phy_dev); | ||||||
| 	return dev_queue_xmit(skb); | 	return dev_queue_xmit(skb); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | @ -622,8 +631,10 @@ static rx_handler_result_t ipvlan_handle_mode_l2(struct sk_buff **pskb, | ||||||
| 			 * when work-queue processes this frame. This is | 			 * when work-queue processes this frame. This is | ||||||
| 			 * achieved by returning RX_HANDLER_PASS. | 			 * achieved by returning RX_HANDLER_PASS. | ||||||
| 			 */ | 			 */ | ||||||
| 			if (nskb) | 			if (nskb) { | ||||||
|  | 				ipvlan_skb_crossing_ns(nskb, NULL); | ||||||
| 				ipvlan_multicast_enqueue(port, nskb); | 				ipvlan_multicast_enqueue(port, nskb); | ||||||
|  | 			} | ||||||
| 		} | 		} | ||||||
| 	} else { | 	} else { | ||||||
| 		struct ipvl_addr *addr; | 		struct ipvl_addr *addr; | ||||||
|  |  | ||||||
		Loading…
	
		Reference in a new issue
	
	 Mahesh Bandewar
						Mahesh Bandewar