forked from mirrors/linux
		
	openvswitch: Remove vport stats.
Since all vport types are now backed by netdev, we can directly use netdev stats. Following patch removes redundant stat from vport. Signed-off-by: Pravin B Shelar <pshelar@nicira.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
		
							parent
							
								
									3eedb41fb4
								
							
						
					
					
						commit
						8c876639c9
					
				
					 5 changed files with 56 additions and 170 deletions
				
			
		|  | @ -43,35 +43,26 @@ static struct internal_dev *internal_dev_priv(struct net_device *netdev) | ||||||
| 	return netdev_priv(netdev); | 	return netdev_priv(netdev); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| /* This function is only called by the kernel network layer.*/ |  | ||||||
| static struct rtnl_link_stats64 *internal_dev_get_stats(struct net_device *netdev, |  | ||||||
| 							struct rtnl_link_stats64 *stats) |  | ||||||
| { |  | ||||||
| 	struct vport *vport = ovs_internal_dev_get_vport(netdev); |  | ||||||
| 	struct ovs_vport_stats vport_stats; |  | ||||||
| 
 |  | ||||||
| 	ovs_vport_get_stats(vport, &vport_stats); |  | ||||||
| 
 |  | ||||||
| 	/* The tx and rx stats need to be swapped because the
 |  | ||||||
| 	 * switch and host OS have opposite perspectives. */ |  | ||||||
| 	stats->rx_packets	= vport_stats.tx_packets; |  | ||||||
| 	stats->tx_packets	= vport_stats.rx_packets; |  | ||||||
| 	stats->rx_bytes		= vport_stats.tx_bytes; |  | ||||||
| 	stats->tx_bytes		= vport_stats.rx_bytes; |  | ||||||
| 	stats->rx_errors	= vport_stats.tx_errors; |  | ||||||
| 	stats->tx_errors	= vport_stats.rx_errors; |  | ||||||
| 	stats->rx_dropped	= vport_stats.tx_dropped; |  | ||||||
| 	stats->tx_dropped	= vport_stats.rx_dropped; |  | ||||||
| 
 |  | ||||||
| 	return stats; |  | ||||||
| } |  | ||||||
| 
 |  | ||||||
| /* Called with rcu_read_lock_bh. */ | /* Called with rcu_read_lock_bh. */ | ||||||
| static int internal_dev_xmit(struct sk_buff *skb, struct net_device *netdev) | static int internal_dev_xmit(struct sk_buff *skb, struct net_device *netdev) | ||||||
| { | { | ||||||
|  | 	int len, err; | ||||||
|  | 
 | ||||||
|  | 	len = skb->len; | ||||||
| 	rcu_read_lock(); | 	rcu_read_lock(); | ||||||
| 	ovs_vport_receive(internal_dev_priv(netdev)->vport, skb, NULL); | 	err = ovs_vport_receive(internal_dev_priv(netdev)->vport, skb, NULL); | ||||||
| 	rcu_read_unlock(); | 	rcu_read_unlock(); | ||||||
|  | 
 | ||||||
|  | 	if (likely(!err)) { | ||||||
|  | 		struct pcpu_sw_netstats *tstats = this_cpu_ptr(netdev->tstats); | ||||||
|  | 
 | ||||||
|  | 		u64_stats_update_begin(&tstats->syncp); | ||||||
|  | 		tstats->tx_bytes += len; | ||||||
|  | 		tstats->tx_packets++; | ||||||
|  | 		u64_stats_update_end(&tstats->syncp); | ||||||
|  | 	} else { | ||||||
|  | 		netdev->stats.tx_errors++; | ||||||
|  | 	} | ||||||
| 	return 0; | 	return 0; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | @ -121,7 +112,6 @@ static const struct net_device_ops internal_dev_netdev_ops = { | ||||||
| 	.ndo_start_xmit = internal_dev_xmit, | 	.ndo_start_xmit = internal_dev_xmit, | ||||||
| 	.ndo_set_mac_address = eth_mac_addr, | 	.ndo_set_mac_address = eth_mac_addr, | ||||||
| 	.ndo_change_mtu = internal_dev_change_mtu, | 	.ndo_change_mtu = internal_dev_change_mtu, | ||||||
| 	.ndo_get_stats64 = internal_dev_get_stats, |  | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| static struct rtnl_link_ops internal_dev_link_ops __read_mostly = { | static struct rtnl_link_ops internal_dev_link_ops __read_mostly = { | ||||||
|  | @ -212,18 +202,17 @@ static void internal_dev_destroy(struct vport *vport) | ||||||
| 	rtnl_unlock(); | 	rtnl_unlock(); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static int internal_dev_recv(struct vport *vport, struct sk_buff *skb) | static void internal_dev_recv(struct vport *vport, struct sk_buff *skb) | ||||||
| { | { | ||||||
| 	struct net_device *netdev = vport->dev; | 	struct net_device *netdev = vport->dev; | ||||||
| 	int len; | 	struct pcpu_sw_netstats *stats; | ||||||
| 
 | 
 | ||||||
| 	if (unlikely(!(netdev->flags & IFF_UP))) { | 	if (unlikely(!(netdev->flags & IFF_UP))) { | ||||||
| 		kfree_skb(skb); | 		kfree_skb(skb); | ||||||
| 		return 0; | 		netdev->stats.rx_dropped++; | ||||||
|  | 		return; | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	len = skb->len; |  | ||||||
| 
 |  | ||||||
| 	skb_dst_drop(skb); | 	skb_dst_drop(skb); | ||||||
| 	nf_reset(skb); | 	nf_reset(skb); | ||||||
| 	secpath_reset(skb); | 	secpath_reset(skb); | ||||||
|  | @ -233,9 +222,13 @@ static int internal_dev_recv(struct vport *vport, struct sk_buff *skb) | ||||||
| 	skb->protocol = eth_type_trans(skb, netdev); | 	skb->protocol = eth_type_trans(skb, netdev); | ||||||
| 	skb_postpull_rcsum(skb, eth_hdr(skb), ETH_HLEN); | 	skb_postpull_rcsum(skb, eth_hdr(skb), ETH_HLEN); | ||||||
| 
 | 
 | ||||||
| 	netif_rx(skb); | 	stats = this_cpu_ptr(netdev->tstats); | ||||||
|  | 	u64_stats_update_begin(&stats->syncp); | ||||||
|  | 	stats->rx_packets++; | ||||||
|  | 	stats->rx_bytes += skb->len; | ||||||
|  | 	u64_stats_update_end(&stats->syncp); | ||||||
| 
 | 
 | ||||||
| 	return len; | 	netif_rx(skb); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| static struct vport_ops ovs_internal_vport_ops = { | static struct vport_ops ovs_internal_vport_ops = { | ||||||
|  |  | ||||||
|  | @ -39,8 +39,11 @@ | ||||||
| static struct vport_ops ovs_netdev_vport_ops; | static struct vport_ops ovs_netdev_vport_ops; | ||||||
| 
 | 
 | ||||||
| /* Must be called with rcu_read_lock. */ | /* Must be called with rcu_read_lock. */ | ||||||
| static void netdev_port_receive(struct vport *vport, struct sk_buff *skb) | static void netdev_port_receive(struct sk_buff *skb) | ||||||
| { | { | ||||||
|  | 	struct vport *vport; | ||||||
|  | 
 | ||||||
|  | 	vport = ovs_netdev_get_vport(skb->dev); | ||||||
| 	if (unlikely(!vport)) | 	if (unlikely(!vport)) | ||||||
| 		goto error; | 		goto error; | ||||||
| 
 | 
 | ||||||
|  | @ -56,10 +59,8 @@ static void netdev_port_receive(struct vport *vport, struct sk_buff *skb) | ||||||
| 
 | 
 | ||||||
| 	skb_push(skb, ETH_HLEN); | 	skb_push(skb, ETH_HLEN); | ||||||
| 	ovs_skb_postpush_rcsum(skb, skb->data, ETH_HLEN); | 	ovs_skb_postpush_rcsum(skb, skb->data, ETH_HLEN); | ||||||
| 
 |  | ||||||
| 	ovs_vport_receive(vport, skb, skb_tunnel_info(skb)); | 	ovs_vport_receive(vport, skb, skb_tunnel_info(skb)); | ||||||
| 	return; | 	return; | ||||||
| 
 |  | ||||||
| error: | error: | ||||||
| 	kfree_skb(skb); | 	kfree_skb(skb); | ||||||
| } | } | ||||||
|  | @ -68,15 +69,11 @@ static void netdev_port_receive(struct vport *vport, struct sk_buff *skb) | ||||||
| static rx_handler_result_t netdev_frame_hook(struct sk_buff **pskb) | static rx_handler_result_t netdev_frame_hook(struct sk_buff **pskb) | ||||||
| { | { | ||||||
| 	struct sk_buff *skb = *pskb; | 	struct sk_buff *skb = *pskb; | ||||||
| 	struct vport *vport; |  | ||||||
| 
 | 
 | ||||||
| 	if (unlikely(skb->pkt_type == PACKET_LOOPBACK)) | 	if (unlikely(skb->pkt_type == PACKET_LOOPBACK)) | ||||||
| 		return RX_HANDLER_PASS; | 		return RX_HANDLER_PASS; | ||||||
| 
 | 
 | ||||||
| 	vport = ovs_netdev_get_vport(skb->dev); | 	netdev_port_receive(skb); | ||||||
| 
 |  | ||||||
| 	netdev_port_receive(vport, skb); |  | ||||||
| 
 |  | ||||||
| 	return RX_HANDLER_CONSUMED; | 	return RX_HANDLER_CONSUMED; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  | @ -203,27 +200,24 @@ static unsigned int packet_length(const struct sk_buff *skb) | ||||||
| 	return length; | 	return length; | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| int ovs_netdev_send(struct vport *vport, struct sk_buff *skb) | void ovs_netdev_send(struct vport *vport, struct sk_buff *skb) | ||||||
| { | { | ||||||
| 	int mtu = vport->dev->mtu; | 	int mtu = vport->dev->mtu; | ||||||
| 	int len; |  | ||||||
| 
 | 
 | ||||||
| 	if (unlikely(packet_length(skb) > mtu && !skb_is_gso(skb))) { | 	if (unlikely(packet_length(skb) > mtu && !skb_is_gso(skb))) { | ||||||
| 		net_warn_ratelimited("%s: dropped over-mtu packet: %d > %d\n", | 		net_warn_ratelimited("%s: dropped over-mtu packet: %d > %d\n", | ||||||
| 				     vport->dev->name, | 				     vport->dev->name, | ||||||
| 				     packet_length(skb), mtu); | 				     packet_length(skb), mtu); | ||||||
|  | 		vport->dev->stats.tx_errors++; | ||||||
| 		goto drop; | 		goto drop; | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	skb->dev = vport->dev; | 	skb->dev = vport->dev; | ||||||
| 	len = skb->len; |  | ||||||
| 	dev_queue_xmit(skb); | 	dev_queue_xmit(skb); | ||||||
| 
 | 	return; | ||||||
| 	return len; |  | ||||||
| 
 | 
 | ||||||
| drop: | drop: | ||||||
| 	kfree_skb(skb); | 	kfree_skb(skb); | ||||||
| 	return 0; |  | ||||||
| } | } | ||||||
| EXPORT_SYMBOL_GPL(ovs_netdev_send); | EXPORT_SYMBOL_GPL(ovs_netdev_send); | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -27,7 +27,7 @@ | ||||||
| struct vport *ovs_netdev_get_vport(struct net_device *dev); | struct vport *ovs_netdev_get_vport(struct net_device *dev); | ||||||
| 
 | 
 | ||||||
| struct vport *ovs_netdev_link(struct vport *vport, const char *name); | struct vport *ovs_netdev_link(struct vport *vport, const char *name); | ||||||
| int ovs_netdev_send(struct vport *vport, struct sk_buff *skb); | void ovs_netdev_send(struct vport *vport, struct sk_buff *skb); | ||||||
| void ovs_netdev_detach_dev(struct vport *); | void ovs_netdev_detach_dev(struct vport *); | ||||||
| 
 | 
 | ||||||
| int __init ovs_netdev_init(void); | int __init ovs_netdev_init(void); | ||||||
|  |  | ||||||
|  | @ -34,9 +34,6 @@ | ||||||
| #include "vport.h" | #include "vport.h" | ||||||
| #include "vport-internal_dev.h" | #include "vport-internal_dev.h" | ||||||
| 
 | 
 | ||||||
| static void ovs_vport_record_error(struct vport *, |  | ||||||
| 				   enum vport_err_type err_type); |  | ||||||
| 
 |  | ||||||
| static LIST_HEAD(vport_ops_list); | static LIST_HEAD(vport_ops_list); | ||||||
| 
 | 
 | ||||||
| /* Protected by RCU read lock for reading, ovs_mutex for writing. */ | /* Protected by RCU read lock for reading, ovs_mutex for writing. */ | ||||||
|  | @ -157,12 +154,6 @@ struct vport *ovs_vport_alloc(int priv_size, const struct vport_ops *ops, | ||||||
| 		return ERR_PTR(-EINVAL); | 		return ERR_PTR(-EINVAL); | ||||||
| 	} | 	} | ||||||
| 
 | 
 | ||||||
| 	vport->percpu_stats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats); |  | ||||||
| 	if (!vport->percpu_stats) { |  | ||||||
| 		kfree(vport); |  | ||||||
| 		return ERR_PTR(-ENOMEM); |  | ||||||
| 	} |  | ||||||
| 
 |  | ||||||
| 	return vport; | 	return vport; | ||||||
| } | } | ||||||
| EXPORT_SYMBOL_GPL(ovs_vport_alloc); | EXPORT_SYMBOL_GPL(ovs_vport_alloc); | ||||||
|  | @ -183,7 +174,6 @@ void ovs_vport_free(struct vport *vport) | ||||||
| 	 * it is safe to use raw dereference. | 	 * it is safe to use raw dereference. | ||||||
| 	 */ | 	 */ | ||||||
| 	kfree(rcu_dereference_raw(vport->upcall_portids)); | 	kfree(rcu_dereference_raw(vport->upcall_portids)); | ||||||
| 	free_percpu(vport->percpu_stats); |  | ||||||
| 	kfree(vport); | 	kfree(vport); | ||||||
| } | } | ||||||
| EXPORT_SYMBOL_GPL(ovs_vport_free); | EXPORT_SYMBOL_GPL(ovs_vport_free); | ||||||
|  | @ -290,30 +280,24 @@ void ovs_vport_del(struct vport *vport) | ||||||
|  */ |  */ | ||||||
| void ovs_vport_get_stats(struct vport *vport, struct ovs_vport_stats *stats) | void ovs_vport_get_stats(struct vport *vport, struct ovs_vport_stats *stats) | ||||||
| { | { | ||||||
|  | 	struct net_device *dev = vport->dev; | ||||||
| 	int i; | 	int i; | ||||||
| 
 | 
 | ||||||
| 	memset(stats, 0, sizeof(*stats)); | 	memset(stats, 0, sizeof(*stats)); | ||||||
|  | 	stats->rx_errors  = dev->stats.rx_errors; | ||||||
|  | 	stats->tx_errors  = dev->stats.tx_errors; | ||||||
|  | 	stats->tx_dropped = dev->stats.tx_dropped; | ||||||
|  | 	stats->rx_dropped = dev->stats.rx_dropped; | ||||||
| 
 | 
 | ||||||
| 	/* We potentially have 2 sources of stats that need to be combined:
 | 	stats->rx_dropped += atomic_long_read(&dev->rx_dropped); | ||||||
| 	 * those we have collected (split into err_stats and percpu_stats) from | 	stats->tx_dropped += atomic_long_read(&dev->tx_dropped); | ||||||
| 	 * set_stats() and device error stats from netdev->get_stats() (for |  | ||||||
| 	 * errors that happen  downstream and therefore aren't reported through |  | ||||||
| 	 * our vport_record_error() function). |  | ||||||
| 	 * Stats from first source are reported by ovs (OVS_VPORT_ATTR_STATS). |  | ||||||
| 	 * netdev-stats can be directly read over netlink-ioctl. |  | ||||||
| 	 */ |  | ||||||
| 
 |  | ||||||
| 	stats->rx_errors  = atomic_long_read(&vport->err_stats.rx_errors); |  | ||||||
| 	stats->tx_errors  = atomic_long_read(&vport->err_stats.tx_errors); |  | ||||||
| 	stats->tx_dropped = atomic_long_read(&vport->err_stats.tx_dropped); |  | ||||||
| 	stats->rx_dropped = atomic_long_read(&vport->err_stats.rx_dropped); |  | ||||||
| 
 | 
 | ||||||
| 	for_each_possible_cpu(i) { | 	for_each_possible_cpu(i) { | ||||||
| 		const struct pcpu_sw_netstats *percpu_stats; | 		const struct pcpu_sw_netstats *percpu_stats; | ||||||
| 		struct pcpu_sw_netstats local_stats; | 		struct pcpu_sw_netstats local_stats; | ||||||
| 		unsigned int start; | 		unsigned int start; | ||||||
| 
 | 
 | ||||||
| 		percpu_stats = per_cpu_ptr(vport->percpu_stats, i); | 		percpu_stats = per_cpu_ptr(dev->tstats, i); | ||||||
| 
 | 
 | ||||||
| 		do { | 		do { | ||||||
| 			start = u64_stats_fetch_begin_irq(&percpu_stats->syncp); | 			start = u64_stats_fetch_begin_irq(&percpu_stats->syncp); | ||||||
|  | @ -468,94 +452,25 @@ u32 ovs_vport_find_upcall_portid(const struct vport *vport, struct sk_buff *skb) | ||||||
|  * Must be called with rcu_read_lock.  The packet cannot be shared and |  * Must be called with rcu_read_lock.  The packet cannot be shared and | ||||||
|  * skb->data should point to the Ethernet header. |  * skb->data should point to the Ethernet header. | ||||||
|  */ |  */ | ||||||
| void ovs_vport_receive(struct vport *vport, struct sk_buff *skb, | int ovs_vport_receive(struct vport *vport, struct sk_buff *skb, | ||||||
| 		       const struct ip_tunnel_info *tun_info) | 		      const struct ip_tunnel_info *tun_info) | ||||||
| { | { | ||||||
| 	struct pcpu_sw_netstats *stats; |  | ||||||
| 	struct sw_flow_key key; | 	struct sw_flow_key key; | ||||||
| 	int error; | 	int error; | ||||||
| 
 | 
 | ||||||
| 	stats = this_cpu_ptr(vport->percpu_stats); |  | ||||||
| 	u64_stats_update_begin(&stats->syncp); |  | ||||||
| 	stats->rx_packets++; |  | ||||||
| 	stats->rx_bytes += skb->len + |  | ||||||
| 			   (skb_vlan_tag_present(skb) ? VLAN_HLEN : 0); |  | ||||||
| 	u64_stats_update_end(&stats->syncp); |  | ||||||
| 
 |  | ||||||
| 	OVS_CB(skb)->input_vport = vport; | 	OVS_CB(skb)->input_vport = vport; | ||||||
| 	OVS_CB(skb)->mru = 0; | 	OVS_CB(skb)->mru = 0; | ||||||
| 	/* Extract flow from 'skb' into 'key'. */ | 	/* Extract flow from 'skb' into 'key'. */ | ||||||
| 	error = ovs_flow_key_extract(tun_info, skb, &key); | 	error = ovs_flow_key_extract(tun_info, skb, &key); | ||||||
| 	if (unlikely(error)) { | 	if (unlikely(error)) { | ||||||
| 		kfree_skb(skb); | 		kfree_skb(skb); | ||||||
| 		return; | 		return error; | ||||||
| 	} | 	} | ||||||
| 	ovs_dp_process_packet(skb, &key); | 	ovs_dp_process_packet(skb, &key); | ||||||
|  | 	return 0; | ||||||
| } | } | ||||||
| EXPORT_SYMBOL_GPL(ovs_vport_receive); | EXPORT_SYMBOL_GPL(ovs_vport_receive); | ||||||
| 
 | 
 | ||||||
| /**
 |  | ||||||
|  *	ovs_vport_send - send a packet on a device |  | ||||||
|  * |  | ||||||
|  * @vport: vport on which to send the packet |  | ||||||
|  * @skb: skb to send |  | ||||||
|  * |  | ||||||
|  * Sends the given packet and returns the length of data sent.  Either ovs |  | ||||||
|  * lock or rcu_read_lock must be held. |  | ||||||
|  */ |  | ||||||
| int ovs_vport_send(struct vport *vport, struct sk_buff *skb) |  | ||||||
| { |  | ||||||
| 	int sent = vport->ops->send(vport, skb); |  | ||||||
| 
 |  | ||||||
| 	if (likely(sent > 0)) { |  | ||||||
| 		struct pcpu_sw_netstats *stats; |  | ||||||
| 
 |  | ||||||
| 		stats = this_cpu_ptr(vport->percpu_stats); |  | ||||||
| 
 |  | ||||||
| 		u64_stats_update_begin(&stats->syncp); |  | ||||||
| 		stats->tx_packets++; |  | ||||||
| 		stats->tx_bytes += sent; |  | ||||||
| 		u64_stats_update_end(&stats->syncp); |  | ||||||
| 	} else if (sent < 0) { |  | ||||||
| 		ovs_vport_record_error(vport, VPORT_E_TX_ERROR); |  | ||||||
| 	} else { |  | ||||||
| 		ovs_vport_record_error(vport, VPORT_E_TX_DROPPED); |  | ||||||
| 	} |  | ||||||
| 	return sent; |  | ||||||
| } |  | ||||||
| 
 |  | ||||||
| /**
 |  | ||||||
|  *	ovs_vport_record_error - indicate device error to generic stats layer |  | ||||||
|  * |  | ||||||
|  * @vport: vport that encountered the error |  | ||||||
|  * @err_type: one of enum vport_err_type types to indicate the error type |  | ||||||
|  * |  | ||||||
|  * If using the vport generic stats layer indicate that an error of the given |  | ||||||
|  * type has occurred. |  | ||||||
|  */ |  | ||||||
| static void ovs_vport_record_error(struct vport *vport, |  | ||||||
| 				   enum vport_err_type err_type) |  | ||||||
| { |  | ||||||
| 	switch (err_type) { |  | ||||||
| 	case VPORT_E_RX_DROPPED: |  | ||||||
| 		atomic_long_inc(&vport->err_stats.rx_dropped); |  | ||||||
| 		break; |  | ||||||
| 
 |  | ||||||
| 	case VPORT_E_RX_ERROR: |  | ||||||
| 		atomic_long_inc(&vport->err_stats.rx_errors); |  | ||||||
| 		break; |  | ||||||
| 
 |  | ||||||
| 	case VPORT_E_TX_DROPPED: |  | ||||||
| 		atomic_long_inc(&vport->err_stats.tx_dropped); |  | ||||||
| 		break; |  | ||||||
| 
 |  | ||||||
| 	case VPORT_E_TX_ERROR: |  | ||||||
| 		atomic_long_inc(&vport->err_stats.tx_errors); |  | ||||||
| 		break; |  | ||||||
| 	} |  | ||||||
| 
 |  | ||||||
| } |  | ||||||
| 
 |  | ||||||
| static void free_vport_rcu(struct rcu_head *rcu) | static void free_vport_rcu(struct rcu_head *rcu) | ||||||
| { | { | ||||||
| 	struct vport *vport = container_of(rcu, struct vport, rcu); | 	struct vport *vport = container_of(rcu, struct vport, rcu); | ||||||
|  |  | ||||||
|  | @ -57,8 +57,6 @@ int ovs_vport_set_upcall_portids(struct vport *, const struct nlattr *pids); | ||||||
| int ovs_vport_get_upcall_portids(const struct vport *, struct sk_buff *); | int ovs_vport_get_upcall_portids(const struct vport *, struct sk_buff *); | ||||||
| u32 ovs_vport_find_upcall_portid(const struct vport *, struct sk_buff *); | u32 ovs_vport_find_upcall_portid(const struct vport *, struct sk_buff *); | ||||||
| 
 | 
 | ||||||
| int ovs_vport_send(struct vport *, struct sk_buff *); |  | ||||||
| 
 |  | ||||||
| int ovs_tunnel_get_egress_info(struct ip_tunnel_info *egress_tun_info, | int ovs_tunnel_get_egress_info(struct ip_tunnel_info *egress_tun_info, | ||||||
| 			       struct net *net, | 			       struct net *net, | ||||||
| 			       struct sk_buff *, | 			       struct sk_buff *, | ||||||
|  | @ -68,14 +66,6 @@ int ovs_tunnel_get_egress_info(struct ip_tunnel_info *egress_tun_info, | ||||||
| int ovs_vport_get_egress_tun_info(struct vport *vport, struct sk_buff *skb, | int ovs_vport_get_egress_tun_info(struct vport *vport, struct sk_buff *skb, | ||||||
| 				  struct ip_tunnel_info *info); | 				  struct ip_tunnel_info *info); | ||||||
| 
 | 
 | ||||||
| /* The following definitions are for implementers of vport devices: */ |  | ||||||
| 
 |  | ||||||
| struct vport_err_stats { |  | ||||||
| 	atomic_long_t rx_dropped; |  | ||||||
| 	atomic_long_t rx_errors; |  | ||||||
| 	atomic_long_t tx_dropped; |  | ||||||
| 	atomic_long_t tx_errors; |  | ||||||
| }; |  | ||||||
| /**
 | /**
 | ||||||
|  * struct vport_portids - array of netlink portids of a vport. |  * struct vport_portids - array of netlink portids of a vport. | ||||||
|  *                        must be protected by rcu. |  *                        must be protected by rcu. | ||||||
|  | @ -101,8 +91,6 @@ struct vport_portids { | ||||||
|  * @hash_node: Element in @dev_table hash table in vport.c. |  * @hash_node: Element in @dev_table hash table in vport.c. | ||||||
|  * @dp_hash_node: Element in @datapath->ports hash table in datapath.c. |  * @dp_hash_node: Element in @datapath->ports hash table in datapath.c. | ||||||
|  * @ops: Class structure. |  * @ops: Class structure. | ||||||
|  * @percpu_stats: Points to per-CPU statistics used and maintained by vport |  | ||||||
|  * @err_stats: Points to error statistics used and maintained by vport |  | ||||||
|  * @detach_list: list used for detaching vport in net-exit call. |  * @detach_list: list used for detaching vport in net-exit call. | ||||||
|  */ |  */ | ||||||
| struct vport { | struct vport { | ||||||
|  | @ -115,9 +103,6 @@ struct vport { | ||||||
| 	struct hlist_node dp_hash_node; | 	struct hlist_node dp_hash_node; | ||||||
| 	const struct vport_ops *ops; | 	const struct vport_ops *ops; | ||||||
| 
 | 
 | ||||||
| 	struct pcpu_sw_netstats __percpu *percpu_stats; |  | ||||||
| 
 |  | ||||||
| 	struct vport_err_stats err_stats; |  | ||||||
| 	struct list_head detach_list; | 	struct list_head detach_list; | ||||||
| 	struct rcu_head rcu; | 	struct rcu_head rcu; | ||||||
| }; | }; | ||||||
|  | @ -156,7 +141,7 @@ struct vport_parms { | ||||||
|  * @get_options: Appends vport-specific attributes for the configuration of an |  * @get_options: Appends vport-specific attributes for the configuration of an | ||||||
|  * existing vport to a &struct sk_buff.  May be %NULL for a vport that does not |  * existing vport to a &struct sk_buff.  May be %NULL for a vport that does not | ||||||
|  * have any configuration. |  * have any configuration. | ||||||
|  * @send: Send a packet on the device.  Returns the length of the packet sent, |  * @send: Send a packet on the device. | ||||||
|  * zero for dropped packets or negative for error. |  * zero for dropped packets or negative for error. | ||||||
|  * @get_egress_tun_info: Get the egress tunnel 5-tuple and other info for |  * @get_egress_tun_info: Get the egress tunnel 5-tuple and other info for | ||||||
|  * a packet. |  * a packet. | ||||||
|  | @ -171,7 +156,7 @@ struct vport_ops { | ||||||
| 	int (*set_options)(struct vport *, struct nlattr *); | 	int (*set_options)(struct vport *, struct nlattr *); | ||||||
| 	int (*get_options)(const struct vport *, struct sk_buff *); | 	int (*get_options)(const struct vport *, struct sk_buff *); | ||||||
| 
 | 
 | ||||||
| 	int (*send)(struct vport *, struct sk_buff *); | 	void (*send)(struct vport *, struct sk_buff *); | ||||||
| 	int (*get_egress_tun_info)(struct vport *, struct sk_buff *, | 	int (*get_egress_tun_info)(struct vport *, struct sk_buff *, | ||||||
| 				   struct ip_tunnel_info *); | 				   struct ip_tunnel_info *); | ||||||
| 
 | 
 | ||||||
|  | @ -179,13 +164,6 @@ struct vport_ops { | ||||||
| 	struct list_head list; | 	struct list_head list; | ||||||
| }; | }; | ||||||
| 
 | 
 | ||||||
| enum vport_err_type { |  | ||||||
| 	VPORT_E_RX_DROPPED, |  | ||||||
| 	VPORT_E_RX_ERROR, |  | ||||||
| 	VPORT_E_TX_DROPPED, |  | ||||||
| 	VPORT_E_TX_ERROR, |  | ||||||
| }; |  | ||||||
| 
 |  | ||||||
| struct vport *ovs_vport_alloc(int priv_size, const struct vport_ops *, | struct vport *ovs_vport_alloc(int priv_size, const struct vport_ops *, | ||||||
| 			      const struct vport_parms *); | 			      const struct vport_parms *); | ||||||
| void ovs_vport_free(struct vport *); | void ovs_vport_free(struct vport *); | ||||||
|  | @ -222,8 +200,8 @@ static inline struct vport *vport_from_priv(void *priv) | ||||||
| 	return (struct vport *)((u8 *)priv - ALIGN(sizeof(struct vport), VPORT_ALIGN)); | 	return (struct vport *)((u8 *)priv - ALIGN(sizeof(struct vport), VPORT_ALIGN)); | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| void ovs_vport_receive(struct vport *, struct sk_buff *, | int ovs_vport_receive(struct vport *, struct sk_buff *, | ||||||
| 		       const struct ip_tunnel_info *); | 		      const struct ip_tunnel_info *); | ||||||
| 
 | 
 | ||||||
| static inline void ovs_skb_postpush_rcsum(struct sk_buff *skb, | static inline void ovs_skb_postpush_rcsum(struct sk_buff *skb, | ||||||
| 				      const void *start, unsigned int len) | 				      const void *start, unsigned int len) | ||||||
|  | @ -258,4 +236,10 @@ static inline struct rtable *ovs_tunnel_route_lookup(struct net *net, | ||||||
| 	rt = ip_route_output_key(net, fl); | 	rt = ip_route_output_key(net, fl); | ||||||
| 	return rt; | 	return rt; | ||||||
| } | } | ||||||
|  | 
 | ||||||
|  | static inline void ovs_vport_send(struct vport *vport, struct sk_buff *skb) | ||||||
|  | { | ||||||
|  | 	vport->ops->send(vport, skb); | ||||||
|  | } | ||||||
|  | 
 | ||||||
| #endif /* vport.h */ | #endif /* vport.h */ | ||||||
|  |  | ||||||
		Loading…
	
		Reference in a new issue
	
	 Pravin B Shelar
						Pravin B Shelar