mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	Reset the adapter through new function ibmveth_reset, called in WARN_ON situations. Removed conflicting and unneeded forward declaration. Signed-off-by: Dave Marquardt <davemarq@linux.ibm.com> Link: https://patch.msgid.link/20250501194944.283729-3-davemarq@linux.ibm.com Signed-off-by: Paolo Abeni <pabeni@redhat.com>
		
			
				
	
	
		
			217 lines
		
	
	
	
		
			6.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			217 lines
		
	
	
	
		
			6.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/* SPDX-License-Identifier: GPL-2.0-or-later */
 | 
						|
/*
 | 
						|
 * IBM Power Virtual Ethernet Device Driver
 | 
						|
 *
 | 
						|
 * Copyright (C) IBM Corporation, 2003, 2010
 | 
						|
 *
 | 
						|
 * Authors: Dave Larson <larson1@us.ibm.com>
 | 
						|
 *	    Santiago Leon <santil@linux.vnet.ibm.com>
 | 
						|
 *	    Brian King <brking@linux.vnet.ibm.com>
 | 
						|
 *	    Robert Jennings <rcj@linux.vnet.ibm.com>
 | 
						|
 *	    Anton Blanchard <anton@au.ibm.com>
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef _IBMVETH_H
 | 
						|
#define _IBMVETH_H
 | 
						|
 | 
						|
/* constants for H_MULTICAST_CTRL */
 | 
						|
#define IbmVethMcastReceptionModifyBit     0x80000UL
 | 
						|
#define IbmVethMcastReceptionEnableBit     0x20000UL
 | 
						|
#define IbmVethMcastFilterModifyBit        0x40000UL
 | 
						|
#define IbmVethMcastFilterEnableBit        0x10000UL
 | 
						|
 | 
						|
#define IbmVethMcastEnableRecv       (IbmVethMcastReceptionModifyBit | IbmVethMcastReceptionEnableBit)
 | 
						|
#define IbmVethMcastDisableRecv      (IbmVethMcastReceptionModifyBit)
 | 
						|
#define IbmVethMcastEnableFiltering  (IbmVethMcastFilterModifyBit | IbmVethMcastFilterEnableBit)
 | 
						|
#define IbmVethMcastDisableFiltering (IbmVethMcastFilterModifyBit)
 | 
						|
#define IbmVethMcastAddFilter        0x1UL
 | 
						|
#define IbmVethMcastRemoveFilter     0x2UL
 | 
						|
#define IbmVethMcastClearFilterTable 0x3UL
 | 
						|
 | 
						|
#define IBMVETH_ILLAN_LRG_SR_ENABLED	0x0000000000010000UL
 | 
						|
#define IBMVETH_ILLAN_LRG_SND_SUPPORT	0x0000000000008000UL
 | 
						|
#define IBMVETH_ILLAN_PADDED_PKT_CSUM	0x0000000000002000UL
 | 
						|
#define IBMVETH_ILLAN_TRUNK_PRI_MASK	0x0000000000000F00UL
 | 
						|
#define IBMVETH_ILLAN_IPV6_TCP_CSUM		0x0000000000000004UL
 | 
						|
#define IBMVETH_ILLAN_IPV4_TCP_CSUM		0x0000000000000002UL
 | 
						|
#define IBMVETH_ILLAN_ACTIVE_TRUNK		0x0000000000000001UL
 | 
						|
 | 
						|
/* hcall macros */
 | 
						|
#define h_register_logical_lan(ua, buflst, rxq, fltlst, mac) \
 | 
						|
  plpar_hcall_norets(H_REGISTER_LOGICAL_LAN, ua, buflst, rxq, fltlst, mac)
 | 
						|
 | 
						|
#define h_free_logical_lan(ua) \
 | 
						|
  plpar_hcall_norets(H_FREE_LOGICAL_LAN, ua)
 | 
						|
 | 
						|
#define h_add_logical_lan_buffer(ua, buf) \
 | 
						|
  plpar_hcall_norets(H_ADD_LOGICAL_LAN_BUFFER, ua, buf)
 | 
						|
 | 
						|
/* FW allows us to send 6 descriptors but we only use one so mark
 | 
						|
 * the other 5 as unused (0)
 | 
						|
 */
 | 
						|
static inline long h_send_logical_lan(unsigned long unit_address,
 | 
						|
		unsigned long desc, unsigned long corellator_in,
 | 
						|
		unsigned long *corellator_out, unsigned long mss,
 | 
						|
		unsigned long large_send_support)
 | 
						|
{
 | 
						|
	long rc;
 | 
						|
	unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
 | 
						|
 | 
						|
	if (large_send_support)
 | 
						|
		rc = plpar_hcall9(H_SEND_LOGICAL_LAN, retbuf, unit_address,
 | 
						|
				  desc, 0, 0, 0, 0, 0, corellator_in, mss);
 | 
						|
	else
 | 
						|
		rc = plpar_hcall9(H_SEND_LOGICAL_LAN, retbuf, unit_address,
 | 
						|
				  desc, 0, 0, 0, 0, 0, corellator_in);
 | 
						|
 | 
						|
	*corellator_out = retbuf[0];
 | 
						|
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
static inline long h_illan_attributes(unsigned long unit_address,
 | 
						|
				      unsigned long reset_mask, unsigned long set_mask,
 | 
						|
				      unsigned long *ret_attributes)
 | 
						|
{
 | 
						|
	long rc;
 | 
						|
	unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
 | 
						|
 | 
						|
	rc = plpar_hcall(H_ILLAN_ATTRIBUTES, retbuf, unit_address,
 | 
						|
			 reset_mask, set_mask);
 | 
						|
 | 
						|
	*ret_attributes = retbuf[0];
 | 
						|
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
#define h_multicast_ctrl(ua, cmd, mac) \
 | 
						|
  plpar_hcall_norets(H_MULTICAST_CTRL, ua, cmd, mac)
 | 
						|
 | 
						|
#define h_change_logical_lan_mac(ua, mac) \
 | 
						|
  plpar_hcall_norets(H_CHANGE_LOGICAL_LAN_MAC, ua, mac)
 | 
						|
 | 
						|
#define IBMVETH_NUM_BUFF_POOLS 5
 | 
						|
#define IBMVETH_IO_ENTITLEMENT_DEFAULT 4243456 /* MTU of 1500 needs 4.2Mb */
 | 
						|
#define IBMVETH_BUFF_OH 22 /* Overhead: 14 ethernet header + 8 opaque handle */
 | 
						|
#define IBMVETH_MIN_MTU 68
 | 
						|
#define IBMVETH_MAX_POOL_COUNT 4096
 | 
						|
#define IBMVETH_BUFF_LIST_SIZE 4096
 | 
						|
#define IBMVETH_FILT_LIST_SIZE 4096
 | 
						|
#define IBMVETH_MAX_BUF_SIZE (1024 * 128)
 | 
						|
#define IBMVETH_MAX_TX_BUF_SIZE (1024 * 64)
 | 
						|
#define IBMVETH_MAX_QUEUES 16U
 | 
						|
#define IBMVETH_DEFAULT_QUEUES 8U
 | 
						|
 | 
						|
static int pool_size[] = { 512, 1024 * 2, 1024 * 16, 1024 * 32, 1024 * 64 };
 | 
						|
static int pool_count[] = { 256, 512, 256, 256, 256 };
 | 
						|
static int pool_count_cmo[] = { 256, 512, 256, 256, 64 };
 | 
						|
static int pool_active[] = { 1, 1, 0, 0, 1};
 | 
						|
 | 
						|
#define IBM_VETH_INVALID_MAP ((u16)0xffff)
 | 
						|
 | 
						|
struct ibmveth_buff_pool {
 | 
						|
    u32 size;
 | 
						|
    u32 index;
 | 
						|
    u32 buff_size;
 | 
						|
    u32 threshold;
 | 
						|
    atomic_t available;
 | 
						|
    u32 consumer_index;
 | 
						|
    u32 producer_index;
 | 
						|
    u16 *free_map;
 | 
						|
    dma_addr_t *dma_addr;
 | 
						|
    struct sk_buff **skbuff;
 | 
						|
    int active;
 | 
						|
    struct kobject kobj;
 | 
						|
};
 | 
						|
 | 
						|
struct ibmveth_rx_q {
 | 
						|
    u64        index;
 | 
						|
    u64        num_slots;
 | 
						|
    u64        toggle;
 | 
						|
    dma_addr_t queue_dma;
 | 
						|
    u32        queue_len;
 | 
						|
    struct ibmveth_rx_q_entry *queue_addr;
 | 
						|
};
 | 
						|
 | 
						|
struct ibmveth_adapter {
 | 
						|
	struct vio_dev *vdev;
 | 
						|
	struct net_device *netdev;
 | 
						|
	struct napi_struct napi;
 | 
						|
	struct work_struct work;
 | 
						|
	unsigned int mcastFilterSize;
 | 
						|
	void *buffer_list_addr;
 | 
						|
	void *filter_list_addr;
 | 
						|
	void *tx_ltb_ptr[IBMVETH_MAX_QUEUES];
 | 
						|
	unsigned int tx_ltb_size;
 | 
						|
	dma_addr_t tx_ltb_dma[IBMVETH_MAX_QUEUES];
 | 
						|
	dma_addr_t buffer_list_dma;
 | 
						|
	dma_addr_t filter_list_dma;
 | 
						|
	struct ibmveth_buff_pool rx_buff_pool[IBMVETH_NUM_BUFF_POOLS];
 | 
						|
	struct ibmveth_rx_q rx_queue;
 | 
						|
	int rx_csum;
 | 
						|
	int large_send;
 | 
						|
	bool is_active_trunk;
 | 
						|
 | 
						|
	u64 fw_ipv6_csum_support;
 | 
						|
	u64 fw_ipv4_csum_support;
 | 
						|
	u64 fw_large_send_support;
 | 
						|
	/* adapter specific stats */
 | 
						|
	u64 replenish_task_cycles;
 | 
						|
	u64 replenish_no_mem;
 | 
						|
	u64 replenish_add_buff_failure;
 | 
						|
	u64 replenish_add_buff_success;
 | 
						|
	u64 rx_invalid_buffer;
 | 
						|
	u64 rx_no_buffer;
 | 
						|
	u64 tx_map_failed;
 | 
						|
	u64 tx_send_failed;
 | 
						|
	u64 tx_large_packets;
 | 
						|
	u64 rx_large_packets;
 | 
						|
	/* Ethtool settings */
 | 
						|
	u8 duplex;
 | 
						|
	u32 speed;
 | 
						|
};
 | 
						|
 | 
						|
/*
 | 
						|
 * We pass struct ibmveth_buf_desc_fields to the hypervisor in registers,
 | 
						|
 * so we don't need to byteswap the two elements. However since we use
 | 
						|
 * a union (ibmveth_buf_desc) to convert from the struct to a u64 we
 | 
						|
 * do end up with endian specific ordering of the elements and that
 | 
						|
 * needs correcting.
 | 
						|
 */
 | 
						|
struct ibmveth_buf_desc_fields {
 | 
						|
#ifdef __BIG_ENDIAN
 | 
						|
	u32 flags_len;
 | 
						|
	u32 address;
 | 
						|
#else
 | 
						|
	u32 address;
 | 
						|
	u32 flags_len;
 | 
						|
#endif
 | 
						|
#define IBMVETH_BUF_VALID	0x80000000
 | 
						|
#define IBMVETH_BUF_TOGGLE	0x40000000
 | 
						|
#define IBMVETH_BUF_LRG_SND     0x04000000
 | 
						|
#define IBMVETH_BUF_NO_CSUM	0x02000000
 | 
						|
#define IBMVETH_BUF_CSUM_GOOD	0x01000000
 | 
						|
#define IBMVETH_BUF_LEN_MASK	0x00FFFFFF
 | 
						|
};
 | 
						|
 | 
						|
union ibmveth_buf_desc {
 | 
						|
    u64 desc;
 | 
						|
    struct ibmveth_buf_desc_fields fields;
 | 
						|
};
 | 
						|
 | 
						|
struct ibmveth_rx_q_entry {
 | 
						|
	__be32 flags_off;
 | 
						|
#define IBMVETH_RXQ_TOGGLE		0x80000000
 | 
						|
#define IBMVETH_RXQ_TOGGLE_SHIFT	31
 | 
						|
#define IBMVETH_RXQ_VALID		0x40000000
 | 
						|
#define IBMVETH_RXQ_LRG_PKT		0x04000000
 | 
						|
#define IBMVETH_RXQ_NO_CSUM		0x02000000
 | 
						|
#define IBMVETH_RXQ_CSUM_GOOD		0x01000000
 | 
						|
#define IBMVETH_RXQ_OFF_MASK		0x0000FFFF
 | 
						|
 | 
						|
	__be32 length;
 | 
						|
	/* correlator is only used by the OS, no need to byte swap */
 | 
						|
	u64 correlator;
 | 
						|
};
 | 
						|
 | 
						|
#endif /* _IBMVETH_H */
 |