The ice driver stores VF structures in a simple array which is allocated once at the time of VF creation. The VF structures are then accessed from the array by their VF ID. The ID must be between 0 and the number of allocated VFs. Multiple threads can access this table: * .ndo operations such as .ndo_get_vf_cfg or .ndo_set_vf_trust * interrupts, such as due to messages from the VF using the virtchnl communication * processing such as device reset * commands to add or remove VFs The current implementation does not keep track of when all threads are done operating on a VF and can potentially result in use-after-free issues caused by one thread accessing a VF structure after it has been released when removing VFs. Some of these are prevented with various state flags and checks. In addition, this structure is quite static and does not support a planned future where virtualization can be more dynamic. As we begin to look at supporting Scalable IOV with the ice driver (as opposed to just supporting Single Root IOV), this structure is not sufficient. In the future, VFs will be able to be added and removed individually and dynamically. To allow for this, and to better protect against a whole class of use-after-free bugs, replace the VF storage with a combination of a hash table and krefs to reference track all of the accesses to VFs through the hash table. A hash table still allows efficient look up of the VF given its ID, but also allows adding and removing VFs. It does not require contiguous VF IDs. The use of krefs allows the cleanup of the VF memory to be delayed until after all threads have released their reference (by calling ice_put_vf). To prevent corruption of the hash table, a combination of RCU and the mutex table_lock are used. Addition and removal from the hash table use the RCU-aware hash macros. This allows simple read-only look ups that iterate to locate a single VF can be fast using RCU. Accesses which modify the hash table, or which can't take RCU because they sleep, will hold the mutex lock. By using this design, we have a stronger guarantee that the VF structure can't be released until after all threads are finished operating on it. We also pave the way for the more dynamic Scalable IOV implementation in the future. Signed-off-by: Jacob Keller <jacob.e.keller@intel.com> Tested-by: Konrad Jankowski <konrad0.jankowski@intel.com> Signed-off-by: Tony Nguyen <anthony.l.nguyen@intel.com>
441 lines
14 KiB
C
441 lines
14 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/* Copyright (c) 2018, Intel Corporation. */
|
|
|
|
#ifndef _ICE_VIRTCHNL_PF_H_
|
|
#define _ICE_VIRTCHNL_PF_H_
|
|
#include "ice.h"
|
|
#include "ice_virtchnl_fdir.h"
|
|
#include "ice_vsi_vlan_ops.h"
|
|
|
|
/* Restrict number of MAC Addr and VLAN that non-trusted VF can programmed */
|
|
#define ICE_MAX_VLAN_PER_VF 8
|
|
/* MAC filters: 1 is reserved for the VF's default/perm_addr/LAA MAC, 1 for
|
|
* broadcast, and 16 for additional unicast/multicast filters
|
|
*/
|
|
#define ICE_MAX_MACADDR_PER_VF 18
|
|
|
|
/* Malicious Driver Detection */
|
|
#define ICE_DFLT_NUM_INVAL_MSGS_ALLOWED 10
|
|
#define ICE_MDD_EVENTS_THRESHOLD 30
|
|
|
|
/* Static VF transaction/status register def */
|
|
#define VF_DEVICE_STATUS 0xAA
|
|
#define VF_TRANS_PENDING_M 0x20
|
|
|
|
/* wait defines for polling PF_PCI_CIAD register status */
|
|
#define ICE_PCI_CIAD_WAIT_COUNT 100
|
|
#define ICE_PCI_CIAD_WAIT_DELAY_US 1
|
|
|
|
/* VF resource constraints */
|
|
#define ICE_MAX_VF_COUNT 256
|
|
#define ICE_MIN_QS_PER_VF 1
|
|
#define ICE_NONQ_VECS_VF 1
|
|
#define ICE_MAX_SCATTER_QS_PER_VF 16
|
|
#define ICE_MAX_RSS_QS_PER_VF 16
|
|
#define ICE_NUM_VF_MSIX_MED 17
|
|
#define ICE_NUM_VF_MSIX_SMALL 5
|
|
#define ICE_NUM_VF_MSIX_MULTIQ_MIN 3
|
|
#define ICE_MIN_INTR_PER_VF (ICE_MIN_QS_PER_VF + 1)
|
|
#define ICE_MAX_VF_RESET_TRIES 40
|
|
#define ICE_MAX_VF_RESET_SLEEP_MS 20
|
|
|
|
/* VF Hash Table access functions
|
|
*
|
|
* These functions provide abstraction for interacting with the VF hash table.
|
|
* In general, direct access to the hash table should be avoided outside of
|
|
* these functions where possible.
|
|
*
|
|
* The VF entries in the hash table are protected by reference counting to
|
|
* track lifetime of accesses from the table. The ice_get_vf_by_id() function
|
|
* obtains a reference to the VF structure which must be dropped by using
|
|
* ice_put_vf().
|
|
*/
|
|
|
|
/**
|
|
* ice_for_each_vf - Iterate over each VF entry
|
|
* @pf: pointer to the PF private structure
|
|
* @bkt: bucket index used for iteration
|
|
* @vf: pointer to the VF entry currently being processed in the loop.
|
|
*
|
|
* The bkt variable is an unsigned integer iterator used to traverse the VF
|
|
* entries. It is *not* guaranteed to be the VF's vf_id. Do not assume it is.
|
|
* Use vf->vf_id to get the id number if needed.
|
|
*
|
|
* The caller is expected to be under the table_lock mutex for the entire
|
|
* loop. Use this iterator if your loop is long or if it might sleep.
|
|
*/
|
|
#define ice_for_each_vf(pf, bkt, vf) \
|
|
hash_for_each((pf)->vfs.table, (bkt), (vf), entry)
|
|
|
|
/**
|
|
* ice_for_each_vf_rcu - Iterate over each VF entry protected by RCU
|
|
* @pf: pointer to the PF private structure
|
|
* @bkt: bucket index used for iteration
|
|
* @vf: pointer to the VF entry currently being processed in the loop.
|
|
*
|
|
* The bkt variable is an unsigned integer iterator used to traverse the VF
|
|
* entries. It is *not* guaranteed to be the VF's vf_id. Do not assume it is.
|
|
* Use vf->vf_id to get the id number if needed.
|
|
*
|
|
* The caller is expected to be under rcu_read_lock() for the entire loop.
|
|
* Only use this iterator if your loop is short and you can guarantee it does
|
|
* not sleep.
|
|
*/
|
|
#define ice_for_each_vf_rcu(pf, bkt, vf) \
|
|
hash_for_each_rcu((pf)->vfs.table, (bkt), (vf), entry)
|
|
|
|
/* Specific VF states */
|
|
enum ice_vf_states {
|
|
ICE_VF_STATE_INIT = 0, /* PF is initializing VF */
|
|
ICE_VF_STATE_ACTIVE, /* VF resources are allocated for use */
|
|
ICE_VF_STATE_QS_ENA, /* VF queue(s) enabled */
|
|
ICE_VF_STATE_DIS,
|
|
ICE_VF_STATE_MC_PROMISC,
|
|
ICE_VF_STATE_UC_PROMISC,
|
|
ICE_VF_STATES_NBITS
|
|
};
|
|
|
|
/* VF capabilities */
|
|
enum ice_virtchnl_cap {
|
|
ICE_VIRTCHNL_VF_CAP_L2 = 0,
|
|
ICE_VIRTCHNL_VF_CAP_PRIVILEGE,
|
|
};
|
|
|
|
struct ice_time_mac {
|
|
unsigned long time_modified;
|
|
u8 addr[ETH_ALEN];
|
|
};
|
|
|
|
/* VF MDD events print structure */
|
|
struct ice_mdd_vf_events {
|
|
u16 count; /* total count of Rx|Tx events */
|
|
/* count number of the last printed event */
|
|
u16 last_printed;
|
|
};
|
|
|
|
struct ice_vf;
|
|
|
|
struct ice_vc_vf_ops {
|
|
int (*get_ver_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*get_vf_res_msg)(struct ice_vf *vf, u8 *msg);
|
|
void (*reset_vf)(struct ice_vf *vf);
|
|
int (*add_mac_addr_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*del_mac_addr_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*cfg_qs_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*ena_qs_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*dis_qs_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*request_qs_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*cfg_irq_map_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*config_rss_key)(struct ice_vf *vf, u8 *msg);
|
|
int (*config_rss_lut)(struct ice_vf *vf, u8 *msg);
|
|
int (*get_stats_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*cfg_promiscuous_mode_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*add_vlan_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*remove_vlan_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*ena_vlan_stripping)(struct ice_vf *vf);
|
|
int (*dis_vlan_stripping)(struct ice_vf *vf);
|
|
int (*handle_rss_cfg_msg)(struct ice_vf *vf, u8 *msg, bool add);
|
|
int (*add_fdir_fltr_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*del_fdir_fltr_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*get_offload_vlan_v2_caps)(struct ice_vf *vf);
|
|
int (*add_vlan_v2_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*remove_vlan_v2_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*ena_vlan_stripping_v2_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*dis_vlan_stripping_v2_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*ena_vlan_insertion_v2_msg)(struct ice_vf *vf, u8 *msg);
|
|
int (*dis_vlan_insertion_v2_msg)(struct ice_vf *vf, u8 *msg);
|
|
};
|
|
|
|
/* Virtchnl/SR-IOV config info */
|
|
struct ice_vfs {
|
|
DECLARE_HASHTABLE(table, 8); /* table of VF entries */
|
|
struct mutex table_lock; /* Lock for protecting the hash table */
|
|
u16 num_supported; /* max supported VFs on this PF */
|
|
u16 num_qps_per; /* number of queue pairs per VF */
|
|
u16 num_msix_per; /* number of MSI-X vectors per VF */
|
|
unsigned long last_printed_mdd_jiffies; /* MDD message rate limit */
|
|
DECLARE_BITMAP(malvfs, ICE_MAX_VF_COUNT); /* malicious VF indicator */
|
|
};
|
|
|
|
/* VF information structure */
|
|
struct ice_vf {
|
|
struct hlist_node entry;
|
|
struct rcu_head rcu;
|
|
struct kref refcnt;
|
|
struct ice_pf *pf;
|
|
|
|
/* Used during virtchnl message handling and NDO ops against the VF
|
|
* that will trigger a VFR
|
|
*/
|
|
struct mutex cfg_lock;
|
|
|
|
u16 vf_id; /* VF ID in the PF space */
|
|
u16 lan_vsi_idx; /* index into PF struct */
|
|
u16 ctrl_vsi_idx;
|
|
struct ice_vf_fdir fdir;
|
|
/* first vector index of this VF in the PF space */
|
|
int first_vector_idx;
|
|
struct ice_sw *vf_sw_id; /* switch ID the VF VSIs connect to */
|
|
struct virtchnl_version_info vf_ver;
|
|
u32 driver_caps; /* reported by VF driver */
|
|
struct virtchnl_ether_addr dev_lan_addr;
|
|
struct virtchnl_ether_addr hw_lan_addr;
|
|
struct ice_time_mac legacy_last_added_umac;
|
|
DECLARE_BITMAP(txq_ena, ICE_MAX_RSS_QS_PER_VF);
|
|
DECLARE_BITMAP(rxq_ena, ICE_MAX_RSS_QS_PER_VF);
|
|
struct ice_vlan port_vlan_info; /* Port VLAN ID, QoS, and TPID */
|
|
struct virtchnl_vlan_caps vlan_v2_caps;
|
|
u8 pf_set_mac:1; /* VF MAC address set by VMM admin */
|
|
u8 trusted:1;
|
|
u8 spoofchk:1;
|
|
u8 link_forced:1;
|
|
u8 link_up:1; /* only valid if VF link is forced */
|
|
/* VSI indices - actual VSI pointers are maintained in the PF structure
|
|
* When assigned, these will be non-zero, because VSI 0 is always
|
|
* the main LAN VSI for the PF.
|
|
*/
|
|
u16 lan_vsi_num; /* ID as used by firmware */
|
|
unsigned int min_tx_rate; /* Minimum Tx bandwidth limit in Mbps */
|
|
unsigned int max_tx_rate; /* Maximum Tx bandwidth limit in Mbps */
|
|
DECLARE_BITMAP(vf_states, ICE_VF_STATES_NBITS); /* VF runtime states */
|
|
|
|
u64 num_inval_msgs; /* number of continuous invalid msgs */
|
|
u64 num_valid_msgs; /* number of valid msgs detected */
|
|
unsigned long vf_caps; /* VF's adv. capabilities */
|
|
u8 num_req_qs; /* num of queue pairs requested by VF */
|
|
u16 num_mac;
|
|
u16 num_vf_qs; /* num of queue configured per VF */
|
|
struct ice_mdd_vf_events mdd_rx_events;
|
|
struct ice_mdd_vf_events mdd_tx_events;
|
|
DECLARE_BITMAP(opcodes_allowlist, VIRTCHNL_OP_MAX);
|
|
|
|
struct ice_repr *repr;
|
|
|
|
struct ice_vc_vf_ops vc_ops;
|
|
|
|
/* devlink port data */
|
|
struct devlink_port devlink_port;
|
|
};
|
|
|
|
#ifdef CONFIG_PCI_IOV
|
|
struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id);
|
|
void ice_put_vf(struct ice_vf *vf);
|
|
bool ice_has_vfs(struct ice_pf *pf);
|
|
u16 ice_get_num_vfs(struct ice_pf *pf);
|
|
struct ice_vsi *ice_get_vf_vsi(struct ice_vf *vf);
|
|
void ice_process_vflr_event(struct ice_pf *pf);
|
|
int ice_sriov_configure(struct pci_dev *pdev, int num_vfs);
|
|
int ice_set_vf_mac(struct net_device *netdev, int vf_id, u8 *mac);
|
|
int
|
|
ice_get_vf_cfg(struct net_device *netdev, int vf_id, struct ifla_vf_info *ivi);
|
|
|
|
void ice_free_vfs(struct ice_pf *pf);
|
|
void ice_vc_process_vf_msg(struct ice_pf *pf, struct ice_rq_event_info *event);
|
|
void ice_vc_notify_link_state(struct ice_pf *pf);
|
|
void ice_vc_notify_reset(struct ice_pf *pf);
|
|
void ice_vc_notify_vf_link_state(struct ice_vf *vf);
|
|
void ice_vc_change_ops_to_repr(struct ice_vc_vf_ops *ops);
|
|
void ice_vc_set_dflt_vf_ops(struct ice_vc_vf_ops *ops);
|
|
bool ice_reset_all_vfs(struct ice_pf *pf, bool is_vflr);
|
|
bool ice_reset_vf(struct ice_vf *vf, bool is_vflr);
|
|
void ice_restore_all_vfs_msi_state(struct pci_dev *pdev);
|
|
bool
|
|
ice_is_malicious_vf(struct ice_pf *pf, struct ice_rq_event_info *event,
|
|
u16 num_msg_proc, u16 num_msg_pending);
|
|
|
|
int
|
|
ice_set_vf_port_vlan(struct net_device *netdev, int vf_id, u16 vlan_id, u8 qos,
|
|
__be16 vlan_proto);
|
|
|
|
int
|
|
ice_set_vf_bw(struct net_device *netdev, int vf_id, int min_tx_rate,
|
|
int max_tx_rate);
|
|
|
|
int ice_set_vf_trust(struct net_device *netdev, int vf_id, bool trusted);
|
|
|
|
int ice_set_vf_link_state(struct net_device *netdev, int vf_id, int link_state);
|
|
|
|
int ice_check_vf_ready_for_cfg(struct ice_vf *vf);
|
|
|
|
bool ice_is_vf_disabled(struct ice_vf *vf);
|
|
|
|
int ice_set_vf_spoofchk(struct net_device *netdev, int vf_id, bool ena);
|
|
|
|
int ice_calc_vf_reg_idx(struct ice_vf *vf, struct ice_q_vector *q_vector);
|
|
|
|
void ice_set_vf_state_qs_dis(struct ice_vf *vf);
|
|
int
|
|
ice_get_vf_stats(struct net_device *netdev, int vf_id,
|
|
struct ifla_vf_stats *vf_stats);
|
|
bool ice_is_any_vf_in_promisc(struct ice_pf *pf);
|
|
void
|
|
ice_vf_lan_overflow_event(struct ice_pf *pf, struct ice_rq_event_info *event);
|
|
void ice_print_vfs_mdd_events(struct ice_pf *pf);
|
|
void ice_print_vf_rx_mdd_event(struct ice_vf *vf);
|
|
bool
|
|
ice_vc_validate_pattern(struct ice_vf *vf, struct virtchnl_proto_hdrs *proto);
|
|
struct ice_vsi *ice_vf_ctrl_vsi_setup(struct ice_vf *vf);
|
|
int
|
|
ice_vc_send_msg_to_vf(struct ice_vf *vf, u32 v_opcode,
|
|
enum virtchnl_status_code v_retval, u8 *msg, u16 msglen);
|
|
bool ice_vc_isvalid_vsi_id(struct ice_vf *vf, u16 vsi_id);
|
|
bool ice_vf_is_port_vlan_ena(struct ice_vf *vf);
|
|
#else /* CONFIG_PCI_IOV */
|
|
static inline struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static inline void ice_put_vf(struct ice_vf *vf)
|
|
{
|
|
}
|
|
|
|
static inline bool ice_has_vfs(struct ice_pf *pf)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static inline u16 ice_get_num_vfs(struct ice_pf *pf)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline void ice_process_vflr_event(struct ice_pf *pf) { }
|
|
static inline void ice_free_vfs(struct ice_pf *pf) { }
|
|
static inline
|
|
void ice_vc_process_vf_msg(struct ice_pf *pf, struct ice_rq_event_info *event) { }
|
|
static inline void ice_vc_notify_link_state(struct ice_pf *pf) { }
|
|
static inline void ice_vc_notify_reset(struct ice_pf *pf) { }
|
|
static inline void ice_vc_notify_vf_link_state(struct ice_vf *vf) { }
|
|
static inline void ice_vc_change_ops_to_repr(struct ice_vc_vf_ops *ops) { }
|
|
static inline void ice_vc_set_dflt_vf_ops(struct ice_vc_vf_ops *ops) { }
|
|
static inline void ice_set_vf_state_qs_dis(struct ice_vf *vf) { }
|
|
static inline
|
|
void ice_vf_lan_overflow_event(struct ice_pf *pf, struct ice_rq_event_info *event) { }
|
|
static inline void ice_print_vfs_mdd_events(struct ice_pf *pf) { }
|
|
static inline void ice_print_vf_rx_mdd_event(struct ice_vf *vf) { }
|
|
static inline void ice_restore_all_vfs_msi_state(struct pci_dev *pdev) { }
|
|
|
|
static inline int ice_check_vf_ready_for_cfg(struct ice_vf *vf)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline bool ice_is_vf_disabled(struct ice_vf *vf)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
static inline struct ice_vsi *ice_get_vf_vsi(struct ice_vf *vf)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static inline bool
|
|
ice_is_malicious_vf(struct ice_pf __always_unused *pf,
|
|
struct ice_rq_event_info __always_unused *event,
|
|
u16 __always_unused num_msg_proc,
|
|
u16 __always_unused num_msg_pending)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static inline bool
|
|
ice_reset_all_vfs(struct ice_pf __always_unused *pf,
|
|
bool __always_unused is_vflr)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
static inline bool
|
|
ice_reset_vf(struct ice_vf __always_unused *vf, bool __always_unused is_vflr)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
static inline int
|
|
ice_sriov_configure(struct pci_dev __always_unused *pdev,
|
|
int __always_unused num_vfs)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_set_vf_mac(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id, u8 __always_unused *mac)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_get_vf_cfg(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id,
|
|
struct ifla_vf_info __always_unused *ivi)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_set_vf_trust(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id, bool __always_unused trusted)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_set_vf_port_vlan(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id, u16 __always_unused vid,
|
|
u8 __always_unused qos, __be16 __always_unused v_proto)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_set_vf_spoofchk(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id, bool __always_unused ena)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_set_vf_link_state(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id, int __always_unused link_state)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_set_vf_bw(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id, int __always_unused min_tx_rate,
|
|
int __always_unused max_tx_rate)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline int
|
|
ice_calc_vf_reg_idx(struct ice_vf __always_unused *vf,
|
|
struct ice_q_vector __always_unused *q_vector)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline int
|
|
ice_get_vf_stats(struct net_device __always_unused *netdev,
|
|
int __always_unused vf_id,
|
|
struct ifla_vf_stats __always_unused *vf_stats)
|
|
{
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
static inline bool ice_is_any_vf_in_promisc(struct ice_pf __always_unused *pf)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static inline bool ice_vf_is_port_vlan_ena(struct ice_vf __always_unused *vf)
|
|
{
|
|
return false;
|
|
}
|
|
#endif /* CONFIG_PCI_IOV */
|
|
#endif /* _ICE_VIRTCHNL_PF_H_ */
|