mirror of
https://github.com/torvalds/linux.git
synced 2024-11-11 14:42:24 +00:00
ixgbe: reset interface on link loss with pending Tx work from the VF
ixgbe initiates a reset of the interface on link loss with pending Tx work in order to clear the rings. This patch extends the pending Tx work check to the VF interfaces with the same purpose. Signed-off-by: Emil Tantilov <emil.s.tantilov@intel.com> Tested-by: Phil Schmitt <phillip.j.schmitt@intel.com> Signed-off-by: Jeff Kirsher <jeffrey.t.kirsher@intel.com>
This commit is contained in:
parent
bcfd3432d1
commit
07923c17b1
@ -6319,25 +6319,55 @@ static void ixgbe_watchdog_link_is_down(struct ixgbe_adapter *adapter)
|
||||
ixgbe_ping_all_vfs(adapter);
|
||||
}
|
||||
|
||||
static bool ixgbe_ring_tx_pending(struct ixgbe_adapter *adapter)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 0; i < adapter->num_tx_queues; i++) {
|
||||
struct ixgbe_ring *tx_ring = adapter->tx_ring[i];
|
||||
|
||||
if (tx_ring->next_to_use != tx_ring->next_to_clean)
|
||||
return true;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool ixgbe_vf_tx_pending(struct ixgbe_adapter *adapter)
|
||||
{
|
||||
struct ixgbe_hw *hw = &adapter->hw;
|
||||
struct ixgbe_ring_feature *vmdq = &adapter->ring_feature[RING_F_VMDQ];
|
||||
u32 q_per_pool = __ALIGN_MASK(1, ~vmdq->mask);
|
||||
|
||||
int i, j;
|
||||
|
||||
if (!adapter->num_vfs)
|
||||
return false;
|
||||
|
||||
for (i = 0; i < adapter->num_vfs; i++) {
|
||||
for (j = 0; j < q_per_pool; j++) {
|
||||
u32 h, t;
|
||||
|
||||
h = IXGBE_READ_REG(hw, IXGBE_PVFTDHN(q_per_pool, i, j));
|
||||
t = IXGBE_READ_REG(hw, IXGBE_PVFTDTN(q_per_pool, i, j));
|
||||
|
||||
if (h != t)
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* ixgbe_watchdog_flush_tx - flush queues on link down
|
||||
* @adapter: pointer to the device adapter structure
|
||||
**/
|
||||
static void ixgbe_watchdog_flush_tx(struct ixgbe_adapter *adapter)
|
||||
{
|
||||
int i;
|
||||
int some_tx_pending = 0;
|
||||
|
||||
if (!netif_carrier_ok(adapter->netdev)) {
|
||||
for (i = 0; i < adapter->num_tx_queues; i++) {
|
||||
struct ixgbe_ring *tx_ring = adapter->tx_ring[i];
|
||||
if (tx_ring->next_to_use != tx_ring->next_to_clean) {
|
||||
some_tx_pending = 1;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (some_tx_pending) {
|
||||
if (ixgbe_ring_tx_pending(adapter) ||
|
||||
ixgbe_vf_tx_pending(adapter)) {
|
||||
/* We've lost link, so the controller stops DMA,
|
||||
* but we've got queued Tx work that's never going
|
||||
* to get done, so reset controller to flush Tx.
|
||||
|
@ -2194,6 +2194,8 @@ enum {
|
||||
#define IXGBE_VFLRE(_i) ((((_i) & 1) ? 0x001C0 : 0x00600))
|
||||
#define IXGBE_VFLREC(_i) (0x00700 + ((_i) * 4))
|
||||
/* Translated register #defines */
|
||||
#define IXGBE_PVFTDH(P) (0x06010 + (0x40 * (P)))
|
||||
#define IXGBE_PVFTDT(P) (0x06018 + (0x40 * (P)))
|
||||
#define IXGBE_PVFTDWBAL(P) (0x06038 + (0x40 * (P)))
|
||||
#define IXGBE_PVFTDWBAH(P) (0x0603C + (0x40 * (P)))
|
||||
|
||||
@ -2202,6 +2204,11 @@ enum {
|
||||
#define IXGBE_PVFTDWBAHn(q_per_pool, vf_number, vf_q_index) \
|
||||
(IXGBE_PVFTDWBAH((q_per_pool)*(vf_number) + (vf_q_index)))
|
||||
|
||||
#define IXGBE_PVFTDHN(q_per_pool, vf_number, vf_q_index) \
|
||||
(IXGBE_PVFTDH((q_per_pool)*(vf_number) + (vf_q_index)))
|
||||
#define IXGBE_PVFTDTN(q_per_pool, vf_number, vf_q_index) \
|
||||
(IXGBE_PVFTDT((q_per_pool)*(vf_number) + (vf_q_index)))
|
||||
|
||||
enum ixgbe_fdir_pballoc_type {
|
||||
IXGBE_FDIR_PBALLOC_NONE = 0,
|
||||
IXGBE_FDIR_PBALLOC_64K = 1,
|
||||
|
Loading…
Reference in New Issue
Block a user