|
|
@@ -3458,13 +3458,18 @@ static int bnxt_hwrm_func_drv_rgtr(struct bnxt *bp)
|
|
|
req.ver_upd = DRV_VER_UPD;
|
|
|
|
|
|
if (BNXT_PF(bp)) {
|
|
|
- DECLARE_BITMAP(vf_req_snif_bmap, 256);
|
|
|
- u32 *data = (u32 *)vf_req_snif_bmap;
|
|
|
+ u32 data[8];
|
|
|
int i;
|
|
|
|
|
|
- memset(vf_req_snif_bmap, 0, sizeof(vf_req_snif_bmap));
|
|
|
- for (i = 0; i < ARRAY_SIZE(bnxt_vf_req_snif); i++)
|
|
|
- __set_bit(bnxt_vf_req_snif[i], vf_req_snif_bmap);
|
|
|
+ memset(data, 0, sizeof(data));
|
|
|
+ for (i = 0; i < ARRAY_SIZE(bnxt_vf_req_snif); i++) {
|
|
|
+ u16 cmd = bnxt_vf_req_snif[i];
|
|
|
+ unsigned int bit, idx;
|
|
|
+
|
|
|
+ idx = cmd / 32;
|
|
|
+ bit = cmd % 32;
|
|
|
+ data[idx] |= 1 << bit;
|
|
|
+ }
|
|
|
|
|
|
for (i = 0; i < 8; i++)
|
|
|
req.vf_req_fwd[i] = cpu_to_le32(data[i]);
|
|
|
@@ -6279,6 +6284,12 @@ static int bnxt_open(struct net_device *dev)
|
|
|
return __bnxt_open_nic(bp, true, true);
|
|
|
}
|
|
|
|
|
|
+static bool bnxt_drv_busy(struct bnxt *bp)
|
|
|
+{
|
|
|
+ return (test_bit(BNXT_STATE_IN_SP_TASK, &bp->state) ||
|
|
|
+ test_bit(BNXT_STATE_READ_STATS, &bp->state));
|
|
|
+}
|
|
|
+
|
|
|
int bnxt_close_nic(struct bnxt *bp, bool irq_re_init, bool link_re_init)
|
|
|
{
|
|
|
int rc = 0;
|
|
|
@@ -6297,7 +6308,7 @@ int bnxt_close_nic(struct bnxt *bp, bool irq_re_init, bool link_re_init)
|
|
|
|
|
|
clear_bit(BNXT_STATE_OPEN, &bp->state);
|
|
|
smp_mb__after_atomic();
|
|
|
- while (test_bit(BNXT_STATE_IN_SP_TASK, &bp->state))
|
|
|
+ while (bnxt_drv_busy(bp))
|
|
|
msleep(20);
|
|
|
|
|
|
/* Flush rings and and disable interrupts */
|
|
|
@@ -6358,8 +6369,15 @@ bnxt_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
|
|
|
u32 i;
|
|
|
struct bnxt *bp = netdev_priv(dev);
|
|
|
|
|
|
- if (!bp->bnapi)
|
|
|
+ set_bit(BNXT_STATE_READ_STATS, &bp->state);
|
|
|
+ /* Make sure bnxt_close_nic() sees that we are reading stats before
|
|
|
+ * we check the BNXT_STATE_OPEN flag.
|
|
|
+ */
|
|
|
+ smp_mb__after_atomic();
|
|
|
+ if (!test_bit(BNXT_STATE_OPEN, &bp->state)) {
|
|
|
+ clear_bit(BNXT_STATE_READ_STATS, &bp->state);
|
|
|
return;
|
|
|
+ }
|
|
|
|
|
|
/* TODO check if we need to synchronize with bnxt_close path */
|
|
|
for (i = 0; i < bp->cp_nr_rings; i++) {
|
|
|
@@ -6406,6 +6424,7 @@ bnxt_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
|
|
|
stats->tx_fifo_errors = le64_to_cpu(tx->tx_fifo_underruns);
|
|
|
stats->tx_errors = le64_to_cpu(tx->tx_err);
|
|
|
}
|
|
|
+ clear_bit(BNXT_STATE_READ_STATS, &bp->state);
|
|
|
}
|
|
|
|
|
|
static bool bnxt_mc_list_updated(struct bnxt *bp, u32 *rx_mask)
|
|
|
@@ -6904,16 +6923,13 @@ static void bnxt_sp_task(struct work_struct *work)
|
|
|
}
|
|
|
|
|
|
/* Under rtnl_lock */
|
|
|
-int bnxt_reserve_rings(struct bnxt *bp, int tx, int rx, int tcs, int tx_xdp)
|
|
|
+int bnxt_reserve_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
|
|
|
+ int tx_xdp)
|
|
|
{
|
|
|
int max_rx, max_tx, tx_sets = 1;
|
|
|
int tx_rings_needed;
|
|
|
- bool sh = true;
|
|
|
int rc;
|
|
|
|
|
|
- if (!(bp->flags & BNXT_FLAG_SHARED_RINGS))
|
|
|
- sh = false;
|
|
|
-
|
|
|
if (tcs)
|
|
|
tx_sets = tcs;
|
|
|
|
|
|
@@ -7121,7 +7137,7 @@ int bnxt_setup_mq_tc(struct net_device *dev, u8 tc)
|
|
|
sh = true;
|
|
|
|
|
|
rc = bnxt_reserve_rings(bp, bp->tx_nr_rings_per_tc, bp->rx_nr_rings,
|
|
|
- tc, bp->tx_nr_rings_xdp);
|
|
|
+ sh, tc, bp->tx_nr_rings_xdp);
|
|
|
if (rc)
|
|
|
return rc;
|
|
|
|