|
@@ -100,12 +100,17 @@ static int hw_atl_b0_hw_reset(struct aq_hw_s *self)
|
|
|
return err;
|
|
|
}
|
|
|
|
|
|
+static int hw_atl_b0_set_fc(struct aq_hw_s *self, u32 fc, u32 tc)
|
|
|
+{
|
|
|
+ hw_atl_rpb_rx_xoff_en_per_tc_set(self, !!(fc & AQ_NIC_FC_RX), tc);
|
|
|
+ return 0;
|
|
|
+}
|
|
|
+
|
|
|
static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
|
|
|
{
|
|
|
u32 tc = 0U;
|
|
|
u32 buff_size = 0U;
|
|
|
unsigned int i_priority = 0U;
|
|
|
- bool is_rx_flow_control = false;
|
|
|
|
|
|
/* TPS Descriptor rate init */
|
|
|
hw_atl_tps_tx_pkt_shed_desc_rate_curr_time_res_set(self, 0x0U);
|
|
@@ -138,7 +143,6 @@ static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
|
|
|
|
|
|
/* QoS Rx buf size per TC */
|
|
|
tc = 0;
|
|
|
- is_rx_flow_control = (AQ_NIC_FC_RX & self->aq_nic_cfg->flow_control);
|
|
|
buff_size = HW_ATL_B0_RXBUF_MAX;
|
|
|
|
|
|
hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, buff_size, tc);
|
|
@@ -150,7 +154,8 @@ static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
|
|
|
(buff_size *
|
|
|
(1024U / 32U) * 50U) /
|
|
|
100U, tc);
|
|
|
- hw_atl_rpb_rx_xoff_en_per_tc_set(self, is_rx_flow_control ? 1U : 0U, tc);
|
|
|
+
|
|
|
+ hw_atl_b0_set_fc(self, self->aq_nic_cfg->flow_control, tc);
|
|
|
|
|
|
/* QoS 802.1p priority -> TC mapping */
|
|
|
for (i_priority = 8U; i_priority--;)
|
|
@@ -963,4 +968,5 @@ const struct aq_hw_ops hw_atl_ops_b0 = {
|
|
|
.hw_get_regs = hw_atl_utils_hw_get_regs,
|
|
|
.hw_get_hw_stats = hw_atl_utils_get_hw_stats,
|
|
|
.hw_get_fw_version = hw_atl_utils_get_fw_version,
|
|
|
+ .hw_set_fc = hw_atl_b0_set_fc,
|
|
|
};
|