|
@@ -821,7 +821,6 @@ static int atl1c_sw_init(struct atl1c_adapter *adapter)
|
|
|
atl1c_set_rxbufsize(adapter, adapter->netdev);
|
|
|
atomic_set(&adapter->irq_sem, 1);
|
|
|
spin_lock_init(&adapter->mdio_lock);
|
|
|
- spin_lock_init(&adapter->tx_lock);
|
|
|
set_bit(__AT_DOWN, &adapter->flags);
|
|
|
|
|
|
return 0;
|
|
@@ -2206,7 +2205,6 @@ static netdev_tx_t atl1c_xmit_frame(struct sk_buff *skb,
|
|
|
struct net_device *netdev)
|
|
|
{
|
|
|
struct atl1c_adapter *adapter = netdev_priv(netdev);
|
|
|
- unsigned long flags;
|
|
|
u16 tpd_req = 1;
|
|
|
struct atl1c_tpd_desc *tpd;
|
|
|
enum atl1c_trans_queue type = atl1c_trans_normal;
|
|
@@ -2217,16 +2215,10 @@ static netdev_tx_t atl1c_xmit_frame(struct sk_buff *skb,
|
|
|
}
|
|
|
|
|
|
tpd_req = atl1c_cal_tpd_req(skb);
|
|
|
- if (!spin_trylock_irqsave(&adapter->tx_lock, flags)) {
|
|
|
- if (netif_msg_pktdata(adapter))
|
|
|
- dev_info(&adapter->pdev->dev, "tx locked\n");
|
|
|
- return NETDEV_TX_LOCKED;
|
|
|
- }
|
|
|
|
|
|
if (atl1c_tpd_avail(adapter, type) < tpd_req) {
|
|
|
/* no enough descriptor, just stop queue */
|
|
|
netif_stop_queue(netdev);
|
|
|
- spin_unlock_irqrestore(&adapter->tx_lock, flags);
|
|
|
return NETDEV_TX_BUSY;
|
|
|
}
|
|
|
|
|
@@ -2234,7 +2226,6 @@ static netdev_tx_t atl1c_xmit_frame(struct sk_buff *skb,
|
|
|
|
|
|
/* do TSO and check sum */
|
|
|
if (atl1c_tso_csum(adapter, skb, &tpd, type) != 0) {
|
|
|
- spin_unlock_irqrestore(&adapter->tx_lock, flags);
|
|
|
dev_kfree_skb_any(skb);
|
|
|
return NETDEV_TX_OK;
|
|
|
}
|
|
@@ -2257,12 +2248,10 @@ static netdev_tx_t atl1c_xmit_frame(struct sk_buff *skb,
|
|
|
"tx-skb droppted due to dma error\n");
|
|
|
/* roll back tpd/buffer */
|
|
|
atl1c_tx_rollback(adapter, tpd, type);
|
|
|
- spin_unlock_irqrestore(&adapter->tx_lock, flags);
|
|
|
dev_kfree_skb_any(skb);
|
|
|
} else {
|
|
|
netdev_sent_queue(adapter->netdev, skb->len);
|
|
|
atl1c_tx_queue(adapter, skb, tpd, type);
|
|
|
- spin_unlock_irqrestore(&adapter->tx_lock, flags);
|
|
|
}
|
|
|
|
|
|
return NETDEV_TX_OK;
|