|
|
@@ -122,7 +122,7 @@ static void queue_process(struct work_struct *work)
|
|
|
txq = netdev_get_tx_queue(dev, q_index);
|
|
|
HARD_TX_LOCK(dev, txq, smp_processor_id());
|
|
|
if (netif_xmit_frozen_or_stopped(txq) ||
|
|
|
- netpoll_start_xmit(skb, dev, txq) != NETDEV_TX_OK) {
|
|
|
+ !dev_xmit_complete(netpoll_start_xmit(skb, dev, txq))) {
|
|
|
skb_queue_head(&npinfo->txq, skb);
|
|
|
HARD_TX_UNLOCK(dev, txq);
|
|
|
local_irq_restore(flags);
|
|
|
@@ -335,7 +335,7 @@ void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb,
|
|
|
|
|
|
HARD_TX_UNLOCK(dev, txq);
|
|
|
|
|
|
- if (status == NETDEV_TX_OK)
|
|
|
+ if (dev_xmit_complete(status))
|
|
|
break;
|
|
|
|
|
|
}
|
|
|
@@ -352,7 +352,7 @@ void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb,
|
|
|
|
|
|
}
|
|
|
|
|
|
- if (status != NETDEV_TX_OK) {
|
|
|
+ if (!dev_xmit_complete(status)) {
|
|
|
skb_queue_tail(&npinfo->txq, skb);
|
|
|
schedule_delayed_work(&npinfo->tx_work,0);
|
|
|
}
|