|
@@ -947,7 +947,7 @@ static bool e1000_clean_rx_irq(struct e1000_ring *rx_ring, int *work_done,
|
|
|
if (*work_done >= work_to_do)
|
|
|
break;
|
|
|
(*work_done)++;
|
|
|
- rmb(); /* read descriptor and rx_buffer_info after status DD */
|
|
|
+ dma_rmb(); /* read descriptor and rx_buffer_info after status DD */
|
|
|
|
|
|
skb = buffer_info->skb;
|
|
|
buffer_info->skb = NULL;
|
|
@@ -1232,7 +1232,7 @@ static bool e1000_clean_tx_irq(struct e1000_ring *tx_ring)
|
|
|
(count < tx_ring->count)) {
|
|
|
bool cleaned = false;
|
|
|
|
|
|
- rmb(); /* read buffer_info after eop_desc */
|
|
|
+ dma_rmb(); /* read buffer_info after eop_desc */
|
|
|
for (; !cleaned; count++) {
|
|
|
tx_desc = E1000_TX_DESC(*tx_ring, i);
|
|
|
buffer_info = &tx_ring->buffer_info[i];
|
|
@@ -1332,7 +1332,7 @@ static bool e1000_clean_rx_irq_ps(struct e1000_ring *rx_ring, int *work_done,
|
|
|
break;
|
|
|
(*work_done)++;
|
|
|
skb = buffer_info->skb;
|
|
|
- rmb(); /* read descriptor and rx_buffer_info after status DD */
|
|
|
+ dma_rmb(); /* read descriptor and rx_buffer_info after status DD */
|
|
|
|
|
|
/* in the packet split case this is header only */
|
|
|
prefetch(skb->data - NET_IP_ALIGN);
|
|
@@ -1536,7 +1536,7 @@ static bool e1000_clean_jumbo_rx_irq(struct e1000_ring *rx_ring, int *work_done,
|
|
|
if (*work_done >= work_to_do)
|
|
|
break;
|
|
|
(*work_done)++;
|
|
|
- rmb(); /* read descriptor and rx_buffer_info after status DD */
|
|
|
+ dma_rmb(); /* read descriptor and rx_buffer_info after status DD */
|
|
|
|
|
|
skb = buffer_info->skb;
|
|
|
buffer_info->skb = NULL;
|