|
@@ -47,7 +47,7 @@ int mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq, struct mlx5e_rx_wqe *wqe, u16 ix)
|
|
|
struct sk_buff *skb;
|
|
|
dma_addr_t dma_addr;
|
|
|
|
|
|
- skb = netdev_alloc_skb(rq->netdev, rq->wqe_sz);
|
|
|
+ skb = napi_alloc_skb(rq->cq.napi, rq->wqe_sz);
|
|
|
if (unlikely(!skb))
|
|
|
return -ENOMEM;
|
|
|
|
|
@@ -61,10 +61,8 @@ int mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq, struct mlx5e_rx_wqe *wqe, u16 ix)
|
|
|
if (unlikely(dma_mapping_error(rq->pdev, dma_addr)))
|
|
|
goto err_free_skb;
|
|
|
|
|
|
- skb_reserve(skb, MLX5E_NET_IP_ALIGN);
|
|
|
-
|
|
|
*((dma_addr_t *)skb->cb) = dma_addr;
|
|
|
- wqe->data.addr = cpu_to_be64(dma_addr + MLX5E_NET_IP_ALIGN);
|
|
|
+ wqe->data.addr = cpu_to_be64(dma_addr);
|
|
|
wqe->data.lkey = rq->mkey_be;
|
|
|
|
|
|
rq->skb[ix] = skb;
|
|
@@ -701,9 +699,9 @@ void mlx5e_handle_rx_cqe_mpwrq(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
|
|
|
goto mpwrq_cqe_out;
|
|
|
}
|
|
|
|
|
|
- skb = netdev_alloc_skb(rq->netdev,
|
|
|
- ALIGN(MLX5_MPWRQ_SMALL_PACKET_THRESHOLD,
|
|
|
- sizeof(long)));
|
|
|
+ skb = napi_alloc_skb(rq->cq.napi,
|
|
|
+ ALIGN(MLX5_MPWRQ_SMALL_PACKET_THRESHOLD,
|
|
|
+ sizeof(long)));
|
|
|
if (unlikely(!skb))
|
|
|
goto mpwrq_cqe_out;
|
|
|
|