|
@@ -96,9 +96,12 @@ void mlx5e_set_rq_type_params(struct mlx5_core_dev *mdev,
|
|
params->log_rq_size = is_kdump_kernel() ?
|
|
params->log_rq_size = is_kdump_kernel() ?
|
|
MLX5E_PARAMS_MINIMUM_LOG_RQ_SIZE :
|
|
MLX5E_PARAMS_MINIMUM_LOG_RQ_SIZE :
|
|
MLX5E_PARAMS_DEFAULT_LOG_RQ_SIZE;
|
|
MLX5E_PARAMS_DEFAULT_LOG_RQ_SIZE;
|
|
|
|
+ params->rq_headroom = params->xdp_prog ?
|
|
|
|
+ XDP_PACKET_HEADROOM : MLX5_RX_HEADROOM;
|
|
|
|
+ params->rq_headroom += NET_IP_ALIGN;
|
|
|
|
|
|
/* Extra room needed for build_skb */
|
|
/* Extra room needed for build_skb */
|
|
- params->lro_wqe_sz -= MLX5_RX_HEADROOM +
|
|
|
|
|
|
+ params->lro_wqe_sz -= params->rq_headroom +
|
|
SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
|
|
SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
|
|
}
|
|
}
|
|
|
|
|
|
@@ -579,13 +582,8 @@ static int mlx5e_alloc_rq(struct mlx5e_channel *c,
|
|
goto err_rq_wq_destroy;
|
|
goto err_rq_wq_destroy;
|
|
}
|
|
}
|
|
|
|
|
|
- if (rq->xdp_prog) {
|
|
|
|
- rq->buff.map_dir = DMA_BIDIRECTIONAL;
|
|
|
|
- rq->rx_headroom = XDP_PACKET_HEADROOM;
|
|
|
|
- } else {
|
|
|
|
- rq->buff.map_dir = DMA_FROM_DEVICE;
|
|
|
|
- rq->rx_headroom = MLX5_RX_HEADROOM;
|
|
|
|
- }
|
|
|
|
|
|
+ rq->buff.map_dir = rq->xdp_prog ? DMA_BIDIRECTIONAL : DMA_FROM_DEVICE;
|
|
|
|
+ rq->rx_headroom = params->rq_headroom;
|
|
|
|
|
|
switch (rq->wq_type) {
|
|
switch (rq->wq_type) {
|
|
case MLX5_WQ_TYPE_LINKED_LIST_STRIDING_RQ:
|
|
case MLX5_WQ_TYPE_LINKED_LIST_STRIDING_RQ:
|