|
@@ -694,14 +694,12 @@ repost:
|
|
|
static inline int post_send(struct ipoib_dev_priv *priv,
|
|
|
struct ipoib_cm_tx *tx,
|
|
|
unsigned int wr_id,
|
|
|
- u64 addr, int len)
|
|
|
+ struct ipoib_tx_buf *tx_req)
|
|
|
{
|
|
|
struct ib_send_wr *bad_wr;
|
|
|
|
|
|
- priv->tx_sge[0].addr = addr;
|
|
|
- priv->tx_sge[0].length = len;
|
|
|
+ ipoib_build_sge(priv, tx_req);
|
|
|
|
|
|
- priv->tx_wr.num_sge = 1;
|
|
|
priv->tx_wr.wr_id = wr_id | IPOIB_OP_CM;
|
|
|
|
|
|
return ib_post_send(tx->qp, &priv->tx_wr, &bad_wr);
|
|
@@ -710,8 +708,7 @@ static inline int post_send(struct ipoib_dev_priv *priv,
|
|
|
void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx)
|
|
|
{
|
|
|
struct ipoib_dev_priv *priv = netdev_priv(dev);
|
|
|
- struct ipoib_cm_tx_buf *tx_req;
|
|
|
- u64 addr;
|
|
|
+ struct ipoib_tx_buf *tx_req;
|
|
|
int rc;
|
|
|
|
|
|
if (unlikely(skb->len > tx->mtu)) {
|
|
@@ -735,24 +732,21 @@ void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_
|
|
|
*/
|
|
|
tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)];
|
|
|
tx_req->skb = skb;
|
|
|
- addr = ib_dma_map_single(priv->ca, skb->data, skb->len, DMA_TO_DEVICE);
|
|
|
- if (unlikely(ib_dma_mapping_error(priv->ca, addr))) {
|
|
|
+
|
|
|
+ if (unlikely(ipoib_dma_map_tx(priv->ca, tx_req))) {
|
|
|
++dev->stats.tx_errors;
|
|
|
dev_kfree_skb_any(skb);
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
- tx_req->mapping = addr;
|
|
|
-
|
|
|
skb_orphan(skb);
|
|
|
skb_dst_drop(skb);
|
|
|
|
|
|
- rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1),
|
|
|
- addr, skb->len);
|
|
|
+ rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), tx_req);
|
|
|
if (unlikely(rc)) {
|
|
|
ipoib_warn(priv, "post_send failed, error %d\n", rc);
|
|
|
++dev->stats.tx_errors;
|
|
|
- ib_dma_unmap_single(priv->ca, addr, skb->len, DMA_TO_DEVICE);
|
|
|
+ ipoib_dma_unmap_tx(priv, tx_req);
|
|
|
dev_kfree_skb_any(skb);
|
|
|
} else {
|
|
|
dev->trans_start = jiffies;
|
|
@@ -777,7 +771,7 @@ void ipoib_cm_handle_tx_wc(struct net_device *dev, struct ib_wc *wc)
|
|
|
struct ipoib_dev_priv *priv = netdev_priv(dev);
|
|
|
struct ipoib_cm_tx *tx = wc->qp->qp_context;
|
|
|
unsigned int wr_id = wc->wr_id & ~IPOIB_OP_CM;
|
|
|
- struct ipoib_cm_tx_buf *tx_req;
|
|
|
+ struct ipoib_tx_buf *tx_req;
|
|
|
unsigned long flags;
|
|
|
|
|
|
ipoib_dbg_data(priv, "cm send completion: id %d, status: %d\n",
|
|
@@ -791,7 +785,7 @@ void ipoib_cm_handle_tx_wc(struct net_device *dev, struct ib_wc *wc)
|
|
|
|
|
|
tx_req = &tx->tx_ring[wr_id];
|
|
|
|
|
|
- ib_dma_unmap_single(priv->ca, tx_req->mapping, tx_req->skb->len, DMA_TO_DEVICE);
|
|
|
+ ipoib_dma_unmap_tx(priv, tx_req);
|
|
|
|
|
|
/* FIXME: is this right? Shouldn't we only increment on success? */
|
|
|
++dev->stats.tx_packets;
|
|
@@ -1036,6 +1030,9 @@ static struct ib_qp *ipoib_cm_create_tx_qp(struct net_device *dev, struct ipoib_
|
|
|
|
|
|
struct ib_qp *tx_qp;
|
|
|
|
|
|
+ if (dev->features & NETIF_F_SG)
|
|
|
+ attr.cap.max_send_sge = MAX_SKB_FRAGS + 1;
|
|
|
+
|
|
|
tx_qp = ib_create_qp(priv->pd, &attr);
|
|
|
if (PTR_ERR(tx_qp) == -EINVAL) {
|
|
|
ipoib_warn(priv, "can't use GFP_NOIO for QPs on device %s, using GFP_KERNEL\n",
|
|
@@ -1170,7 +1167,7 @@ err_tx:
|
|
|
static void ipoib_cm_tx_destroy(struct ipoib_cm_tx *p)
|
|
|
{
|
|
|
struct ipoib_dev_priv *priv = netdev_priv(p->dev);
|
|
|
- struct ipoib_cm_tx_buf *tx_req;
|
|
|
+ struct ipoib_tx_buf *tx_req;
|
|
|
unsigned long begin;
|
|
|
|
|
|
ipoib_dbg(priv, "Destroy active connection 0x%x head 0x%x tail 0x%x\n",
|
|
@@ -1197,8 +1194,7 @@ timeout:
|
|
|
|
|
|
while ((int) p->tx_tail - (int) p->tx_head < 0) {
|
|
|
tx_req = &p->tx_ring[p->tx_tail & (ipoib_sendq_size - 1)];
|
|
|
- ib_dma_unmap_single(priv->ca, tx_req->mapping, tx_req->skb->len,
|
|
|
- DMA_TO_DEVICE);
|
|
|
+ ipoib_dma_unmap_tx(priv, tx_req);
|
|
|
dev_kfree_skb_any(tx_req->skb);
|
|
|
++p->tx_tail;
|
|
|
netif_tx_lock_bh(p->dev);
|
|
@@ -1455,7 +1451,6 @@ static void ipoib_cm_stale_task(struct work_struct *work)
|
|
|
spin_unlock_irq(&priv->lock);
|
|
|
}
|
|
|
|
|
|
-
|
|
|
static ssize_t show_mode(struct device *d, struct device_attribute *attr,
|
|
|
char *buf)
|
|
|
{
|