|
@@ -51,6 +51,7 @@
|
|
|
#include <linux/slab.h>
|
|
|
#include <linux/prefetch.h>
|
|
|
#include <linux/sunrpc/addr.h>
|
|
|
+#include <linux/sunrpc/svc_rdma.h>
|
|
|
#include <asm/bitops.h>
|
|
|
#include <linux/module.h> /* try_module_get()/module_put() */
|
|
|
|
|
@@ -923,7 +924,7 @@ rpcrdma_buffer_create(struct rpcrdma_xprt *r_xprt)
|
|
|
}
|
|
|
|
|
|
INIT_LIST_HEAD(&buf->rb_recv_bufs);
|
|
|
- for (i = 0; i < buf->rb_max_requests + 2; i++) {
|
|
|
+ for (i = 0; i < buf->rb_max_requests + RPCRDMA_MAX_BC_REQUESTS; i++) {
|
|
|
struct rpcrdma_rep *rep;
|
|
|
|
|
|
rep = rpcrdma_create_rep(r_xprt);
|
|
@@ -1018,6 +1019,7 @@ rpcrdma_buffer_destroy(struct rpcrdma_buffer *buf)
|
|
|
rep = rpcrdma_buffer_get_rep_locked(buf);
|
|
|
rpcrdma_destroy_rep(ia, rep);
|
|
|
}
|
|
|
+ buf->rb_send_count = 0;
|
|
|
|
|
|
spin_lock(&buf->rb_reqslock);
|
|
|
while (!list_empty(&buf->rb_allreqs)) {
|
|
@@ -1032,6 +1034,7 @@ rpcrdma_buffer_destroy(struct rpcrdma_buffer *buf)
|
|
|
spin_lock(&buf->rb_reqslock);
|
|
|
}
|
|
|
spin_unlock(&buf->rb_reqslock);
|
|
|
+ buf->rb_recv_count = 0;
|
|
|
|
|
|
rpcrdma_destroy_mrs(buf);
|
|
|
}
|
|
@@ -1074,6 +1077,23 @@ rpcrdma_put_mw(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mw *mw)
|
|
|
spin_unlock(&buf->rb_mwlock);
|
|
|
}
|
|
|
|
|
|
+static struct rpcrdma_rep *
|
|
|
+rpcrdma_buffer_get_rep(struct rpcrdma_buffer *buffers)
|
|
|
+{
|
|
|
+ /* If an RPC previously completed without a reply (say, a
|
|
|
+ * credential problem or a soft timeout occurs) then hold off
|
|
|
+ * on supplying more Receive buffers until the number of new
|
|
|
+ * pending RPCs catches up to the number of posted Receives.
|
|
|
+ */
|
|
|
+ if (unlikely(buffers->rb_send_count < buffers->rb_recv_count))
|
|
|
+ return NULL;
|
|
|
+
|
|
|
+ if (unlikely(list_empty(&buffers->rb_recv_bufs)))
|
|
|
+ return NULL;
|
|
|
+ buffers->rb_recv_count++;
|
|
|
+ return rpcrdma_buffer_get_rep_locked(buffers);
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* Get a set of request/reply buffers.
|
|
|
*
|
|
@@ -1087,10 +1107,9 @@ rpcrdma_buffer_get(struct rpcrdma_buffer *buffers)
|
|
|
spin_lock(&buffers->rb_lock);
|
|
|
if (list_empty(&buffers->rb_send_bufs))
|
|
|
goto out_reqbuf;
|
|
|
+ buffers->rb_send_count++;
|
|
|
req = rpcrdma_buffer_get_req_locked(buffers);
|
|
|
- if (list_empty(&buffers->rb_recv_bufs))
|
|
|
- goto out_repbuf;
|
|
|
- req->rl_reply = rpcrdma_buffer_get_rep_locked(buffers);
|
|
|
+ req->rl_reply = rpcrdma_buffer_get_rep(buffers);
|
|
|
spin_unlock(&buffers->rb_lock);
|
|
|
return req;
|
|
|
|
|
@@ -1098,11 +1117,6 @@ out_reqbuf:
|
|
|
spin_unlock(&buffers->rb_lock);
|
|
|
pr_warn("RPC: %s: out of request buffers\n", __func__);
|
|
|
return NULL;
|
|
|
-out_repbuf:
|
|
|
- spin_unlock(&buffers->rb_lock);
|
|
|
- pr_warn("RPC: %s: out of reply buffers\n", __func__);
|
|
|
- req->rl_reply = NULL;
|
|
|
- return req;
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -1119,9 +1133,12 @@ rpcrdma_buffer_put(struct rpcrdma_req *req)
|
|
|
req->rl_reply = NULL;
|
|
|
|
|
|
spin_lock(&buffers->rb_lock);
|
|
|
+ buffers->rb_send_count--;
|
|
|
list_add_tail(&req->rl_free, &buffers->rb_send_bufs);
|
|
|
- if (rep)
|
|
|
+ if (rep) {
|
|
|
+ buffers->rb_recv_count--;
|
|
|
list_add_tail(&rep->rr_list, &buffers->rb_recv_bufs);
|
|
|
+ }
|
|
|
spin_unlock(&buffers->rb_lock);
|
|
|
}
|
|
|
|
|
@@ -1135,8 +1152,7 @@ rpcrdma_recv_buffer_get(struct rpcrdma_req *req)
|
|
|
struct rpcrdma_buffer *buffers = req->rl_buffer;
|
|
|
|
|
|
spin_lock(&buffers->rb_lock);
|
|
|
- if (!list_empty(&buffers->rb_recv_bufs))
|
|
|
- req->rl_reply = rpcrdma_buffer_get_rep_locked(buffers);
|
|
|
+ req->rl_reply = rpcrdma_buffer_get_rep(buffers);
|
|
|
spin_unlock(&buffers->rb_lock);
|
|
|
}
|
|
|
|
|
@@ -1150,6 +1166,7 @@ rpcrdma_recv_buffer_put(struct rpcrdma_rep *rep)
|
|
|
struct rpcrdma_buffer *buffers = &rep->rr_rxprt->rx_buf;
|
|
|
|
|
|
spin_lock(&buffers->rb_lock);
|
|
|
+ buffers->rb_recv_count--;
|
|
|
list_add_tail(&rep->rr_list, &buffers->rb_recv_bufs);
|
|
|
spin_unlock(&buffers->rb_lock);
|
|
|
}
|