|
@@ -2090,7 +2090,8 @@ int smbd_recv(struct smbd_connection *info, struct msghdr *msg)
|
|
|
* rqst: the data to write
|
|
* rqst: the data to write
|
|
|
* return value: 0 if successfully write, otherwise error code
|
|
* return value: 0 if successfully write, otherwise error code
|
|
|
*/
|
|
*/
|
|
|
-int smbd_send(struct TCP_Server_Info *server, struct smb_rqst *rqst)
|
|
|
|
|
|
|
+int smbd_send(struct TCP_Server_Info *server,
|
|
|
|
|
+ int num_rqst, struct smb_rqst *rqst_array)
|
|
|
{
|
|
{
|
|
|
struct smbd_connection *info = server->smbd_conn;
|
|
struct smbd_connection *info = server->smbd_conn;
|
|
|
struct kvec vec;
|
|
struct kvec vec;
|
|
@@ -2102,6 +2103,8 @@ int smbd_send(struct TCP_Server_Info *server, struct smb_rqst *rqst)
|
|
|
info->max_send_size - sizeof(struct smbd_data_transfer);
|
|
info->max_send_size - sizeof(struct smbd_data_transfer);
|
|
|
struct kvec *iov;
|
|
struct kvec *iov;
|
|
|
int rc;
|
|
int rc;
|
|
|
|
|
+ struct smb_rqst *rqst;
|
|
|
|
|
+ int rqst_idx;
|
|
|
|
|
|
|
|
info->smbd_send_pending++;
|
|
info->smbd_send_pending++;
|
|
|
if (info->transport_status != SMBD_CONNECTED) {
|
|
if (info->transport_status != SMBD_CONNECTED) {
|
|
@@ -2109,47 +2112,41 @@ int smbd_send(struct TCP_Server_Info *server, struct smb_rqst *rqst)
|
|
|
goto done;
|
|
goto done;
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
- /*
|
|
|
|
|
- * Skip the RFC1002 length defined in MS-SMB2 section 2.1
|
|
|
|
|
- * It is used only for TCP transport in the iov[0]
|
|
|
|
|
- * In future we may want to add a transport layer under protocol
|
|
|
|
|
- * layer so this will only be issued to TCP transport
|
|
|
|
|
- */
|
|
|
|
|
-
|
|
|
|
|
- if (rqst->rq_iov[0].iov_len != 4) {
|
|
|
|
|
- log_write(ERR, "expected the pdu length in 1st iov, but got %zu\n", rqst->rq_iov[0].iov_len);
|
|
|
|
|
- return -EINVAL;
|
|
|
|
|
- }
|
|
|
|
|
-
|
|
|
|
|
/*
|
|
/*
|
|
|
* Add in the page array if there is one. The caller needs to set
|
|
* Add in the page array if there is one. The caller needs to set
|
|
|
* rq_tailsz to PAGE_SIZE when the buffer has multiple pages and
|
|
* rq_tailsz to PAGE_SIZE when the buffer has multiple pages and
|
|
|
* ends at page boundary
|
|
* ends at page boundary
|
|
|
*/
|
|
*/
|
|
|
- buflen = smb_rqst_len(server, rqst);
|
|
|
|
|
|
|
+ remaining_data_length = 0;
|
|
|
|
|
+ for (i = 0; i < num_rqst; i++)
|
|
|
|
|
+ remaining_data_length += smb_rqst_len(server, &rqst_array[i]);
|
|
|
|
|
|
|
|
- if (buflen + sizeof(struct smbd_data_transfer) >
|
|
|
|
|
|
|
+ if (remaining_data_length + sizeof(struct smbd_data_transfer) >
|
|
|
info->max_fragmented_send_size) {
|
|
info->max_fragmented_send_size) {
|
|
|
log_write(ERR, "payload size %d > max size %d\n",
|
|
log_write(ERR, "payload size %d > max size %d\n",
|
|
|
- buflen, info->max_fragmented_send_size);
|
|
|
|
|
|
|
+ remaining_data_length, info->max_fragmented_send_size);
|
|
|
rc = -EINVAL;
|
|
rc = -EINVAL;
|
|
|
goto done;
|
|
goto done;
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
- iov = &rqst->rq_iov[1];
|
|
|
|
|
|
|
+ rqst_idx = 0;
|
|
|
|
|
+
|
|
|
|
|
+next_rqst:
|
|
|
|
|
+ rqst = &rqst_array[rqst_idx];
|
|
|
|
|
+ iov = rqst->rq_iov;
|
|
|
|
|
|
|
|
- cifs_dbg(FYI, "Sending smb (RDMA): smb_len=%u\n", buflen);
|
|
|
|
|
- for (i = 0; i < rqst->rq_nvec-1; i++)
|
|
|
|
|
|
|
+ cifs_dbg(FYI, "Sending smb (RDMA): idx=%d smb_len=%lu\n",
|
|
|
|
|
+ rqst_idx, smb_rqst_len(server, rqst));
|
|
|
|
|
+ for (i = 0; i < rqst->rq_nvec; i++)
|
|
|
dump_smb(iov[i].iov_base, iov[i].iov_len);
|
|
dump_smb(iov[i].iov_base, iov[i].iov_len);
|
|
|
|
|
|
|
|
- remaining_data_length = buflen;
|
|
|
|
|
|
|
|
|
|
- log_write(INFO, "rqst->rq_nvec=%d rqst->rq_npages=%d rq_pagesz=%d "
|
|
|
|
|
- "rq_tailsz=%d buflen=%d\n",
|
|
|
|
|
- rqst->rq_nvec, rqst->rq_npages, rqst->rq_pagesz,
|
|
|
|
|
- rqst->rq_tailsz, buflen);
|
|
|
|
|
|
|
+ log_write(INFO, "rqst_idx=%d nvec=%d rqst->rq_npages=%d rq_pagesz=%d "
|
|
|
|
|
+ "rq_tailsz=%d buflen=%lu\n",
|
|
|
|
|
+ rqst_idx, rqst->rq_nvec, rqst->rq_npages, rqst->rq_pagesz,
|
|
|
|
|
+ rqst->rq_tailsz, smb_rqst_len(server, rqst));
|
|
|
|
|
|
|
|
- start = i = iov[0].iov_len ? 0 : 1;
|
|
|
|
|
|
|
+ start = i = 0;
|
|
|
buflen = 0;
|
|
buflen = 0;
|
|
|
while (true) {
|
|
while (true) {
|
|
|
buflen += iov[i].iov_len;
|
|
buflen += iov[i].iov_len;
|
|
@@ -2197,14 +2194,14 @@ int smbd_send(struct TCP_Server_Info *server, struct smb_rqst *rqst)
|
|
|
goto done;
|
|
goto done;
|
|
|
}
|
|
}
|
|
|
i++;
|
|
i++;
|
|
|
- if (i == rqst->rq_nvec-1)
|
|
|
|
|
|
|
+ if (i == rqst->rq_nvec)
|
|
|
break;
|
|
break;
|
|
|
}
|
|
}
|
|
|
start = i;
|
|
start = i;
|
|
|
buflen = 0;
|
|
buflen = 0;
|
|
|
} else {
|
|
} else {
|
|
|
i++;
|
|
i++;
|
|
|
- if (i == rqst->rq_nvec-1) {
|
|
|
|
|
|
|
+ if (i == rqst->rq_nvec) {
|
|
|
/* send out all remaining vecs */
|
|
/* send out all remaining vecs */
|
|
|
remaining_data_length -= buflen;
|
|
remaining_data_length -= buflen;
|
|
|
log_write(INFO,
|
|
log_write(INFO,
|
|
@@ -2248,6 +2245,10 @@ int smbd_send(struct TCP_Server_Info *server, struct smb_rqst *rqst)
|
|
|
}
|
|
}
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
+ rqst_idx++;
|
|
|
|
|
+ if (rqst_idx < num_rqst)
|
|
|
|
|
+ goto next_rqst;
|
|
|
|
|
+
|
|
|
done:
|
|
done:
|
|
|
/*
|
|
/*
|
|
|
* As an optimization, we don't wait for individual I/O to finish
|
|
* As an optimization, we don't wait for individual I/O to finish
|