|
@@ -1910,8 +1910,7 @@ nvmet_fc_transfer_fcp_data(struct nvmet_fc_tgtport *tgtport,
|
|
spin_lock_irqsave(&fod->flock, flags);
|
|
spin_lock_irqsave(&fod->flock, flags);
|
|
fod->writedataactive = false;
|
|
fod->writedataactive = false;
|
|
spin_unlock_irqrestore(&fod->flock, flags);
|
|
spin_unlock_irqrestore(&fod->flock, flags);
|
|
- nvmet_req_complete(&fod->req,
|
|
|
|
- NVME_SC_FC_TRANSPORT_ERROR);
|
|
|
|
|
|
+ nvmet_req_complete(&fod->req, NVME_SC_INTERNAL);
|
|
} else /* NVMET_FCOP_READDATA or NVMET_FCOP_READDATA_RSP */ {
|
|
} else /* NVMET_FCOP_READDATA or NVMET_FCOP_READDATA_RSP */ {
|
|
fcpreq->fcp_error = ret;
|
|
fcpreq->fcp_error = ret;
|
|
fcpreq->transferred_length = 0;
|
|
fcpreq->transferred_length = 0;
|
|
@@ -1929,8 +1928,7 @@ __nvmet_fc_fod_op_abort(struct nvmet_fc_fcp_iod *fod, bool abort)
|
|
/* if in the middle of an io and we need to tear down */
|
|
/* if in the middle of an io and we need to tear down */
|
|
if (abort) {
|
|
if (abort) {
|
|
if (fcpreq->op == NVMET_FCOP_WRITEDATA) {
|
|
if (fcpreq->op == NVMET_FCOP_WRITEDATA) {
|
|
- nvmet_req_complete(&fod->req,
|
|
|
|
- NVME_SC_FC_TRANSPORT_ERROR);
|
|
|
|
|
|
+ nvmet_req_complete(&fod->req, NVME_SC_INTERNAL);
|
|
return true;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -1968,8 +1966,7 @@ nvmet_fc_fod_op_done(struct nvmet_fc_fcp_iod *fod)
|
|
fod->abort = true;
|
|
fod->abort = true;
|
|
spin_unlock(&fod->flock);
|
|
spin_unlock(&fod->flock);
|
|
|
|
|
|
- nvmet_req_complete(&fod->req,
|
|
|
|
- NVME_SC_FC_TRANSPORT_ERROR);
|
|
|
|
|
|
+ nvmet_req_complete(&fod->req, NVME_SC_INTERNAL);
|
|
return;
|
|
return;
|
|
}
|
|
}
|
|
|
|
|