|
@@ -38,7 +38,6 @@
|
|
|
#include <linux/miscdevice.h>
|
|
|
#include <asm/unaligned.h>
|
|
|
#include <scsi/scsi.h>
|
|
|
-#include <scsi/scsi_tcq.h>
|
|
|
#include <target/target_core_base.h>
|
|
|
#include <target/target_core_fabric.h>
|
|
|
#include <target/target_core_fabric_configfs.h>
|
|
@@ -52,13 +51,13 @@
|
|
|
|
|
|
#include "vhost.h"
|
|
|
|
|
|
-#define TCM_VHOST_VERSION "v0.1"
|
|
|
-#define TCM_VHOST_NAMELEN 256
|
|
|
-#define TCM_VHOST_MAX_CDB_SIZE 32
|
|
|
-#define TCM_VHOST_DEFAULT_TAGS 256
|
|
|
-#define TCM_VHOST_PREALLOC_SGLS 2048
|
|
|
-#define TCM_VHOST_PREALLOC_UPAGES 2048
|
|
|
-#define TCM_VHOST_PREALLOC_PROT_SGLS 512
|
|
|
+#define VHOST_SCSI_VERSION "v0.1"
|
|
|
+#define VHOST_SCSI_NAMELEN 256
|
|
|
+#define VHOST_SCSI_MAX_CDB_SIZE 32
|
|
|
+#define VHOST_SCSI_DEFAULT_TAGS 256
|
|
|
+#define VHOST_SCSI_PREALLOC_SGLS 2048
|
|
|
+#define VHOST_SCSI_PREALLOC_UPAGES 2048
|
|
|
+#define VHOST_SCSI_PREALLOC_PROT_SGLS 512
|
|
|
|
|
|
struct vhost_scsi_inflight {
|
|
|
/* Wait for the flush operation to finish */
|
|
@@ -67,11 +66,13 @@ struct vhost_scsi_inflight {
|
|
|
struct kref kref;
|
|
|
};
|
|
|
|
|
|
-struct tcm_vhost_cmd {
|
|
|
+struct vhost_scsi_cmd {
|
|
|
/* Descriptor from vhost_get_vq_desc() for virt_queue segment */
|
|
|
int tvc_vq_desc;
|
|
|
/* virtio-scsi initiator task attribute */
|
|
|
int tvc_task_attr;
|
|
|
+ /* virtio-scsi response incoming iovecs */
|
|
|
+ int tvc_in_iovs;
|
|
|
/* virtio-scsi initiator data direction */
|
|
|
enum dma_data_direction tvc_data_direction;
|
|
|
/* Expected data transfer length from virtio-scsi header */
|
|
@@ -81,26 +82,26 @@ struct tcm_vhost_cmd {
|
|
|
/* The number of scatterlists associated with this cmd */
|
|
|
u32 tvc_sgl_count;
|
|
|
u32 tvc_prot_sgl_count;
|
|
|
- /* Saved unpacked SCSI LUN for tcm_vhost_submission_work() */
|
|
|
+ /* Saved unpacked SCSI LUN for vhost_scsi_submission_work() */
|
|
|
u32 tvc_lun;
|
|
|
/* Pointer to the SGL formatted memory from virtio-scsi */
|
|
|
struct scatterlist *tvc_sgl;
|
|
|
struct scatterlist *tvc_prot_sgl;
|
|
|
struct page **tvc_upages;
|
|
|
- /* Pointer to response */
|
|
|
- struct virtio_scsi_cmd_resp __user *tvc_resp;
|
|
|
+ /* Pointer to response header iovec */
|
|
|
+ struct iovec *tvc_resp_iov;
|
|
|
/* Pointer to vhost_scsi for our device */
|
|
|
struct vhost_scsi *tvc_vhost;
|
|
|
/* Pointer to vhost_virtqueue for the cmd */
|
|
|
struct vhost_virtqueue *tvc_vq;
|
|
|
/* Pointer to vhost nexus memory */
|
|
|
- struct tcm_vhost_nexus *tvc_nexus;
|
|
|
+ struct vhost_scsi_nexus *tvc_nexus;
|
|
|
/* The TCM I/O descriptor that is accessed via container_of() */
|
|
|
struct se_cmd tvc_se_cmd;
|
|
|
- /* work item used for cmwq dispatch to tcm_vhost_submission_work() */
|
|
|
+ /* work item used for cmwq dispatch to vhost_scsi_submission_work() */
|
|
|
struct work_struct work;
|
|
|
/* Copy of the incoming SCSI command descriptor block (CDB) */
|
|
|
- unsigned char tvc_cdb[TCM_VHOST_MAX_CDB_SIZE];
|
|
|
+ unsigned char tvc_cdb[VHOST_SCSI_MAX_CDB_SIZE];
|
|
|
/* Sense buffer that will be mapped into outgoing status */
|
|
|
unsigned char tvc_sense_buf[TRANSPORT_SENSE_BUFFER];
|
|
|
/* Completed commands list, serviced from vhost worker thread */
|
|
@@ -109,53 +110,53 @@ struct tcm_vhost_cmd {
|
|
|
struct vhost_scsi_inflight *inflight;
|
|
|
};
|
|
|
|
|
|
-struct tcm_vhost_nexus {
|
|
|
+struct vhost_scsi_nexus {
|
|
|
/* Pointer to TCM session for I_T Nexus */
|
|
|
struct se_session *tvn_se_sess;
|
|
|
};
|
|
|
|
|
|
-struct tcm_vhost_nacl {
|
|
|
+struct vhost_scsi_nacl {
|
|
|
/* Binary World Wide unique Port Name for Vhost Initiator port */
|
|
|
u64 iport_wwpn;
|
|
|
/* ASCII formatted WWPN for Sas Initiator port */
|
|
|
- char iport_name[TCM_VHOST_NAMELEN];
|
|
|
- /* Returned by tcm_vhost_make_nodeacl() */
|
|
|
+ char iport_name[VHOST_SCSI_NAMELEN];
|
|
|
+ /* Returned by vhost_scsi_make_nodeacl() */
|
|
|
struct se_node_acl se_node_acl;
|
|
|
};
|
|
|
|
|
|
-struct tcm_vhost_tpg {
|
|
|
+struct vhost_scsi_tpg {
|
|
|
/* Vhost port target portal group tag for TCM */
|
|
|
u16 tport_tpgt;
|
|
|
/* Used to track number of TPG Port/Lun Links wrt to explict I_T Nexus shutdown */
|
|
|
int tv_tpg_port_count;
|
|
|
/* Used for vhost_scsi device reference to tpg_nexus, protected by tv_tpg_mutex */
|
|
|
int tv_tpg_vhost_count;
|
|
|
- /* list for tcm_vhost_list */
|
|
|
+ /* list for vhost_scsi_list */
|
|
|
struct list_head tv_tpg_list;
|
|
|
/* Used to protect access for tpg_nexus */
|
|
|
struct mutex tv_tpg_mutex;
|
|
|
/* Pointer to the TCM VHost I_T Nexus for this TPG endpoint */
|
|
|
- struct tcm_vhost_nexus *tpg_nexus;
|
|
|
- /* Pointer back to tcm_vhost_tport */
|
|
|
- struct tcm_vhost_tport *tport;
|
|
|
- /* Returned by tcm_vhost_make_tpg() */
|
|
|
+ struct vhost_scsi_nexus *tpg_nexus;
|
|
|
+ /* Pointer back to vhost_scsi_tport */
|
|
|
+ struct vhost_scsi_tport *tport;
|
|
|
+ /* Returned by vhost_scsi_make_tpg() */
|
|
|
struct se_portal_group se_tpg;
|
|
|
/* Pointer back to vhost_scsi, protected by tv_tpg_mutex */
|
|
|
struct vhost_scsi *vhost_scsi;
|
|
|
};
|
|
|
|
|
|
-struct tcm_vhost_tport {
|
|
|
+struct vhost_scsi_tport {
|
|
|
/* SCSI protocol the tport is providing */
|
|
|
u8 tport_proto_id;
|
|
|
/* Binary World Wide unique Port Name for Vhost Target port */
|
|
|
u64 tport_wwpn;
|
|
|
/* ASCII formatted WWPN for Vhost Target port */
|
|
|
- char tport_name[TCM_VHOST_NAMELEN];
|
|
|
- /* Returned by tcm_vhost_make_tport() */
|
|
|
+ char tport_name[VHOST_SCSI_NAMELEN];
|
|
|
+ /* Returned by vhost_scsi_make_tport() */
|
|
|
struct se_wwn tport_wwn;
|
|
|
};
|
|
|
|
|
|
-struct tcm_vhost_evt {
|
|
|
+struct vhost_scsi_evt {
|
|
|
/* event to be sent to guest */
|
|
|
struct virtio_scsi_event event;
|
|
|
/* event list, serviced from vhost worker thread */
|
|
@@ -171,7 +172,9 @@ enum {
|
|
|
/* Note: can't set VIRTIO_F_VERSION_1 yet, since that implies ANY_LAYOUT. */
|
|
|
enum {
|
|
|
VHOST_SCSI_FEATURES = VHOST_FEATURES | (1ULL << VIRTIO_SCSI_F_HOTPLUG) |
|
|
|
- (1ULL << VIRTIO_SCSI_F_T10_PI)
|
|
|
+ (1ULL << VIRTIO_SCSI_F_T10_PI) |
|
|
|
+ (1ULL << VIRTIO_F_ANY_LAYOUT) |
|
|
|
+ (1ULL << VIRTIO_F_VERSION_1)
|
|
|
};
|
|
|
|
|
|
#define VHOST_SCSI_MAX_TARGET 256
|
|
@@ -195,7 +198,7 @@ struct vhost_scsi_virtqueue {
|
|
|
|
|
|
struct vhost_scsi {
|
|
|
/* Protected by vhost_scsi->dev.mutex */
|
|
|
- struct tcm_vhost_tpg **vs_tpg;
|
|
|
+ struct vhost_scsi_tpg **vs_tpg;
|
|
|
char vs_vhost_wwpn[TRANSPORT_IQN_LEN];
|
|
|
|
|
|
struct vhost_dev dev;
|
|
@@ -212,21 +215,21 @@ struct vhost_scsi {
|
|
|
};
|
|
|
|
|
|
/* Local pointer to allocated TCM configfs fabric module */
|
|
|
-static struct target_fabric_configfs *tcm_vhost_fabric_configfs;
|
|
|
+static struct target_fabric_configfs *vhost_scsi_fabric_configfs;
|
|
|
|
|
|
-static struct workqueue_struct *tcm_vhost_workqueue;
|
|
|
+static struct workqueue_struct *vhost_scsi_workqueue;
|
|
|
|
|
|
-/* Global spinlock to protect tcm_vhost TPG list for vhost IOCTL access */
|
|
|
-static DEFINE_MUTEX(tcm_vhost_mutex);
|
|
|
-static LIST_HEAD(tcm_vhost_list);
|
|
|
+/* Global spinlock to protect vhost_scsi TPG list for vhost IOCTL access */
|
|
|
+static DEFINE_MUTEX(vhost_scsi_mutex);
|
|
|
+static LIST_HEAD(vhost_scsi_list);
|
|
|
|
|
|
-static int iov_num_pages(struct iovec *iov)
|
|
|
+static int iov_num_pages(void __user *iov_base, size_t iov_len)
|
|
|
{
|
|
|
- return (PAGE_ALIGN((unsigned long)iov->iov_base + iov->iov_len) -
|
|
|
- ((unsigned long)iov->iov_base & PAGE_MASK)) >> PAGE_SHIFT;
|
|
|
+ return (PAGE_ALIGN((unsigned long)iov_base + iov_len) -
|
|
|
+ ((unsigned long)iov_base & PAGE_MASK)) >> PAGE_SHIFT;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_done_inflight(struct kref *kref)
|
|
|
+static void vhost_scsi_done_inflight(struct kref *kref)
|
|
|
{
|
|
|
struct vhost_scsi_inflight *inflight;
|
|
|
|
|
@@ -234,7 +237,7 @@ static void tcm_vhost_done_inflight(struct kref *kref)
|
|
|
complete(&inflight->comp);
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_init_inflight(struct vhost_scsi *vs,
|
|
|
+static void vhost_scsi_init_inflight(struct vhost_scsi *vs,
|
|
|
struct vhost_scsi_inflight *old_inflight[])
|
|
|
{
|
|
|
struct vhost_scsi_inflight *new_inflight;
|
|
@@ -262,7 +265,7 @@ static void tcm_vhost_init_inflight(struct vhost_scsi *vs,
|
|
|
}
|
|
|
|
|
|
static struct vhost_scsi_inflight *
|
|
|
-tcm_vhost_get_inflight(struct vhost_virtqueue *vq)
|
|
|
+vhost_scsi_get_inflight(struct vhost_virtqueue *vq)
|
|
|
{
|
|
|
struct vhost_scsi_inflight *inflight;
|
|
|
struct vhost_scsi_virtqueue *svq;
|
|
@@ -274,31 +277,31 @@ tcm_vhost_get_inflight(struct vhost_virtqueue *vq)
|
|
|
return inflight;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_put_inflight(struct vhost_scsi_inflight *inflight)
|
|
|
+static void vhost_scsi_put_inflight(struct vhost_scsi_inflight *inflight)
|
|
|
{
|
|
|
- kref_put(&inflight->kref, tcm_vhost_done_inflight);
|
|
|
+ kref_put(&inflight->kref, vhost_scsi_done_inflight);
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_check_true(struct se_portal_group *se_tpg)
|
|
|
+static int vhost_scsi_check_true(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
return 1;
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_check_false(struct se_portal_group *se_tpg)
|
|
|
+static int vhost_scsi_check_false(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static char *tcm_vhost_get_fabric_name(void)
|
|
|
+static char *vhost_scsi_get_fabric_name(void)
|
|
|
{
|
|
|
return "vhost";
|
|
|
}
|
|
|
|
|
|
-static u8 tcm_vhost_get_fabric_proto_ident(struct se_portal_group *se_tpg)
|
|
|
+static u8 vhost_scsi_get_fabric_proto_ident(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_tport *tport = tpg->tport;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tport *tport = tpg->tport;
|
|
|
|
|
|
switch (tport->tport_proto_id) {
|
|
|
case SCSI_PROTOCOL_SAS:
|
|
@@ -316,37 +319,37 @@ static u8 tcm_vhost_get_fabric_proto_ident(struct se_portal_group *se_tpg)
|
|
|
return sas_get_fabric_proto_ident(se_tpg);
|
|
|
}
|
|
|
|
|
|
-static char *tcm_vhost_get_fabric_wwn(struct se_portal_group *se_tpg)
|
|
|
+static char *vhost_scsi_get_fabric_wwn(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_tport *tport = tpg->tport;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tport *tport = tpg->tport;
|
|
|
|
|
|
return &tport->tport_name[0];
|
|
|
}
|
|
|
|
|
|
-static u16 tcm_vhost_get_tag(struct se_portal_group *se_tpg)
|
|
|
+static u16 vhost_scsi_get_tpgt(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
return tpg->tport_tpgt;
|
|
|
}
|
|
|
|
|
|
-static u32 tcm_vhost_get_default_depth(struct se_portal_group *se_tpg)
|
|
|
+static u32 vhost_scsi_get_default_depth(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
return 1;
|
|
|
}
|
|
|
|
|
|
static u32
|
|
|
-tcm_vhost_get_pr_transport_id(struct se_portal_group *se_tpg,
|
|
|
+vhost_scsi_get_pr_transport_id(struct se_portal_group *se_tpg,
|
|
|
struct se_node_acl *se_nacl,
|
|
|
struct t10_pr_registration *pr_reg,
|
|
|
int *format_code,
|
|
|
unsigned char *buf)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_tport *tport = tpg->tport;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tport *tport = tpg->tport;
|
|
|
|
|
|
switch (tport->tport_proto_id) {
|
|
|
case SCSI_PROTOCOL_SAS:
|
|
@@ -369,14 +372,14 @@ tcm_vhost_get_pr_transport_id(struct se_portal_group *se_tpg,
|
|
|
}
|
|
|
|
|
|
static u32
|
|
|
-tcm_vhost_get_pr_transport_id_len(struct se_portal_group *se_tpg,
|
|
|
+vhost_scsi_get_pr_transport_id_len(struct se_portal_group *se_tpg,
|
|
|
struct se_node_acl *se_nacl,
|
|
|
struct t10_pr_registration *pr_reg,
|
|
|
int *format_code)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_tport *tport = tpg->tport;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tport *tport = tpg->tport;
|
|
|
|
|
|
switch (tport->tport_proto_id) {
|
|
|
case SCSI_PROTOCOL_SAS:
|
|
@@ -399,14 +402,14 @@ tcm_vhost_get_pr_transport_id_len(struct se_portal_group *se_tpg,
|
|
|
}
|
|
|
|
|
|
static char *
|
|
|
-tcm_vhost_parse_pr_out_transport_id(struct se_portal_group *se_tpg,
|
|
|
+vhost_scsi_parse_pr_out_transport_id(struct se_portal_group *se_tpg,
|
|
|
const char *buf,
|
|
|
u32 *out_tid_len,
|
|
|
char **port_nexus_ptr)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_tport *tport = tpg->tport;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tport *tport = tpg->tport;
|
|
|
|
|
|
switch (tport->tport_proto_id) {
|
|
|
case SCSI_PROTOCOL_SAS:
|
|
@@ -429,13 +432,13 @@ tcm_vhost_parse_pr_out_transport_id(struct se_portal_group *se_tpg,
|
|
|
}
|
|
|
|
|
|
static struct se_node_acl *
|
|
|
-tcm_vhost_alloc_fabric_acl(struct se_portal_group *se_tpg)
|
|
|
+vhost_scsi_alloc_fabric_acl(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
- struct tcm_vhost_nacl *nacl;
|
|
|
+ struct vhost_scsi_nacl *nacl;
|
|
|
|
|
|
- nacl = kzalloc(sizeof(struct tcm_vhost_nacl), GFP_KERNEL);
|
|
|
+ nacl = kzalloc(sizeof(struct vhost_scsi_nacl), GFP_KERNEL);
|
|
|
if (!nacl) {
|
|
|
- pr_err("Unable to allocate struct tcm_vhost_nacl\n");
|
|
|
+ pr_err("Unable to allocate struct vhost_scsi_nacl\n");
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
@@ -443,24 +446,24 @@ tcm_vhost_alloc_fabric_acl(struct se_portal_group *se_tpg)
|
|
|
}
|
|
|
|
|
|
static void
|
|
|
-tcm_vhost_release_fabric_acl(struct se_portal_group *se_tpg,
|
|
|
+vhost_scsi_release_fabric_acl(struct se_portal_group *se_tpg,
|
|
|
struct se_node_acl *se_nacl)
|
|
|
{
|
|
|
- struct tcm_vhost_nacl *nacl = container_of(se_nacl,
|
|
|
- struct tcm_vhost_nacl, se_node_acl);
|
|
|
+ struct vhost_scsi_nacl *nacl = container_of(se_nacl,
|
|
|
+ struct vhost_scsi_nacl, se_node_acl);
|
|
|
kfree(nacl);
|
|
|
}
|
|
|
|
|
|
-static u32 tcm_vhost_tpg_get_inst_index(struct se_portal_group *se_tpg)
|
|
|
+static u32 vhost_scsi_tpg_get_inst_index(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
return 1;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_release_cmd(struct se_cmd *se_cmd)
|
|
|
+static void vhost_scsi_release_cmd(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
- struct tcm_vhost_cmd *tv_cmd = container_of(se_cmd,
|
|
|
- struct tcm_vhost_cmd, tvc_se_cmd);
|
|
|
- struct se_session *se_sess = se_cmd->se_sess;
|
|
|
+ struct vhost_scsi_cmd *tv_cmd = container_of(se_cmd,
|
|
|
+ struct vhost_scsi_cmd, tvc_se_cmd);
|
|
|
+ struct se_session *se_sess = tv_cmd->tvc_nexus->tvn_se_sess;
|
|
|
int i;
|
|
|
|
|
|
if (tv_cmd->tvc_sgl_count) {
|
|
@@ -472,53 +475,53 @@ static void tcm_vhost_release_cmd(struct se_cmd *se_cmd)
|
|
|
put_page(sg_page(&tv_cmd->tvc_prot_sgl[i]));
|
|
|
}
|
|
|
|
|
|
- tcm_vhost_put_inflight(tv_cmd->inflight);
|
|
|
+ vhost_scsi_put_inflight(tv_cmd->inflight);
|
|
|
percpu_ida_free(&se_sess->sess_tag_pool, se_cmd->map_tag);
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_shutdown_session(struct se_session *se_sess)
|
|
|
+static int vhost_scsi_shutdown_session(struct se_session *se_sess)
|
|
|
{
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_close_session(struct se_session *se_sess)
|
|
|
+static void vhost_scsi_close_session(struct se_session *se_sess)
|
|
|
{
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
-static u32 tcm_vhost_sess_get_index(struct se_session *se_sess)
|
|
|
+static u32 vhost_scsi_sess_get_index(struct se_session *se_sess)
|
|
|
{
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_write_pending(struct se_cmd *se_cmd)
|
|
|
+static int vhost_scsi_write_pending(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
/* Go ahead and process the write immediately */
|
|
|
target_execute_cmd(se_cmd);
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_write_pending_status(struct se_cmd *se_cmd)
|
|
|
+static int vhost_scsi_write_pending_status(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_set_default_node_attrs(struct se_node_acl *nacl)
|
|
|
+static void vhost_scsi_set_default_node_attrs(struct se_node_acl *nacl)
|
|
|
{
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
-static u32 tcm_vhost_get_task_tag(struct se_cmd *se_cmd)
|
|
|
+static u32 vhost_scsi_get_task_tag(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_get_cmd_state(struct se_cmd *se_cmd)
|
|
|
+static int vhost_scsi_get_cmd_state(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void vhost_scsi_complete_cmd(struct tcm_vhost_cmd *cmd)
|
|
|
+static void vhost_scsi_complete_cmd(struct vhost_scsi_cmd *cmd)
|
|
|
{
|
|
|
struct vhost_scsi *vs = cmd->tvc_vhost;
|
|
|
|
|
@@ -527,44 +530,44 @@ static void vhost_scsi_complete_cmd(struct tcm_vhost_cmd *cmd)
|
|
|
vhost_work_queue(&vs->dev, &vs->vs_completion_work);
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_queue_data_in(struct se_cmd *se_cmd)
|
|
|
+static int vhost_scsi_queue_data_in(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
- struct tcm_vhost_cmd *cmd = container_of(se_cmd,
|
|
|
- struct tcm_vhost_cmd, tvc_se_cmd);
|
|
|
+ struct vhost_scsi_cmd *cmd = container_of(se_cmd,
|
|
|
+ struct vhost_scsi_cmd, tvc_se_cmd);
|
|
|
vhost_scsi_complete_cmd(cmd);
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_queue_status(struct se_cmd *se_cmd)
|
|
|
+static int vhost_scsi_queue_status(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
- struct tcm_vhost_cmd *cmd = container_of(se_cmd,
|
|
|
- struct tcm_vhost_cmd, tvc_se_cmd);
|
|
|
+ struct vhost_scsi_cmd *cmd = container_of(se_cmd,
|
|
|
+ struct vhost_scsi_cmd, tvc_se_cmd);
|
|
|
vhost_scsi_complete_cmd(cmd);
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_queue_tm_rsp(struct se_cmd *se_cmd)
|
|
|
+static void vhost_scsi_queue_tm_rsp(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_aborted_task(struct se_cmd *se_cmd)
|
|
|
+static void vhost_scsi_aborted_task(struct se_cmd *se_cmd)
|
|
|
{
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_free_evt(struct vhost_scsi *vs, struct tcm_vhost_evt *evt)
|
|
|
+static void vhost_scsi_free_evt(struct vhost_scsi *vs, struct vhost_scsi_evt *evt)
|
|
|
{
|
|
|
vs->vs_events_nr--;
|
|
|
kfree(evt);
|
|
|
}
|
|
|
|
|
|
-static struct tcm_vhost_evt *
|
|
|
-tcm_vhost_allocate_evt(struct vhost_scsi *vs,
|
|
|
+static struct vhost_scsi_evt *
|
|
|
+vhost_scsi_allocate_evt(struct vhost_scsi *vs,
|
|
|
u32 event, u32 reason)
|
|
|
{
|
|
|
struct vhost_virtqueue *vq = &vs->vqs[VHOST_SCSI_VQ_EVT].vq;
|
|
|
- struct tcm_vhost_evt *evt;
|
|
|
+ struct vhost_scsi_evt *evt;
|
|
|
|
|
|
if (vs->vs_events_nr > VHOST_SCSI_MAX_EVENT) {
|
|
|
vs->vs_events_missed = true;
|
|
@@ -573,7 +576,7 @@ tcm_vhost_allocate_evt(struct vhost_scsi *vs,
|
|
|
|
|
|
evt = kzalloc(sizeof(*evt), GFP_KERNEL);
|
|
|
if (!evt) {
|
|
|
- vq_err(vq, "Failed to allocate tcm_vhost_evt\n");
|
|
|
+ vq_err(vq, "Failed to allocate vhost_scsi_evt\n");
|
|
|
vs->vs_events_missed = true;
|
|
|
return NULL;
|
|
|
}
|
|
@@ -585,7 +588,7 @@ tcm_vhost_allocate_evt(struct vhost_scsi *vs,
|
|
|
return evt;
|
|
|
}
|
|
|
|
|
|
-static void vhost_scsi_free_cmd(struct tcm_vhost_cmd *cmd)
|
|
|
+static void vhost_scsi_free_cmd(struct vhost_scsi_cmd *cmd)
|
|
|
{
|
|
|
struct se_cmd *se_cmd = &cmd->tvc_se_cmd;
|
|
|
|
|
@@ -600,7 +603,7 @@ static int vhost_scsi_check_stop_free(struct se_cmd *se_cmd)
|
|
|
}
|
|
|
|
|
|
static void
|
|
|
-tcm_vhost_do_evt_work(struct vhost_scsi *vs, struct tcm_vhost_evt *evt)
|
|
|
+vhost_scsi_do_evt_work(struct vhost_scsi *vs, struct vhost_scsi_evt *evt)
|
|
|
{
|
|
|
struct vhost_virtqueue *vq = &vs->vqs[VHOST_SCSI_VQ_EVT].vq;
|
|
|
struct virtio_scsi_event *event = &evt->event;
|
|
@@ -646,24 +649,24 @@ again:
|
|
|
if (!ret)
|
|
|
vhost_add_used_and_signal(&vs->dev, vq, head, 0);
|
|
|
else
|
|
|
- vq_err(vq, "Faulted on tcm_vhost_send_event\n");
|
|
|
+ vq_err(vq, "Faulted on vhost_scsi_send_event\n");
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_evt_work(struct vhost_work *work)
|
|
|
+static void vhost_scsi_evt_work(struct vhost_work *work)
|
|
|
{
|
|
|
struct vhost_scsi *vs = container_of(work, struct vhost_scsi,
|
|
|
vs_event_work);
|
|
|
struct vhost_virtqueue *vq = &vs->vqs[VHOST_SCSI_VQ_EVT].vq;
|
|
|
- struct tcm_vhost_evt *evt;
|
|
|
+ struct vhost_scsi_evt *evt;
|
|
|
struct llist_node *llnode;
|
|
|
|
|
|
mutex_lock(&vq->mutex);
|
|
|
llnode = llist_del_all(&vs->vs_event_list);
|
|
|
while (llnode) {
|
|
|
- evt = llist_entry(llnode, struct tcm_vhost_evt, list);
|
|
|
+ evt = llist_entry(llnode, struct vhost_scsi_evt, list);
|
|
|
llnode = llist_next(llnode);
|
|
|
- tcm_vhost_do_evt_work(vs, evt);
|
|
|
- tcm_vhost_free_evt(vs, evt);
|
|
|
+ vhost_scsi_do_evt_work(vs, evt);
|
|
|
+ vhost_scsi_free_evt(vs, evt);
|
|
|
}
|
|
|
mutex_unlock(&vq->mutex);
|
|
|
}
|
|
@@ -679,15 +682,16 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
|
|
|
vs_completion_work);
|
|
|
DECLARE_BITMAP(signal, VHOST_SCSI_MAX_VQ);
|
|
|
struct virtio_scsi_cmd_resp v_rsp;
|
|
|
- struct tcm_vhost_cmd *cmd;
|
|
|
+ struct vhost_scsi_cmd *cmd;
|
|
|
struct llist_node *llnode;
|
|
|
struct se_cmd *se_cmd;
|
|
|
+ struct iov_iter iov_iter;
|
|
|
int ret, vq;
|
|
|
|
|
|
bitmap_zero(signal, VHOST_SCSI_MAX_VQ);
|
|
|
llnode = llist_del_all(&vs->vs_completion_list);
|
|
|
while (llnode) {
|
|
|
- cmd = llist_entry(llnode, struct tcm_vhost_cmd,
|
|
|
+ cmd = llist_entry(llnode, struct vhost_scsi_cmd,
|
|
|
tvc_completion_list);
|
|
|
llnode = llist_next(llnode);
|
|
|
se_cmd = &cmd->tvc_se_cmd;
|
|
@@ -703,8 +707,11 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
|
|
|
se_cmd->scsi_sense_length);
|
|
|
memcpy(v_rsp.sense, cmd->tvc_sense_buf,
|
|
|
se_cmd->scsi_sense_length);
|
|
|
- ret = copy_to_user(cmd->tvc_resp, &v_rsp, sizeof(v_rsp));
|
|
|
- if (likely(ret == 0)) {
|
|
|
+
|
|
|
+ iov_iter_init(&iov_iter, READ, cmd->tvc_resp_iov,
|
|
|
+ cmd->tvc_in_iovs, sizeof(v_rsp));
|
|
|
+ ret = copy_to_iter(&v_rsp, sizeof(v_rsp), &iov_iter);
|
|
|
+ if (likely(ret == sizeof(v_rsp))) {
|
|
|
struct vhost_scsi_virtqueue *q;
|
|
|
vhost_add_used(cmd->tvc_vq, cmd->tvc_vq_desc, 0);
|
|
|
q = container_of(cmd->tvc_vq, struct vhost_scsi_virtqueue, vq);
|
|
@@ -722,13 +729,13 @@ static void vhost_scsi_complete_cmd_work(struct vhost_work *work)
|
|
|
vhost_signal(&vs->dev, &vs->vqs[vq].vq);
|
|
|
}
|
|
|
|
|
|
-static struct tcm_vhost_cmd *
|
|
|
-vhost_scsi_get_tag(struct vhost_virtqueue *vq, struct tcm_vhost_tpg *tpg,
|
|
|
+static struct vhost_scsi_cmd *
|
|
|
+vhost_scsi_get_tag(struct vhost_virtqueue *vq, struct vhost_scsi_tpg *tpg,
|
|
|
unsigned char *cdb, u64 scsi_tag, u16 lun, u8 task_attr,
|
|
|
u32 exp_data_len, int data_direction)
|
|
|
{
|
|
|
- struct tcm_vhost_cmd *cmd;
|
|
|
- struct tcm_vhost_nexus *tv_nexus;
|
|
|
+ struct vhost_scsi_cmd *cmd;
|
|
|
+ struct vhost_scsi_nexus *tv_nexus;
|
|
|
struct se_session *se_sess;
|
|
|
struct scatterlist *sg, *prot_sg;
|
|
|
struct page **pages;
|
|
@@ -736,22 +743,22 @@ vhost_scsi_get_tag(struct vhost_virtqueue *vq, struct tcm_vhost_tpg *tpg,
|
|
|
|
|
|
tv_nexus = tpg->tpg_nexus;
|
|
|
if (!tv_nexus) {
|
|
|
- pr_err("Unable to locate active struct tcm_vhost_nexus\n");
|
|
|
+ pr_err("Unable to locate active struct vhost_scsi_nexus\n");
|
|
|
return ERR_PTR(-EIO);
|
|
|
}
|
|
|
se_sess = tv_nexus->tvn_se_sess;
|
|
|
|
|
|
tag = percpu_ida_alloc(&se_sess->sess_tag_pool, TASK_RUNNING);
|
|
|
if (tag < 0) {
|
|
|
- pr_err("Unable to obtain tag for tcm_vhost_cmd\n");
|
|
|
+ pr_err("Unable to obtain tag for vhost_scsi_cmd\n");
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
}
|
|
|
|
|
|
- cmd = &((struct tcm_vhost_cmd *)se_sess->sess_cmd_map)[tag];
|
|
|
+ cmd = &((struct vhost_scsi_cmd *)se_sess->sess_cmd_map)[tag];
|
|
|
sg = cmd->tvc_sgl;
|
|
|
prot_sg = cmd->tvc_prot_sgl;
|
|
|
pages = cmd->tvc_upages;
|
|
|
- memset(cmd, 0, sizeof(struct tcm_vhost_cmd));
|
|
|
+ memset(cmd, 0, sizeof(struct vhost_scsi_cmd));
|
|
|
|
|
|
cmd->tvc_sgl = sg;
|
|
|
cmd->tvc_prot_sgl = prot_sg;
|
|
@@ -763,9 +770,9 @@ vhost_scsi_get_tag(struct vhost_virtqueue *vq, struct tcm_vhost_tpg *tpg,
|
|
|
cmd->tvc_exp_data_len = exp_data_len;
|
|
|
cmd->tvc_data_direction = data_direction;
|
|
|
cmd->tvc_nexus = tv_nexus;
|
|
|
- cmd->inflight = tcm_vhost_get_inflight(vq);
|
|
|
+ cmd->inflight = vhost_scsi_get_inflight(vq);
|
|
|
|
|
|
- memcpy(cmd->tvc_cdb, cdb, TCM_VHOST_MAX_CDB_SIZE);
|
|
|
+ memcpy(cmd->tvc_cdb, cdb, VHOST_SCSI_MAX_CDB_SIZE);
|
|
|
|
|
|
return cmd;
|
|
|
}
|
|
@@ -776,29 +783,22 @@ vhost_scsi_get_tag(struct vhost_virtqueue *vq, struct tcm_vhost_tpg *tpg,
|
|
|
* Returns the number of scatterlist entries used or -errno on error.
|
|
|
*/
|
|
|
static int
|
|
|
-vhost_scsi_map_to_sgl(struct tcm_vhost_cmd *tv_cmd,
|
|
|
+vhost_scsi_map_to_sgl(struct vhost_scsi_cmd *cmd,
|
|
|
+ void __user *ptr,
|
|
|
+ size_t len,
|
|
|
struct scatterlist *sgl,
|
|
|
- unsigned int sgl_count,
|
|
|
- struct iovec *iov,
|
|
|
- struct page **pages,
|
|
|
bool write)
|
|
|
{
|
|
|
- unsigned int npages = 0, pages_nr, offset, nbytes;
|
|
|
+ unsigned int npages = 0, offset, nbytes;
|
|
|
+ unsigned int pages_nr = iov_num_pages(ptr, len);
|
|
|
struct scatterlist *sg = sgl;
|
|
|
- void __user *ptr = iov->iov_base;
|
|
|
- size_t len = iov->iov_len;
|
|
|
+ struct page **pages = cmd->tvc_upages;
|
|
|
int ret, i;
|
|
|
|
|
|
- pages_nr = iov_num_pages(iov);
|
|
|
- if (pages_nr > sgl_count) {
|
|
|
- pr_err("vhost_scsi_map_to_sgl() pages_nr: %u greater than"
|
|
|
- " sgl_count: %u\n", pages_nr, sgl_count);
|
|
|
- return -ENOBUFS;
|
|
|
- }
|
|
|
- if (pages_nr > TCM_VHOST_PREALLOC_UPAGES) {
|
|
|
+ if (pages_nr > VHOST_SCSI_PREALLOC_UPAGES) {
|
|
|
pr_err("vhost_scsi_map_to_sgl() pages_nr: %u greater than"
|
|
|
- " preallocated TCM_VHOST_PREALLOC_UPAGES: %u\n",
|
|
|
- pages_nr, TCM_VHOST_PREALLOC_UPAGES);
|
|
|
+ " preallocated VHOST_SCSI_PREALLOC_UPAGES: %u\n",
|
|
|
+ pages_nr, VHOST_SCSI_PREALLOC_UPAGES);
|
|
|
return -ENOBUFS;
|
|
|
}
|
|
|
|
|
@@ -829,84 +829,94 @@ out:
|
|
|
}
|
|
|
|
|
|
static int
|
|
|
-vhost_scsi_map_iov_to_sgl(struct tcm_vhost_cmd *cmd,
|
|
|
- struct iovec *iov,
|
|
|
- int niov,
|
|
|
- bool write)
|
|
|
+vhost_scsi_calc_sgls(struct iov_iter *iter, size_t bytes, int max_sgls)
|
|
|
{
|
|
|
- struct scatterlist *sg = cmd->tvc_sgl;
|
|
|
- unsigned int sgl_count = 0;
|
|
|
- int ret, i;
|
|
|
+ int sgl_count = 0;
|
|
|
|
|
|
- for (i = 0; i < niov; i++)
|
|
|
- sgl_count += iov_num_pages(&iov[i]);
|
|
|
+ if (!iter || !iter->iov) {
|
|
|
+ pr_err("%s: iter->iov is NULL, but expected bytes: %zu"
|
|
|
+ " present\n", __func__, bytes);
|
|
|
+ return -EINVAL;
|
|
|
+ }
|
|
|
|
|
|
- if (sgl_count > TCM_VHOST_PREALLOC_SGLS) {
|
|
|
- pr_err("vhost_scsi_map_iov_to_sgl() sgl_count: %u greater than"
|
|
|
- " preallocated TCM_VHOST_PREALLOC_SGLS: %u\n",
|
|
|
- sgl_count, TCM_VHOST_PREALLOC_SGLS);
|
|
|
- return -ENOBUFS;
|
|
|
+ sgl_count = iov_iter_npages(iter, 0xffff);
|
|
|
+ if (sgl_count > max_sgls) {
|
|
|
+ pr_err("%s: requested sgl_count: %d exceeds pre-allocated"
|
|
|
+ " max_sgls: %d\n", __func__, sgl_count, max_sgls);
|
|
|
+ return -EINVAL;
|
|
|
}
|
|
|
+ return sgl_count;
|
|
|
+}
|
|
|
|
|
|
- pr_debug("%s sg %p sgl_count %u\n", __func__, sg, sgl_count);
|
|
|
- sg_init_table(sg, sgl_count);
|
|
|
- cmd->tvc_sgl_count = sgl_count;
|
|
|
+static int
|
|
|
+vhost_scsi_iov_to_sgl(struct vhost_scsi_cmd *cmd, bool write,
|
|
|
+ struct iov_iter *iter,
|
|
|
+ struct scatterlist *sg, int sg_count)
|
|
|
+{
|
|
|
+ size_t off = iter->iov_offset;
|
|
|
+ int i, ret;
|
|
|
|
|
|
- pr_debug("Mapping iovec %p for %u pages\n", &iov[0], sgl_count);
|
|
|
+ for (i = 0; i < iter->nr_segs; i++) {
|
|
|
+ void __user *base = iter->iov[i].iov_base + off;
|
|
|
+ size_t len = iter->iov[i].iov_len - off;
|
|
|
|
|
|
- for (i = 0; i < niov; i++) {
|
|
|
- ret = vhost_scsi_map_to_sgl(cmd, sg, sgl_count, &iov[i],
|
|
|
- cmd->tvc_upages, write);
|
|
|
+ ret = vhost_scsi_map_to_sgl(cmd, base, len, sg, write);
|
|
|
if (ret < 0) {
|
|
|
- for (i = 0; i < cmd->tvc_sgl_count; i++)
|
|
|
- put_page(sg_page(&cmd->tvc_sgl[i]));
|
|
|
-
|
|
|
- cmd->tvc_sgl_count = 0;
|
|
|
+ for (i = 0; i < sg_count; i++) {
|
|
|
+ struct page *page = sg_page(&sg[i]);
|
|
|
+ if (page)
|
|
|
+ put_page(page);
|
|
|
+ }
|
|
|
return ret;
|
|
|
}
|
|
|
sg += ret;
|
|
|
- sgl_count -= ret;
|
|
|
+ off = 0;
|
|
|
}
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
static int
|
|
|
-vhost_scsi_map_iov_to_prot(struct tcm_vhost_cmd *cmd,
|
|
|
- struct iovec *iov,
|
|
|
- int niov,
|
|
|
- bool write)
|
|
|
-{
|
|
|
- struct scatterlist *prot_sg = cmd->tvc_prot_sgl;
|
|
|
- unsigned int prot_sgl_count = 0;
|
|
|
- int ret, i;
|
|
|
-
|
|
|
- for (i = 0; i < niov; i++)
|
|
|
- prot_sgl_count += iov_num_pages(&iov[i]);
|
|
|
-
|
|
|
- if (prot_sgl_count > TCM_VHOST_PREALLOC_PROT_SGLS) {
|
|
|
- pr_err("vhost_scsi_map_iov_to_prot() sgl_count: %u greater than"
|
|
|
- " preallocated TCM_VHOST_PREALLOC_PROT_SGLS: %u\n",
|
|
|
- prot_sgl_count, TCM_VHOST_PREALLOC_PROT_SGLS);
|
|
|
- return -ENOBUFS;
|
|
|
- }
|
|
|
-
|
|
|
- pr_debug("%s prot_sg %p prot_sgl_count %u\n", __func__,
|
|
|
- prot_sg, prot_sgl_count);
|
|
|
- sg_init_table(prot_sg, prot_sgl_count);
|
|
|
- cmd->tvc_prot_sgl_count = prot_sgl_count;
|
|
|
-
|
|
|
- for (i = 0; i < niov; i++) {
|
|
|
- ret = vhost_scsi_map_to_sgl(cmd, prot_sg, prot_sgl_count, &iov[i],
|
|
|
- cmd->tvc_upages, write);
|
|
|
+vhost_scsi_mapal(struct vhost_scsi_cmd *cmd,
|
|
|
+ size_t prot_bytes, struct iov_iter *prot_iter,
|
|
|
+ size_t data_bytes, struct iov_iter *data_iter)
|
|
|
+{
|
|
|
+ int sgl_count, ret;
|
|
|
+ bool write = (cmd->tvc_data_direction == DMA_FROM_DEVICE);
|
|
|
+
|
|
|
+ if (prot_bytes) {
|
|
|
+ sgl_count = vhost_scsi_calc_sgls(prot_iter, prot_bytes,
|
|
|
+ VHOST_SCSI_PREALLOC_PROT_SGLS);
|
|
|
+ if (sgl_count < 0)
|
|
|
+ return sgl_count;
|
|
|
+
|
|
|
+ sg_init_table(cmd->tvc_prot_sgl, sgl_count);
|
|
|
+ cmd->tvc_prot_sgl_count = sgl_count;
|
|
|
+ pr_debug("%s prot_sg %p prot_sgl_count %u\n", __func__,
|
|
|
+ cmd->tvc_prot_sgl, cmd->tvc_prot_sgl_count);
|
|
|
+
|
|
|
+ ret = vhost_scsi_iov_to_sgl(cmd, write, prot_iter,
|
|
|
+ cmd->tvc_prot_sgl,
|
|
|
+ cmd->tvc_prot_sgl_count);
|
|
|
if (ret < 0) {
|
|
|
- for (i = 0; i < cmd->tvc_prot_sgl_count; i++)
|
|
|
- put_page(sg_page(&cmd->tvc_prot_sgl[i]));
|
|
|
-
|
|
|
cmd->tvc_prot_sgl_count = 0;
|
|
|
return ret;
|
|
|
}
|
|
|
- prot_sg += ret;
|
|
|
- prot_sgl_count -= ret;
|
|
|
+ }
|
|
|
+ sgl_count = vhost_scsi_calc_sgls(data_iter, data_bytes,
|
|
|
+ VHOST_SCSI_PREALLOC_SGLS);
|
|
|
+ if (sgl_count < 0)
|
|
|
+ return sgl_count;
|
|
|
+
|
|
|
+ sg_init_table(cmd->tvc_sgl, sgl_count);
|
|
|
+ cmd->tvc_sgl_count = sgl_count;
|
|
|
+ pr_debug("%s data_sg %p data_sgl_count %u\n", __func__,
|
|
|
+ cmd->tvc_sgl, cmd->tvc_sgl_count);
|
|
|
+
|
|
|
+ ret = vhost_scsi_iov_to_sgl(cmd, write, data_iter,
|
|
|
+ cmd->tvc_sgl, cmd->tvc_sgl_count);
|
|
|
+ if (ret < 0) {
|
|
|
+ cmd->tvc_sgl_count = 0;
|
|
|
+ return ret;
|
|
|
}
|
|
|
return 0;
|
|
|
}
|
|
@@ -928,11 +938,11 @@ static int vhost_scsi_to_tcm_attr(int attr)
|
|
|
return TCM_SIMPLE_TAG;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_submission_work(struct work_struct *work)
|
|
|
+static void vhost_scsi_submission_work(struct work_struct *work)
|
|
|
{
|
|
|
- struct tcm_vhost_cmd *cmd =
|
|
|
- container_of(work, struct tcm_vhost_cmd, work);
|
|
|
- struct tcm_vhost_nexus *tv_nexus;
|
|
|
+ struct vhost_scsi_cmd *cmd =
|
|
|
+ container_of(work, struct vhost_scsi_cmd, work);
|
|
|
+ struct vhost_scsi_nexus *tv_nexus;
|
|
|
struct se_cmd *se_cmd = &cmd->tvc_se_cmd;
|
|
|
struct scatterlist *sg_ptr, *sg_prot_ptr = NULL;
|
|
|
int rc;
|
|
@@ -986,19 +996,20 @@ vhost_scsi_send_bad_target(struct vhost_scsi *vs,
|
|
|
static void
|
|
|
vhost_scsi_handle_vq(struct vhost_scsi *vs, struct vhost_virtqueue *vq)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg **vs_tpg;
|
|
|
+ struct vhost_scsi_tpg **vs_tpg, *tpg;
|
|
|
struct virtio_scsi_cmd_req v_req;
|
|
|
struct virtio_scsi_cmd_req_pi v_req_pi;
|
|
|
- struct tcm_vhost_tpg *tpg;
|
|
|
- struct tcm_vhost_cmd *cmd;
|
|
|
+ struct vhost_scsi_cmd *cmd;
|
|
|
+ struct iov_iter out_iter, in_iter, prot_iter, data_iter;
|
|
|
u64 tag;
|
|
|
- u32 exp_data_len, data_first, data_num, data_direction, prot_first;
|
|
|
- unsigned out, in, i;
|
|
|
- int head, ret, data_niov, prot_niov, prot_bytes;
|
|
|
- size_t req_size;
|
|
|
+ u32 exp_data_len, data_direction;
|
|
|
+ unsigned out, in;
|
|
|
+ int head, ret, prot_bytes;
|
|
|
+ size_t req_size, rsp_size = sizeof(struct virtio_scsi_cmd_resp);
|
|
|
+ size_t out_size, in_size;
|
|
|
u16 lun;
|
|
|
u8 *target, *lunp, task_attr;
|
|
|
- bool hdr_pi;
|
|
|
+ bool t10_pi = vhost_has_feature(vq, VIRTIO_SCSI_F_T10_PI);
|
|
|
void *req, *cdb;
|
|
|
|
|
|
mutex_lock(&vq->mutex);
|
|
@@ -1014,10 +1025,10 @@ vhost_scsi_handle_vq(struct vhost_scsi *vs, struct vhost_virtqueue *vq)
|
|
|
|
|
|
for (;;) {
|
|
|
head = vhost_get_vq_desc(vq, vq->iov,
|
|
|
- ARRAY_SIZE(vq->iov), &out, &in,
|
|
|
- NULL, NULL);
|
|
|
+ ARRAY_SIZE(vq->iov), &out, &in,
|
|
|
+ NULL, NULL);
|
|
|
pr_debug("vhost_get_vq_desc: head: %d, out: %u in: %u\n",
|
|
|
- head, out, in);
|
|
|
+ head, out, in);
|
|
|
/* On error, stop handling until the next kick. */
|
|
|
if (unlikely(head < 0))
|
|
|
break;
|
|
@@ -1029,113 +1040,134 @@ vhost_scsi_handle_vq(struct vhost_scsi *vs, struct vhost_virtqueue *vq)
|
|
|
}
|
|
|
break;
|
|
|
}
|
|
|
-
|
|
|
- /* FIXME: BIDI operation */
|
|
|
- if (out == 1 && in == 1) {
|
|
|
- data_direction = DMA_NONE;
|
|
|
- data_first = 0;
|
|
|
- data_num = 0;
|
|
|
- } else if (out == 1 && in > 1) {
|
|
|
- data_direction = DMA_FROM_DEVICE;
|
|
|
- data_first = out + 1;
|
|
|
- data_num = in - 1;
|
|
|
- } else if (out > 1 && in == 1) {
|
|
|
- data_direction = DMA_TO_DEVICE;
|
|
|
- data_first = 1;
|
|
|
- data_num = out - 1;
|
|
|
- } else {
|
|
|
- vq_err(vq, "Invalid buffer layout out: %u in: %u\n",
|
|
|
- out, in);
|
|
|
- break;
|
|
|
- }
|
|
|
-
|
|
|
/*
|
|
|
- * Check for a sane resp buffer so we can report errors to
|
|
|
- * the guest.
|
|
|
+ * Check for a sane response buffer so we can report early
|
|
|
+ * errors back to the guest.
|
|
|
*/
|
|
|
- if (unlikely(vq->iov[out].iov_len !=
|
|
|
- sizeof(struct virtio_scsi_cmd_resp))) {
|
|
|
- vq_err(vq, "Expecting virtio_scsi_cmd_resp, got %zu"
|
|
|
- " bytes\n", vq->iov[out].iov_len);
|
|
|
+ if (unlikely(vq->iov[out].iov_len < rsp_size)) {
|
|
|
+ vq_err(vq, "Expecting at least virtio_scsi_cmd_resp"
|
|
|
+ " size, got %zu bytes\n", vq->iov[out].iov_len);
|
|
|
break;
|
|
|
}
|
|
|
-
|
|
|
- if (vhost_has_feature(vq, VIRTIO_SCSI_F_T10_PI)) {
|
|
|
+ /*
|
|
|
+ * Setup pointers and values based upon different virtio-scsi
|
|
|
+ * request header if T10_PI is enabled in KVM guest.
|
|
|
+ */
|
|
|
+ if (t10_pi) {
|
|
|
req = &v_req_pi;
|
|
|
+ req_size = sizeof(v_req_pi);
|
|
|
lunp = &v_req_pi.lun[0];
|
|
|
target = &v_req_pi.lun[1];
|
|
|
- req_size = sizeof(v_req_pi);
|
|
|
- hdr_pi = true;
|
|
|
} else {
|
|
|
req = &v_req;
|
|
|
+ req_size = sizeof(v_req);
|
|
|
lunp = &v_req.lun[0];
|
|
|
target = &v_req.lun[1];
|
|
|
- req_size = sizeof(v_req);
|
|
|
- hdr_pi = false;
|
|
|
}
|
|
|
+ /*
|
|
|
+ * FIXME: Not correct for BIDI operation
|
|
|
+ */
|
|
|
+ out_size = iov_length(vq->iov, out);
|
|
|
+ in_size = iov_length(&vq->iov[out], in);
|
|
|
|
|
|
- if (unlikely(vq->iov[0].iov_len < req_size)) {
|
|
|
- pr_err("Expecting virtio-scsi header: %zu, got %zu\n",
|
|
|
- req_size, vq->iov[0].iov_len);
|
|
|
- break;
|
|
|
- }
|
|
|
- ret = copy_from_user(req, vq->iov[0].iov_base, req_size);
|
|
|
- if (unlikely(ret)) {
|
|
|
- vq_err(vq, "Faulted on virtio_scsi_cmd_req\n");
|
|
|
- break;
|
|
|
- }
|
|
|
+ /*
|
|
|
+ * Copy over the virtio-scsi request header, which for a
|
|
|
+ * ANY_LAYOUT enabled guest may span multiple iovecs, or a
|
|
|
+ * single iovec may contain both the header + outgoing
|
|
|
+ * WRITE payloads.
|
|
|
+ *
|
|
|
+ * copy_from_iter() will advance out_iter, so that it will
|
|
|
+ * point at the start of the outgoing WRITE payload, if
|
|
|
+ * DMA_TO_DEVICE is set.
|
|
|
+ */
|
|
|
+ iov_iter_init(&out_iter, WRITE, vq->iov, out, out_size);
|
|
|
|
|
|
+ ret = copy_from_iter(req, req_size, &out_iter);
|
|
|
+ if (unlikely(ret != req_size)) {
|
|
|
+ vq_err(vq, "Faulted on copy_from_iter\n");
|
|
|
+ vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
+ continue;
|
|
|
+ }
|
|
|
/* virtio-scsi spec requires byte 0 of the lun to be 1 */
|
|
|
if (unlikely(*lunp != 1)) {
|
|
|
+ vq_err(vq, "Illegal virtio-scsi lun: %u\n", *lunp);
|
|
|
vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
continue;
|
|
|
}
|
|
|
|
|
|
tpg = ACCESS_ONCE(vs_tpg[*target]);
|
|
|
-
|
|
|
- /* Target does not exist, fail the request */
|
|
|
if (unlikely(!tpg)) {
|
|
|
+ /* Target does not exist, fail the request */
|
|
|
vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
continue;
|
|
|
}
|
|
|
-
|
|
|
- data_niov = data_num;
|
|
|
- prot_niov = prot_first = prot_bytes = 0;
|
|
|
/*
|
|
|
- * Determine if any protection information iovecs are preceeding
|
|
|
- * the actual data payload, and adjust data_first + data_niov
|
|
|
- * values accordingly for vhost_scsi_map_iov_to_sgl() below.
|
|
|
+ * Determine data_direction by calculating the total outgoing
|
|
|
+ * iovec sizes + incoming iovec sizes vs. virtio-scsi request +
|
|
|
+ * response headers respectively.
|
|
|
*
|
|
|
- * Also extract virtio_scsi header bits for vhost_scsi_get_tag()
|
|
|
+ * For DMA_TO_DEVICE this is out_iter, which is already pointing
|
|
|
+ * to the right place.
|
|
|
+ *
|
|
|
+ * For DMA_FROM_DEVICE, the iovec will be just past the end
|
|
|
+ * of the virtio-scsi response header in either the same
|
|
|
+ * or immediately following iovec.
|
|
|
+ *
|
|
|
+ * Any associated T10_PI bytes for the outgoing / incoming
|
|
|
+ * payloads are included in calculation of exp_data_len here.
|
|
|
*/
|
|
|
- if (hdr_pi) {
|
|
|
+ prot_bytes = 0;
|
|
|
+
|
|
|
+ if (out_size > req_size) {
|
|
|
+ data_direction = DMA_TO_DEVICE;
|
|
|
+ exp_data_len = out_size - req_size;
|
|
|
+ data_iter = out_iter;
|
|
|
+ } else if (in_size > rsp_size) {
|
|
|
+ data_direction = DMA_FROM_DEVICE;
|
|
|
+ exp_data_len = in_size - rsp_size;
|
|
|
+
|
|
|
+ iov_iter_init(&in_iter, READ, &vq->iov[out], in,
|
|
|
+ rsp_size + exp_data_len);
|
|
|
+ iov_iter_advance(&in_iter, rsp_size);
|
|
|
+ data_iter = in_iter;
|
|
|
+ } else {
|
|
|
+ data_direction = DMA_NONE;
|
|
|
+ exp_data_len = 0;
|
|
|
+ }
|
|
|
+ /*
|
|
|
+ * If T10_PI header + payload is present, setup prot_iter values
|
|
|
+ * and recalculate data_iter for vhost_scsi_mapal() mapping to
|
|
|
+ * host scatterlists via get_user_pages_fast().
|
|
|
+ */
|
|
|
+ if (t10_pi) {
|
|
|
if (v_req_pi.pi_bytesout) {
|
|
|
if (data_direction != DMA_TO_DEVICE) {
|
|
|
- vq_err(vq, "Received non zero do_pi_niov"
|
|
|
- ", but wrong data_direction\n");
|
|
|
- goto err_cmd;
|
|
|
+ vq_err(vq, "Received non zero pi_bytesout,"
|
|
|
+ " but wrong data_direction\n");
|
|
|
+ vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
+ continue;
|
|
|
}
|
|
|
prot_bytes = vhost32_to_cpu(vq, v_req_pi.pi_bytesout);
|
|
|
} else if (v_req_pi.pi_bytesin) {
|
|
|
if (data_direction != DMA_FROM_DEVICE) {
|
|
|
- vq_err(vq, "Received non zero di_pi_niov"
|
|
|
- ", but wrong data_direction\n");
|
|
|
- goto err_cmd;
|
|
|
+ vq_err(vq, "Received non zero pi_bytesin,"
|
|
|
+ " but wrong data_direction\n");
|
|
|
+ vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
+ continue;
|
|
|
}
|
|
|
prot_bytes = vhost32_to_cpu(vq, v_req_pi.pi_bytesin);
|
|
|
}
|
|
|
+ /*
|
|
|
+ * Set prot_iter to data_iter, and advance past any
|
|
|
+ * preceeding prot_bytes that may be present.
|
|
|
+ *
|
|
|
+ * Also fix up the exp_data_len to reflect only the
|
|
|
+ * actual data payload length.
|
|
|
+ */
|
|
|
if (prot_bytes) {
|
|
|
- int tmp = 0;
|
|
|
-
|
|
|
- for (i = 0; i < data_num; i++) {
|
|
|
- tmp += vq->iov[data_first + i].iov_len;
|
|
|
- prot_niov++;
|
|
|
- if (tmp >= prot_bytes)
|
|
|
- break;
|
|
|
- }
|
|
|
- prot_first = data_first;
|
|
|
- data_first += prot_niov;
|
|
|
- data_niov = data_num - prot_niov;
|
|
|
+ exp_data_len -= prot_bytes;
|
|
|
+ prot_iter = data_iter;
|
|
|
+ iov_iter_advance(&data_iter, prot_bytes);
|
|
|
}
|
|
|
tag = vhost64_to_cpu(vq, v_req_pi.tag);
|
|
|
task_attr = v_req_pi.task_attr;
|
|
@@ -1147,83 +1179,65 @@ vhost_scsi_handle_vq(struct vhost_scsi *vs, struct vhost_virtqueue *vq)
|
|
|
cdb = &v_req.cdb[0];
|
|
|
lun = ((v_req.lun[2] << 8) | v_req.lun[3]) & 0x3FFF;
|
|
|
}
|
|
|
- exp_data_len = 0;
|
|
|
- for (i = 0; i < data_niov; i++)
|
|
|
- exp_data_len += vq->iov[data_first + i].iov_len;
|
|
|
/*
|
|
|
- * Check that the recieved CDB size does not exceeded our
|
|
|
- * hardcoded max for vhost-scsi
|
|
|
+ * Check that the received CDB size does not exceeded our
|
|
|
+ * hardcoded max for vhost-scsi, then get a pre-allocated
|
|
|
+ * cmd descriptor for the new virtio-scsi tag.
|
|
|
*
|
|
|
* TODO what if cdb was too small for varlen cdb header?
|
|
|
*/
|
|
|
- if (unlikely(scsi_command_size(cdb) > TCM_VHOST_MAX_CDB_SIZE)) {
|
|
|
+ if (unlikely(scsi_command_size(cdb) > VHOST_SCSI_MAX_CDB_SIZE)) {
|
|
|
vq_err(vq, "Received SCSI CDB with command_size: %d that"
|
|
|
" exceeds SCSI_MAX_VARLEN_CDB_SIZE: %d\n",
|
|
|
- scsi_command_size(cdb), TCM_VHOST_MAX_CDB_SIZE);
|
|
|
- goto err_cmd;
|
|
|
+ scsi_command_size(cdb), VHOST_SCSI_MAX_CDB_SIZE);
|
|
|
+ vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
+ continue;
|
|
|
}
|
|
|
-
|
|
|
cmd = vhost_scsi_get_tag(vq, tpg, cdb, tag, lun, task_attr,
|
|
|
exp_data_len + prot_bytes,
|
|
|
data_direction);
|
|
|
if (IS_ERR(cmd)) {
|
|
|
vq_err(vq, "vhost_scsi_get_tag failed %ld\n",
|
|
|
- PTR_ERR(cmd));
|
|
|
- goto err_cmd;
|
|
|
+ PTR_ERR(cmd));
|
|
|
+ vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
+ continue;
|
|
|
}
|
|
|
-
|
|
|
- pr_debug("Allocated tv_cmd: %p exp_data_len: %d, data_direction"
|
|
|
- ": %d\n", cmd, exp_data_len, data_direction);
|
|
|
-
|
|
|
cmd->tvc_vhost = vs;
|
|
|
cmd->tvc_vq = vq;
|
|
|
- cmd->tvc_resp = vq->iov[out].iov_base;
|
|
|
+ cmd->tvc_resp_iov = &vq->iov[out];
|
|
|
+ cmd->tvc_in_iovs = in;
|
|
|
|
|
|
pr_debug("vhost_scsi got command opcode: %#02x, lun: %d\n",
|
|
|
- cmd->tvc_cdb[0], cmd->tvc_lun);
|
|
|
+ cmd->tvc_cdb[0], cmd->tvc_lun);
|
|
|
+ pr_debug("cmd: %p exp_data_len: %d, prot_bytes: %d data_direction:"
|
|
|
+ " %d\n", cmd, exp_data_len, prot_bytes, data_direction);
|
|
|
|
|
|
- if (prot_niov) {
|
|
|
- ret = vhost_scsi_map_iov_to_prot(cmd,
|
|
|
- &vq->iov[prot_first], prot_niov,
|
|
|
- data_direction == DMA_FROM_DEVICE);
|
|
|
- if (unlikely(ret)) {
|
|
|
- vq_err(vq, "Failed to map iov to"
|
|
|
- " prot_sgl\n");
|
|
|
- goto err_free;
|
|
|
- }
|
|
|
- }
|
|
|
if (data_direction != DMA_NONE) {
|
|
|
- ret = vhost_scsi_map_iov_to_sgl(cmd,
|
|
|
- &vq->iov[data_first], data_niov,
|
|
|
- data_direction == DMA_FROM_DEVICE);
|
|
|
+ ret = vhost_scsi_mapal(cmd,
|
|
|
+ prot_bytes, &prot_iter,
|
|
|
+ exp_data_len, &data_iter);
|
|
|
if (unlikely(ret)) {
|
|
|
vq_err(vq, "Failed to map iov to sgl\n");
|
|
|
- goto err_free;
|
|
|
+ vhost_scsi_release_cmd(&cmd->tvc_se_cmd);
|
|
|
+ vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
+ continue;
|
|
|
}
|
|
|
}
|
|
|
/*
|
|
|
* Save the descriptor from vhost_get_vq_desc() to be used to
|
|
|
* complete the virtio-scsi request in TCM callback context via
|
|
|
- * tcm_vhost_queue_data_in() and tcm_vhost_queue_status()
|
|
|
+ * vhost_scsi_queue_data_in() and vhost_scsi_queue_status()
|
|
|
*/
|
|
|
cmd->tvc_vq_desc = head;
|
|
|
/*
|
|
|
- * Dispatch tv_cmd descriptor for cmwq execution in process
|
|
|
- * context provided by tcm_vhost_workqueue. This also ensures
|
|
|
- * tv_cmd is executed on the same kworker CPU as this vhost
|
|
|
- * thread to gain positive L2 cache locality effects..
|
|
|
+ * Dispatch cmd descriptor for cmwq execution in process
|
|
|
+ * context provided by vhost_scsi_workqueue. This also ensures
|
|
|
+ * cmd is executed on the same kworker CPU as this vhost
|
|
|
+ * thread to gain positive L2 cache locality effects.
|
|
|
*/
|
|
|
- INIT_WORK(&cmd->work, tcm_vhost_submission_work);
|
|
|
- queue_work(tcm_vhost_workqueue, &cmd->work);
|
|
|
+ INIT_WORK(&cmd->work, vhost_scsi_submission_work);
|
|
|
+ queue_work(vhost_scsi_workqueue, &cmd->work);
|
|
|
}
|
|
|
-
|
|
|
- mutex_unlock(&vq->mutex);
|
|
|
- return;
|
|
|
-
|
|
|
-err_free:
|
|
|
- vhost_scsi_free_cmd(cmd);
|
|
|
-err_cmd:
|
|
|
- vhost_scsi_send_bad_target(vs, vq, head, out);
|
|
|
out:
|
|
|
mutex_unlock(&vq->mutex);
|
|
|
}
|
|
@@ -1234,15 +1248,15 @@ static void vhost_scsi_ctl_handle_kick(struct vhost_work *work)
|
|
|
}
|
|
|
|
|
|
static void
|
|
|
-tcm_vhost_send_evt(struct vhost_scsi *vs,
|
|
|
- struct tcm_vhost_tpg *tpg,
|
|
|
+vhost_scsi_send_evt(struct vhost_scsi *vs,
|
|
|
+ struct vhost_scsi_tpg *tpg,
|
|
|
struct se_lun *lun,
|
|
|
u32 event,
|
|
|
u32 reason)
|
|
|
{
|
|
|
- struct tcm_vhost_evt *evt;
|
|
|
+ struct vhost_scsi_evt *evt;
|
|
|
|
|
|
- evt = tcm_vhost_allocate_evt(vs, event, reason);
|
|
|
+ evt = vhost_scsi_allocate_evt(vs, event, reason);
|
|
|
if (!evt)
|
|
|
return;
|
|
|
|
|
@@ -1253,7 +1267,7 @@ tcm_vhost_send_evt(struct vhost_scsi *vs,
|
|
|
* lun[4-7] need to be zero according to virtio-scsi spec.
|
|
|
*/
|
|
|
evt->event.lun[0] = 0x01;
|
|
|
- evt->event.lun[1] = tpg->tport_tpgt & 0xFF;
|
|
|
+ evt->event.lun[1] = tpg->tport_tpgt;
|
|
|
if (lun->unpacked_lun >= 256)
|
|
|
evt->event.lun[2] = lun->unpacked_lun >> 8 | 0x40 ;
|
|
|
evt->event.lun[3] = lun->unpacked_lun & 0xFF;
|
|
@@ -1274,7 +1288,7 @@ static void vhost_scsi_evt_handle_kick(struct vhost_work *work)
|
|
|
goto out;
|
|
|
|
|
|
if (vs->vs_events_missed)
|
|
|
- tcm_vhost_send_evt(vs, NULL, NULL, VIRTIO_SCSI_T_NO_EVENT, 0);
|
|
|
+ vhost_scsi_send_evt(vs, NULL, NULL, VIRTIO_SCSI_T_NO_EVENT, 0);
|
|
|
out:
|
|
|
mutex_unlock(&vq->mutex);
|
|
|
}
|
|
@@ -1300,7 +1314,7 @@ static void vhost_scsi_flush(struct vhost_scsi *vs)
|
|
|
int i;
|
|
|
|
|
|
/* Init new inflight and remember the old inflight */
|
|
|
- tcm_vhost_init_inflight(vs, old_inflight);
|
|
|
+ vhost_scsi_init_inflight(vs, old_inflight);
|
|
|
|
|
|
/*
|
|
|
* The inflight->kref was initialized to 1. We decrement it here to
|
|
@@ -1308,7 +1322,7 @@ static void vhost_scsi_flush(struct vhost_scsi *vs)
|
|
|
* when all the reqs are finished.
|
|
|
*/
|
|
|
for (i = 0; i < VHOST_SCSI_MAX_VQ; i++)
|
|
|
- kref_put(&old_inflight[i]->kref, tcm_vhost_done_inflight);
|
|
|
+ kref_put(&old_inflight[i]->kref, vhost_scsi_done_inflight);
|
|
|
|
|
|
/* Flush both the vhost poll and vhost work */
|
|
|
for (i = 0; i < VHOST_SCSI_MAX_VQ; i++)
|
|
@@ -1323,24 +1337,24 @@ static void vhost_scsi_flush(struct vhost_scsi *vs)
|
|
|
|
|
|
/*
|
|
|
* Called from vhost_scsi_ioctl() context to walk the list of available
|
|
|
- * tcm_vhost_tpg with an active struct tcm_vhost_nexus
|
|
|
+ * vhost_scsi_tpg with an active struct vhost_scsi_nexus
|
|
|
*
|
|
|
* The lock nesting rule is:
|
|
|
- * tcm_vhost_mutex -> vs->dev.mutex -> tpg->tv_tpg_mutex -> vq->mutex
|
|
|
+ * vhost_scsi_mutex -> vs->dev.mutex -> tpg->tv_tpg_mutex -> vq->mutex
|
|
|
*/
|
|
|
static int
|
|
|
vhost_scsi_set_endpoint(struct vhost_scsi *vs,
|
|
|
struct vhost_scsi_target *t)
|
|
|
{
|
|
|
struct se_portal_group *se_tpg;
|
|
|
- struct tcm_vhost_tport *tv_tport;
|
|
|
- struct tcm_vhost_tpg *tpg;
|
|
|
- struct tcm_vhost_tpg **vs_tpg;
|
|
|
+ struct vhost_scsi_tport *tv_tport;
|
|
|
+ struct vhost_scsi_tpg *tpg;
|
|
|
+ struct vhost_scsi_tpg **vs_tpg;
|
|
|
struct vhost_virtqueue *vq;
|
|
|
int index, ret, i, len;
|
|
|
bool match = false;
|
|
|
|
|
|
- mutex_lock(&tcm_vhost_mutex);
|
|
|
+ mutex_lock(&vhost_scsi_mutex);
|
|
|
mutex_lock(&vs->dev.mutex);
|
|
|
|
|
|
/* Verify that ring has been setup correctly. */
|
|
@@ -1361,7 +1375,7 @@ vhost_scsi_set_endpoint(struct vhost_scsi *vs,
|
|
|
if (vs->vs_tpg)
|
|
|
memcpy(vs_tpg, vs->vs_tpg, len);
|
|
|
|
|
|
- list_for_each_entry(tpg, &tcm_vhost_list, tv_tpg_list) {
|
|
|
+ list_for_each_entry(tpg, &vhost_scsi_list, tv_tpg_list) {
|
|
|
mutex_lock(&tpg->tv_tpg_mutex);
|
|
|
if (!tpg->tpg_nexus) {
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
@@ -1429,7 +1443,7 @@ vhost_scsi_set_endpoint(struct vhost_scsi *vs,
|
|
|
|
|
|
out:
|
|
|
mutex_unlock(&vs->dev.mutex);
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
@@ -1438,14 +1452,14 @@ vhost_scsi_clear_endpoint(struct vhost_scsi *vs,
|
|
|
struct vhost_scsi_target *t)
|
|
|
{
|
|
|
struct se_portal_group *se_tpg;
|
|
|
- struct tcm_vhost_tport *tv_tport;
|
|
|
- struct tcm_vhost_tpg *tpg;
|
|
|
+ struct vhost_scsi_tport *tv_tport;
|
|
|
+ struct vhost_scsi_tpg *tpg;
|
|
|
struct vhost_virtqueue *vq;
|
|
|
bool match = false;
|
|
|
int index, ret, i;
|
|
|
u8 target;
|
|
|
|
|
|
- mutex_lock(&tcm_vhost_mutex);
|
|
|
+ mutex_lock(&vhost_scsi_mutex);
|
|
|
mutex_lock(&vs->dev.mutex);
|
|
|
/* Verify that ring has been setup correctly. */
|
|
|
for (index = 0; index < vs->dev.nvqs; ++index) {
|
|
@@ -1511,14 +1525,14 @@ vhost_scsi_clear_endpoint(struct vhost_scsi *vs,
|
|
|
vs->vs_tpg = NULL;
|
|
|
WARN_ON(vs->vs_events_nr);
|
|
|
mutex_unlock(&vs->dev.mutex);
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
return 0;
|
|
|
|
|
|
err_tpg:
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
err_dev:
|
|
|
mutex_unlock(&vs->dev.mutex);
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
@@ -1565,7 +1579,7 @@ static int vhost_scsi_open(struct inode *inode, struct file *f)
|
|
|
goto err_vqs;
|
|
|
|
|
|
vhost_work_init(&vs->vs_completion_work, vhost_scsi_complete_cmd_work);
|
|
|
- vhost_work_init(&vs->vs_event_work, tcm_vhost_evt_work);
|
|
|
+ vhost_work_init(&vs->vs_event_work, vhost_scsi_evt_work);
|
|
|
|
|
|
vs->vs_events_nr = 0;
|
|
|
vs->vs_events_missed = false;
|
|
@@ -1580,7 +1594,7 @@ static int vhost_scsi_open(struct inode *inode, struct file *f)
|
|
|
}
|
|
|
vhost_dev_init(&vs->dev, vqs, VHOST_SCSI_MAX_VQ);
|
|
|
|
|
|
- tcm_vhost_init_inflight(vs, NULL);
|
|
|
+ vhost_scsi_init_inflight(vs, NULL);
|
|
|
|
|
|
f->private_data = vs;
|
|
|
return 0;
|
|
@@ -1712,7 +1726,7 @@ static int vhost_scsi_deregister(void)
|
|
|
return misc_deregister(&vhost_scsi_misc);
|
|
|
}
|
|
|
|
|
|
-static char *tcm_vhost_dump_proto_id(struct tcm_vhost_tport *tport)
|
|
|
+static char *vhost_scsi_dump_proto_id(struct vhost_scsi_tport *tport)
|
|
|
{
|
|
|
switch (tport->tport_proto_id) {
|
|
|
case SCSI_PROTOCOL_SAS:
|
|
@@ -1729,7 +1743,7 @@ static char *tcm_vhost_dump_proto_id(struct tcm_vhost_tport *tport)
|
|
|
}
|
|
|
|
|
|
static void
|
|
|
-tcm_vhost_do_plug(struct tcm_vhost_tpg *tpg,
|
|
|
+vhost_scsi_do_plug(struct vhost_scsi_tpg *tpg,
|
|
|
struct se_lun *lun, bool plug)
|
|
|
{
|
|
|
|
|
@@ -1750,71 +1764,71 @@ tcm_vhost_do_plug(struct tcm_vhost_tpg *tpg,
|
|
|
vq = &vs->vqs[VHOST_SCSI_VQ_EVT].vq;
|
|
|
mutex_lock(&vq->mutex);
|
|
|
if (vhost_has_feature(vq, VIRTIO_SCSI_F_HOTPLUG))
|
|
|
- tcm_vhost_send_evt(vs, tpg, lun,
|
|
|
+ vhost_scsi_send_evt(vs, tpg, lun,
|
|
|
VIRTIO_SCSI_T_TRANSPORT_RESET, reason);
|
|
|
mutex_unlock(&vq->mutex);
|
|
|
mutex_unlock(&vs->dev.mutex);
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_hotplug(struct tcm_vhost_tpg *tpg, struct se_lun *lun)
|
|
|
+static void vhost_scsi_hotplug(struct vhost_scsi_tpg *tpg, struct se_lun *lun)
|
|
|
{
|
|
|
- tcm_vhost_do_plug(tpg, lun, true);
|
|
|
+ vhost_scsi_do_plug(tpg, lun, true);
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_hotunplug(struct tcm_vhost_tpg *tpg, struct se_lun *lun)
|
|
|
+static void vhost_scsi_hotunplug(struct vhost_scsi_tpg *tpg, struct se_lun *lun)
|
|
|
{
|
|
|
- tcm_vhost_do_plug(tpg, lun, false);
|
|
|
+ vhost_scsi_do_plug(tpg, lun, false);
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_port_link(struct se_portal_group *se_tpg,
|
|
|
+static int vhost_scsi_port_link(struct se_portal_group *se_tpg,
|
|
|
struct se_lun *lun)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
|
|
|
- mutex_lock(&tcm_vhost_mutex);
|
|
|
+ mutex_lock(&vhost_scsi_mutex);
|
|
|
|
|
|
mutex_lock(&tpg->tv_tpg_mutex);
|
|
|
tpg->tv_tpg_port_count++;
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
|
|
|
- tcm_vhost_hotplug(tpg, lun);
|
|
|
+ vhost_scsi_hotplug(tpg, lun);
|
|
|
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_port_unlink(struct se_portal_group *se_tpg,
|
|
|
+static void vhost_scsi_port_unlink(struct se_portal_group *se_tpg,
|
|
|
struct se_lun *lun)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
|
|
|
- mutex_lock(&tcm_vhost_mutex);
|
|
|
+ mutex_lock(&vhost_scsi_mutex);
|
|
|
|
|
|
mutex_lock(&tpg->tv_tpg_mutex);
|
|
|
tpg->tv_tpg_port_count--;
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
|
|
|
- tcm_vhost_hotunplug(tpg, lun);
|
|
|
+ vhost_scsi_hotunplug(tpg, lun);
|
|
|
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
}
|
|
|
|
|
|
static struct se_node_acl *
|
|
|
-tcm_vhost_make_nodeacl(struct se_portal_group *se_tpg,
|
|
|
+vhost_scsi_make_nodeacl(struct se_portal_group *se_tpg,
|
|
|
struct config_group *group,
|
|
|
const char *name)
|
|
|
{
|
|
|
struct se_node_acl *se_nacl, *se_nacl_new;
|
|
|
- struct tcm_vhost_nacl *nacl;
|
|
|
+ struct vhost_scsi_nacl *nacl;
|
|
|
u64 wwpn = 0;
|
|
|
u32 nexus_depth;
|
|
|
|
|
|
- /* tcm_vhost_parse_wwn(name, &wwpn, 1) < 0)
|
|
|
+ /* vhost_scsi_parse_wwn(name, &wwpn, 1) < 0)
|
|
|
return ERR_PTR(-EINVAL); */
|
|
|
- se_nacl_new = tcm_vhost_alloc_fabric_acl(se_tpg);
|
|
|
+ se_nacl_new = vhost_scsi_alloc_fabric_acl(se_tpg);
|
|
|
if (!se_nacl_new)
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
@@ -1826,37 +1840,37 @@ tcm_vhost_make_nodeacl(struct se_portal_group *se_tpg,
|
|
|
se_nacl = core_tpg_add_initiator_node_acl(se_tpg, se_nacl_new,
|
|
|
name, nexus_depth);
|
|
|
if (IS_ERR(se_nacl)) {
|
|
|
- tcm_vhost_release_fabric_acl(se_tpg, se_nacl_new);
|
|
|
+ vhost_scsi_release_fabric_acl(se_tpg, se_nacl_new);
|
|
|
return se_nacl;
|
|
|
}
|
|
|
/*
|
|
|
- * Locate our struct tcm_vhost_nacl and set the FC Nport WWPN
|
|
|
+ * Locate our struct vhost_scsi_nacl and set the FC Nport WWPN
|
|
|
*/
|
|
|
- nacl = container_of(se_nacl, struct tcm_vhost_nacl, se_node_acl);
|
|
|
+ nacl = container_of(se_nacl, struct vhost_scsi_nacl, se_node_acl);
|
|
|
nacl->iport_wwpn = wwpn;
|
|
|
|
|
|
return se_nacl;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_drop_nodeacl(struct se_node_acl *se_acl)
|
|
|
+static void vhost_scsi_drop_nodeacl(struct se_node_acl *se_acl)
|
|
|
{
|
|
|
- struct tcm_vhost_nacl *nacl = container_of(se_acl,
|
|
|
- struct tcm_vhost_nacl, se_node_acl);
|
|
|
+ struct vhost_scsi_nacl *nacl = container_of(se_acl,
|
|
|
+ struct vhost_scsi_nacl, se_node_acl);
|
|
|
core_tpg_del_initiator_node_acl(se_acl->se_tpg, se_acl, 1);
|
|
|
kfree(nacl);
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_free_cmd_map_res(struct tcm_vhost_nexus *nexus,
|
|
|
+static void vhost_scsi_free_cmd_map_res(struct vhost_scsi_nexus *nexus,
|
|
|
struct se_session *se_sess)
|
|
|
{
|
|
|
- struct tcm_vhost_cmd *tv_cmd;
|
|
|
+ struct vhost_scsi_cmd *tv_cmd;
|
|
|
unsigned int i;
|
|
|
|
|
|
if (!se_sess->sess_cmd_map)
|
|
|
return;
|
|
|
|
|
|
- for (i = 0; i < TCM_VHOST_DEFAULT_TAGS; i++) {
|
|
|
- tv_cmd = &((struct tcm_vhost_cmd *)se_sess->sess_cmd_map)[i];
|
|
|
+ for (i = 0; i < VHOST_SCSI_DEFAULT_TAGS; i++) {
|
|
|
+ tv_cmd = &((struct vhost_scsi_cmd *)se_sess->sess_cmd_map)[i];
|
|
|
|
|
|
kfree(tv_cmd->tvc_sgl);
|
|
|
kfree(tv_cmd->tvc_prot_sgl);
|
|
@@ -1864,13 +1878,13 @@ static void tcm_vhost_free_cmd_map_res(struct tcm_vhost_nexus *nexus,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
+static int vhost_scsi_make_nexus(struct vhost_scsi_tpg *tpg,
|
|
|
const char *name)
|
|
|
{
|
|
|
struct se_portal_group *se_tpg;
|
|
|
struct se_session *se_sess;
|
|
|
- struct tcm_vhost_nexus *tv_nexus;
|
|
|
- struct tcm_vhost_cmd *tv_cmd;
|
|
|
+ struct vhost_scsi_nexus *tv_nexus;
|
|
|
+ struct vhost_scsi_cmd *tv_cmd;
|
|
|
unsigned int i;
|
|
|
|
|
|
mutex_lock(&tpg->tv_tpg_mutex);
|
|
@@ -1881,19 +1895,19 @@ static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
}
|
|
|
se_tpg = &tpg->se_tpg;
|
|
|
|
|
|
- tv_nexus = kzalloc(sizeof(struct tcm_vhost_nexus), GFP_KERNEL);
|
|
|
+ tv_nexus = kzalloc(sizeof(struct vhost_scsi_nexus), GFP_KERNEL);
|
|
|
if (!tv_nexus) {
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
- pr_err("Unable to allocate struct tcm_vhost_nexus\n");
|
|
|
+ pr_err("Unable to allocate struct vhost_scsi_nexus\n");
|
|
|
return -ENOMEM;
|
|
|
}
|
|
|
/*
|
|
|
* Initialize the struct se_session pointer and setup tagpool
|
|
|
- * for struct tcm_vhost_cmd descriptors
|
|
|
+ * for struct vhost_scsi_cmd descriptors
|
|
|
*/
|
|
|
tv_nexus->tvn_se_sess = transport_init_session_tags(
|
|
|
- TCM_VHOST_DEFAULT_TAGS,
|
|
|
- sizeof(struct tcm_vhost_cmd),
|
|
|
+ VHOST_SCSI_DEFAULT_TAGS,
|
|
|
+ sizeof(struct vhost_scsi_cmd),
|
|
|
TARGET_PROT_DIN_PASS | TARGET_PROT_DOUT_PASS);
|
|
|
if (IS_ERR(tv_nexus->tvn_se_sess)) {
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
@@ -1901,11 +1915,11 @@ static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
return -ENOMEM;
|
|
|
}
|
|
|
se_sess = tv_nexus->tvn_se_sess;
|
|
|
- for (i = 0; i < TCM_VHOST_DEFAULT_TAGS; i++) {
|
|
|
- tv_cmd = &((struct tcm_vhost_cmd *)se_sess->sess_cmd_map)[i];
|
|
|
+ for (i = 0; i < VHOST_SCSI_DEFAULT_TAGS; i++) {
|
|
|
+ tv_cmd = &((struct vhost_scsi_cmd *)se_sess->sess_cmd_map)[i];
|
|
|
|
|
|
tv_cmd->tvc_sgl = kzalloc(sizeof(struct scatterlist) *
|
|
|
- TCM_VHOST_PREALLOC_SGLS, GFP_KERNEL);
|
|
|
+ VHOST_SCSI_PREALLOC_SGLS, GFP_KERNEL);
|
|
|
if (!tv_cmd->tvc_sgl) {
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
pr_err("Unable to allocate tv_cmd->tvc_sgl\n");
|
|
@@ -1913,7 +1927,7 @@ static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
}
|
|
|
|
|
|
tv_cmd->tvc_upages = kzalloc(sizeof(struct page *) *
|
|
|
- TCM_VHOST_PREALLOC_UPAGES, GFP_KERNEL);
|
|
|
+ VHOST_SCSI_PREALLOC_UPAGES, GFP_KERNEL);
|
|
|
if (!tv_cmd->tvc_upages) {
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
pr_err("Unable to allocate tv_cmd->tvc_upages\n");
|
|
@@ -1921,7 +1935,7 @@ static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
}
|
|
|
|
|
|
tv_cmd->tvc_prot_sgl = kzalloc(sizeof(struct scatterlist) *
|
|
|
- TCM_VHOST_PREALLOC_PROT_SGLS, GFP_KERNEL);
|
|
|
+ VHOST_SCSI_PREALLOC_PROT_SGLS, GFP_KERNEL);
|
|
|
if (!tv_cmd->tvc_prot_sgl) {
|
|
|
mutex_unlock(&tpg->tv_tpg_mutex);
|
|
|
pr_err("Unable to allocate tv_cmd->tvc_prot_sgl\n");
|
|
@@ -1930,7 +1944,7 @@ static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
}
|
|
|
/*
|
|
|
* Since we are running in 'demo mode' this call with generate a
|
|
|
- * struct se_node_acl for the tcm_vhost struct se_portal_group with
|
|
|
+ * struct se_node_acl for the vhost_scsi struct se_portal_group with
|
|
|
* the SCSI Initiator port name of the passed configfs group 'name'.
|
|
|
*/
|
|
|
tv_nexus->tvn_se_sess->se_node_acl = core_tpg_check_initiator_node_acl(
|
|
@@ -1953,16 +1967,16 @@ static int tcm_vhost_make_nexus(struct tcm_vhost_tpg *tpg,
|
|
|
return 0;
|
|
|
|
|
|
out:
|
|
|
- tcm_vhost_free_cmd_map_res(tv_nexus, se_sess);
|
|
|
+ vhost_scsi_free_cmd_map_res(tv_nexus, se_sess);
|
|
|
transport_free_session(se_sess);
|
|
|
kfree(tv_nexus);
|
|
|
return -ENOMEM;
|
|
|
}
|
|
|
|
|
|
-static int tcm_vhost_drop_nexus(struct tcm_vhost_tpg *tpg)
|
|
|
+static int vhost_scsi_drop_nexus(struct vhost_scsi_tpg *tpg)
|
|
|
{
|
|
|
struct se_session *se_sess;
|
|
|
- struct tcm_vhost_nexus *tv_nexus;
|
|
|
+ struct vhost_scsi_nexus *tv_nexus;
|
|
|
|
|
|
mutex_lock(&tpg->tv_tpg_mutex);
|
|
|
tv_nexus = tpg->tpg_nexus;
|
|
@@ -1994,10 +2008,10 @@ static int tcm_vhost_drop_nexus(struct tcm_vhost_tpg *tpg)
|
|
|
}
|
|
|
|
|
|
pr_debug("TCM_vhost_ConfigFS: Removing I_T Nexus to emulated"
|
|
|
- " %s Initiator Port: %s\n", tcm_vhost_dump_proto_id(tpg->tport),
|
|
|
+ " %s Initiator Port: %s\n", vhost_scsi_dump_proto_id(tpg->tport),
|
|
|
tv_nexus->tvn_se_sess->se_node_acl->initiatorname);
|
|
|
|
|
|
- tcm_vhost_free_cmd_map_res(tv_nexus, se_sess);
|
|
|
+ vhost_scsi_free_cmd_map_res(tv_nexus, se_sess);
|
|
|
/*
|
|
|
* Release the SCSI I_T Nexus to the emulated vhost Target Port
|
|
|
*/
|
|
@@ -2009,12 +2023,12 @@ static int tcm_vhost_drop_nexus(struct tcm_vhost_tpg *tpg)
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
-static ssize_t tcm_vhost_tpg_show_nexus(struct se_portal_group *se_tpg,
|
|
|
+static ssize_t vhost_scsi_tpg_show_nexus(struct se_portal_group *se_tpg,
|
|
|
char *page)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_nexus *tv_nexus;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_nexus *tv_nexus;
|
|
|
ssize_t ret;
|
|
|
|
|
|
mutex_lock(&tpg->tv_tpg_mutex);
|
|
@@ -2030,40 +2044,40 @@ static ssize_t tcm_vhost_tpg_show_nexus(struct se_portal_group *se_tpg,
|
|
|
return ret;
|
|
|
}
|
|
|
|
|
|
-static ssize_t tcm_vhost_tpg_store_nexus(struct se_portal_group *se_tpg,
|
|
|
+static ssize_t vhost_scsi_tpg_store_nexus(struct se_portal_group *se_tpg,
|
|
|
const char *page,
|
|
|
size_t count)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
- struct tcm_vhost_tport *tport_wwn = tpg->tport;
|
|
|
- unsigned char i_port[TCM_VHOST_NAMELEN], *ptr, *port_ptr;
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tport *tport_wwn = tpg->tport;
|
|
|
+ unsigned char i_port[VHOST_SCSI_NAMELEN], *ptr, *port_ptr;
|
|
|
int ret;
|
|
|
/*
|
|
|
* Shutdown the active I_T nexus if 'NULL' is passed..
|
|
|
*/
|
|
|
if (!strncmp(page, "NULL", 4)) {
|
|
|
- ret = tcm_vhost_drop_nexus(tpg);
|
|
|
+ ret = vhost_scsi_drop_nexus(tpg);
|
|
|
return (!ret) ? count : ret;
|
|
|
}
|
|
|
/*
|
|
|
* Otherwise make sure the passed virtual Initiator port WWN matches
|
|
|
- * the fabric protocol_id set in tcm_vhost_make_tport(), and call
|
|
|
- * tcm_vhost_make_nexus().
|
|
|
+ * the fabric protocol_id set in vhost_scsi_make_tport(), and call
|
|
|
+ * vhost_scsi_make_nexus().
|
|
|
*/
|
|
|
- if (strlen(page) >= TCM_VHOST_NAMELEN) {
|
|
|
+ if (strlen(page) >= VHOST_SCSI_NAMELEN) {
|
|
|
pr_err("Emulated NAA Sas Address: %s, exceeds"
|
|
|
- " max: %d\n", page, TCM_VHOST_NAMELEN);
|
|
|
+ " max: %d\n", page, VHOST_SCSI_NAMELEN);
|
|
|
return -EINVAL;
|
|
|
}
|
|
|
- snprintf(&i_port[0], TCM_VHOST_NAMELEN, "%s", page);
|
|
|
+ snprintf(&i_port[0], VHOST_SCSI_NAMELEN, "%s", page);
|
|
|
|
|
|
ptr = strstr(i_port, "naa.");
|
|
|
if (ptr) {
|
|
|
if (tport_wwn->tport_proto_id != SCSI_PROTOCOL_SAS) {
|
|
|
pr_err("Passed SAS Initiator Port %s does not"
|
|
|
" match target port protoid: %s\n", i_port,
|
|
|
- tcm_vhost_dump_proto_id(tport_wwn));
|
|
|
+ vhost_scsi_dump_proto_id(tport_wwn));
|
|
|
return -EINVAL;
|
|
|
}
|
|
|
port_ptr = &i_port[0];
|
|
@@ -2074,7 +2088,7 @@ static ssize_t tcm_vhost_tpg_store_nexus(struct se_portal_group *se_tpg,
|
|
|
if (tport_wwn->tport_proto_id != SCSI_PROTOCOL_FCP) {
|
|
|
pr_err("Passed FCP Initiator Port %s does not"
|
|
|
" match target port protoid: %s\n", i_port,
|
|
|
- tcm_vhost_dump_proto_id(tport_wwn));
|
|
|
+ vhost_scsi_dump_proto_id(tport_wwn));
|
|
|
return -EINVAL;
|
|
|
}
|
|
|
port_ptr = &i_port[3]; /* Skip over "fc." */
|
|
@@ -2085,7 +2099,7 @@ static ssize_t tcm_vhost_tpg_store_nexus(struct se_portal_group *se_tpg,
|
|
|
if (tport_wwn->tport_proto_id != SCSI_PROTOCOL_ISCSI) {
|
|
|
pr_err("Passed iSCSI Initiator Port %s does not"
|
|
|
" match target port protoid: %s\n", i_port,
|
|
|
- tcm_vhost_dump_proto_id(tport_wwn));
|
|
|
+ vhost_scsi_dump_proto_id(tport_wwn));
|
|
|
return -EINVAL;
|
|
|
}
|
|
|
port_ptr = &i_port[0];
|
|
@@ -2101,40 +2115,40 @@ check_newline:
|
|
|
if (i_port[strlen(i_port)-1] == '\n')
|
|
|
i_port[strlen(i_port)-1] = '\0';
|
|
|
|
|
|
- ret = tcm_vhost_make_nexus(tpg, port_ptr);
|
|
|
+ ret = vhost_scsi_make_nexus(tpg, port_ptr);
|
|
|
if (ret < 0)
|
|
|
return ret;
|
|
|
|
|
|
return count;
|
|
|
}
|
|
|
|
|
|
-TF_TPG_BASE_ATTR(tcm_vhost, nexus, S_IRUGO | S_IWUSR);
|
|
|
+TF_TPG_BASE_ATTR(vhost_scsi, nexus, S_IRUGO | S_IWUSR);
|
|
|
|
|
|
-static struct configfs_attribute *tcm_vhost_tpg_attrs[] = {
|
|
|
- &tcm_vhost_tpg_nexus.attr,
|
|
|
+static struct configfs_attribute *vhost_scsi_tpg_attrs[] = {
|
|
|
+ &vhost_scsi_tpg_nexus.attr,
|
|
|
NULL,
|
|
|
};
|
|
|
|
|
|
static struct se_portal_group *
|
|
|
-tcm_vhost_make_tpg(struct se_wwn *wwn,
|
|
|
+vhost_scsi_make_tpg(struct se_wwn *wwn,
|
|
|
struct config_group *group,
|
|
|
const char *name)
|
|
|
{
|
|
|
- struct tcm_vhost_tport *tport = container_of(wwn,
|
|
|
- struct tcm_vhost_tport, tport_wwn);
|
|
|
+ struct vhost_scsi_tport *tport = container_of(wwn,
|
|
|
+ struct vhost_scsi_tport, tport_wwn);
|
|
|
|
|
|
- struct tcm_vhost_tpg *tpg;
|
|
|
- unsigned long tpgt;
|
|
|
+ struct vhost_scsi_tpg *tpg;
|
|
|
+ u16 tpgt;
|
|
|
int ret;
|
|
|
|
|
|
if (strstr(name, "tpgt_") != name)
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
- if (kstrtoul(name + 5, 10, &tpgt) || tpgt > UINT_MAX)
|
|
|
+ if (kstrtou16(name + 5, 10, &tpgt) || tpgt >= VHOST_SCSI_MAX_TARGET)
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
- tpg = kzalloc(sizeof(struct tcm_vhost_tpg), GFP_KERNEL);
|
|
|
+ tpg = kzalloc(sizeof(struct vhost_scsi_tpg), GFP_KERNEL);
|
|
|
if (!tpg) {
|
|
|
- pr_err("Unable to allocate struct tcm_vhost_tpg");
|
|
|
+ pr_err("Unable to allocate struct vhost_scsi_tpg");
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
}
|
|
|
mutex_init(&tpg->tv_tpg_mutex);
|
|
@@ -2142,31 +2156,31 @@ tcm_vhost_make_tpg(struct se_wwn *wwn,
|
|
|
tpg->tport = tport;
|
|
|
tpg->tport_tpgt = tpgt;
|
|
|
|
|
|
- ret = core_tpg_register(&tcm_vhost_fabric_configfs->tf_ops, wwn,
|
|
|
+ ret = core_tpg_register(&vhost_scsi_fabric_configfs->tf_ops, wwn,
|
|
|
&tpg->se_tpg, tpg, TRANSPORT_TPG_TYPE_NORMAL);
|
|
|
if (ret < 0) {
|
|
|
kfree(tpg);
|
|
|
return NULL;
|
|
|
}
|
|
|
- mutex_lock(&tcm_vhost_mutex);
|
|
|
- list_add_tail(&tpg->tv_tpg_list, &tcm_vhost_list);
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_lock(&vhost_scsi_mutex);
|
|
|
+ list_add_tail(&tpg->tv_tpg_list, &vhost_scsi_list);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
|
|
|
return &tpg->se_tpg;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_drop_tpg(struct se_portal_group *se_tpg)
|
|
|
+static void vhost_scsi_drop_tpg(struct se_portal_group *se_tpg)
|
|
|
{
|
|
|
- struct tcm_vhost_tpg *tpg = container_of(se_tpg,
|
|
|
- struct tcm_vhost_tpg, se_tpg);
|
|
|
+ struct vhost_scsi_tpg *tpg = container_of(se_tpg,
|
|
|
+ struct vhost_scsi_tpg, se_tpg);
|
|
|
|
|
|
- mutex_lock(&tcm_vhost_mutex);
|
|
|
+ mutex_lock(&vhost_scsi_mutex);
|
|
|
list_del(&tpg->tv_tpg_list);
|
|
|
- mutex_unlock(&tcm_vhost_mutex);
|
|
|
+ mutex_unlock(&vhost_scsi_mutex);
|
|
|
/*
|
|
|
* Release the virtual I_T Nexus for this vhost TPG
|
|
|
*/
|
|
|
- tcm_vhost_drop_nexus(tpg);
|
|
|
+ vhost_scsi_drop_nexus(tpg);
|
|
|
/*
|
|
|
* Deregister the se_tpg from TCM..
|
|
|
*/
|
|
@@ -2175,21 +2189,21 @@ static void tcm_vhost_drop_tpg(struct se_portal_group *se_tpg)
|
|
|
}
|
|
|
|
|
|
static struct se_wwn *
|
|
|
-tcm_vhost_make_tport(struct target_fabric_configfs *tf,
|
|
|
+vhost_scsi_make_tport(struct target_fabric_configfs *tf,
|
|
|
struct config_group *group,
|
|
|
const char *name)
|
|
|
{
|
|
|
- struct tcm_vhost_tport *tport;
|
|
|
+ struct vhost_scsi_tport *tport;
|
|
|
char *ptr;
|
|
|
u64 wwpn = 0;
|
|
|
int off = 0;
|
|
|
|
|
|
- /* if (tcm_vhost_parse_wwn(name, &wwpn, 1) < 0)
|
|
|
+ /* if (vhost_scsi_parse_wwn(name, &wwpn, 1) < 0)
|
|
|
return ERR_PTR(-EINVAL); */
|
|
|
|
|
|
- tport = kzalloc(sizeof(struct tcm_vhost_tport), GFP_KERNEL);
|
|
|
+ tport = kzalloc(sizeof(struct vhost_scsi_tport), GFP_KERNEL);
|
|
|
if (!tport) {
|
|
|
- pr_err("Unable to allocate struct tcm_vhost_tport");
|
|
|
+ pr_err("Unable to allocate struct vhost_scsi_tport");
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
}
|
|
|
tport->tport_wwpn = wwpn;
|
|
@@ -2220,102 +2234,102 @@ tcm_vhost_make_tport(struct target_fabric_configfs *tf,
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
check_len:
|
|
|
- if (strlen(name) >= TCM_VHOST_NAMELEN) {
|
|
|
+ if (strlen(name) >= VHOST_SCSI_NAMELEN) {
|
|
|
pr_err("Emulated %s Address: %s, exceeds"
|
|
|
- " max: %d\n", name, tcm_vhost_dump_proto_id(tport),
|
|
|
- TCM_VHOST_NAMELEN);
|
|
|
+ " max: %d\n", name, vhost_scsi_dump_proto_id(tport),
|
|
|
+ VHOST_SCSI_NAMELEN);
|
|
|
kfree(tport);
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
}
|
|
|
- snprintf(&tport->tport_name[0], TCM_VHOST_NAMELEN, "%s", &name[off]);
|
|
|
+ snprintf(&tport->tport_name[0], VHOST_SCSI_NAMELEN, "%s", &name[off]);
|
|
|
|
|
|
pr_debug("TCM_VHost_ConfigFS: Allocated emulated Target"
|
|
|
- " %s Address: %s\n", tcm_vhost_dump_proto_id(tport), name);
|
|
|
+ " %s Address: %s\n", vhost_scsi_dump_proto_id(tport), name);
|
|
|
|
|
|
return &tport->tport_wwn;
|
|
|
}
|
|
|
|
|
|
-static void tcm_vhost_drop_tport(struct se_wwn *wwn)
|
|
|
+static void vhost_scsi_drop_tport(struct se_wwn *wwn)
|
|
|
{
|
|
|
- struct tcm_vhost_tport *tport = container_of(wwn,
|
|
|
- struct tcm_vhost_tport, tport_wwn);
|
|
|
+ struct vhost_scsi_tport *tport = container_of(wwn,
|
|
|
+ struct vhost_scsi_tport, tport_wwn);
|
|
|
|
|
|
pr_debug("TCM_VHost_ConfigFS: Deallocating emulated Target"
|
|
|
- " %s Address: %s\n", tcm_vhost_dump_proto_id(tport),
|
|
|
+ " %s Address: %s\n", vhost_scsi_dump_proto_id(tport),
|
|
|
tport->tport_name);
|
|
|
|
|
|
kfree(tport);
|
|
|
}
|
|
|
|
|
|
static ssize_t
|
|
|
-tcm_vhost_wwn_show_attr_version(struct target_fabric_configfs *tf,
|
|
|
+vhost_scsi_wwn_show_attr_version(struct target_fabric_configfs *tf,
|
|
|
char *page)
|
|
|
{
|
|
|
return sprintf(page, "TCM_VHOST fabric module %s on %s/%s"
|
|
|
- "on "UTS_RELEASE"\n", TCM_VHOST_VERSION, utsname()->sysname,
|
|
|
+ "on "UTS_RELEASE"\n", VHOST_SCSI_VERSION, utsname()->sysname,
|
|
|
utsname()->machine);
|
|
|
}
|
|
|
|
|
|
-TF_WWN_ATTR_RO(tcm_vhost, version);
|
|
|
+TF_WWN_ATTR_RO(vhost_scsi, version);
|
|
|
|
|
|
-static struct configfs_attribute *tcm_vhost_wwn_attrs[] = {
|
|
|
- &tcm_vhost_wwn_version.attr,
|
|
|
+static struct configfs_attribute *vhost_scsi_wwn_attrs[] = {
|
|
|
+ &vhost_scsi_wwn_version.attr,
|
|
|
NULL,
|
|
|
};
|
|
|
|
|
|
-static struct target_core_fabric_ops tcm_vhost_ops = {
|
|
|
- .get_fabric_name = tcm_vhost_get_fabric_name,
|
|
|
- .get_fabric_proto_ident = tcm_vhost_get_fabric_proto_ident,
|
|
|
- .tpg_get_wwn = tcm_vhost_get_fabric_wwn,
|
|
|
- .tpg_get_tag = tcm_vhost_get_tag,
|
|
|
- .tpg_get_default_depth = tcm_vhost_get_default_depth,
|
|
|
- .tpg_get_pr_transport_id = tcm_vhost_get_pr_transport_id,
|
|
|
- .tpg_get_pr_transport_id_len = tcm_vhost_get_pr_transport_id_len,
|
|
|
- .tpg_parse_pr_out_transport_id = tcm_vhost_parse_pr_out_transport_id,
|
|
|
- .tpg_check_demo_mode = tcm_vhost_check_true,
|
|
|
- .tpg_check_demo_mode_cache = tcm_vhost_check_true,
|
|
|
- .tpg_check_demo_mode_write_protect = tcm_vhost_check_false,
|
|
|
- .tpg_check_prod_mode_write_protect = tcm_vhost_check_false,
|
|
|
- .tpg_alloc_fabric_acl = tcm_vhost_alloc_fabric_acl,
|
|
|
- .tpg_release_fabric_acl = tcm_vhost_release_fabric_acl,
|
|
|
- .tpg_get_inst_index = tcm_vhost_tpg_get_inst_index,
|
|
|
- .release_cmd = tcm_vhost_release_cmd,
|
|
|
+static struct target_core_fabric_ops vhost_scsi_ops = {
|
|
|
+ .get_fabric_name = vhost_scsi_get_fabric_name,
|
|
|
+ .get_fabric_proto_ident = vhost_scsi_get_fabric_proto_ident,
|
|
|
+ .tpg_get_wwn = vhost_scsi_get_fabric_wwn,
|
|
|
+ .tpg_get_tag = vhost_scsi_get_tpgt,
|
|
|
+ .tpg_get_default_depth = vhost_scsi_get_default_depth,
|
|
|
+ .tpg_get_pr_transport_id = vhost_scsi_get_pr_transport_id,
|
|
|
+ .tpg_get_pr_transport_id_len = vhost_scsi_get_pr_transport_id_len,
|
|
|
+ .tpg_parse_pr_out_transport_id = vhost_scsi_parse_pr_out_transport_id,
|
|
|
+ .tpg_check_demo_mode = vhost_scsi_check_true,
|
|
|
+ .tpg_check_demo_mode_cache = vhost_scsi_check_true,
|
|
|
+ .tpg_check_demo_mode_write_protect = vhost_scsi_check_false,
|
|
|
+ .tpg_check_prod_mode_write_protect = vhost_scsi_check_false,
|
|
|
+ .tpg_alloc_fabric_acl = vhost_scsi_alloc_fabric_acl,
|
|
|
+ .tpg_release_fabric_acl = vhost_scsi_release_fabric_acl,
|
|
|
+ .tpg_get_inst_index = vhost_scsi_tpg_get_inst_index,
|
|
|
+ .release_cmd = vhost_scsi_release_cmd,
|
|
|
.check_stop_free = vhost_scsi_check_stop_free,
|
|
|
- .shutdown_session = tcm_vhost_shutdown_session,
|
|
|
- .close_session = tcm_vhost_close_session,
|
|
|
- .sess_get_index = tcm_vhost_sess_get_index,
|
|
|
+ .shutdown_session = vhost_scsi_shutdown_session,
|
|
|
+ .close_session = vhost_scsi_close_session,
|
|
|
+ .sess_get_index = vhost_scsi_sess_get_index,
|
|
|
.sess_get_initiator_sid = NULL,
|
|
|
- .write_pending = tcm_vhost_write_pending,
|
|
|
- .write_pending_status = tcm_vhost_write_pending_status,
|
|
|
- .set_default_node_attributes = tcm_vhost_set_default_node_attrs,
|
|
|
- .get_task_tag = tcm_vhost_get_task_tag,
|
|
|
- .get_cmd_state = tcm_vhost_get_cmd_state,
|
|
|
- .queue_data_in = tcm_vhost_queue_data_in,
|
|
|
- .queue_status = tcm_vhost_queue_status,
|
|
|
- .queue_tm_rsp = tcm_vhost_queue_tm_rsp,
|
|
|
- .aborted_task = tcm_vhost_aborted_task,
|
|
|
+ .write_pending = vhost_scsi_write_pending,
|
|
|
+ .write_pending_status = vhost_scsi_write_pending_status,
|
|
|
+ .set_default_node_attributes = vhost_scsi_set_default_node_attrs,
|
|
|
+ .get_task_tag = vhost_scsi_get_task_tag,
|
|
|
+ .get_cmd_state = vhost_scsi_get_cmd_state,
|
|
|
+ .queue_data_in = vhost_scsi_queue_data_in,
|
|
|
+ .queue_status = vhost_scsi_queue_status,
|
|
|
+ .queue_tm_rsp = vhost_scsi_queue_tm_rsp,
|
|
|
+ .aborted_task = vhost_scsi_aborted_task,
|
|
|
/*
|
|
|
* Setup callers for generic logic in target_core_fabric_configfs.c
|
|
|
*/
|
|
|
- .fabric_make_wwn = tcm_vhost_make_tport,
|
|
|
- .fabric_drop_wwn = tcm_vhost_drop_tport,
|
|
|
- .fabric_make_tpg = tcm_vhost_make_tpg,
|
|
|
- .fabric_drop_tpg = tcm_vhost_drop_tpg,
|
|
|
- .fabric_post_link = tcm_vhost_port_link,
|
|
|
- .fabric_pre_unlink = tcm_vhost_port_unlink,
|
|
|
+ .fabric_make_wwn = vhost_scsi_make_tport,
|
|
|
+ .fabric_drop_wwn = vhost_scsi_drop_tport,
|
|
|
+ .fabric_make_tpg = vhost_scsi_make_tpg,
|
|
|
+ .fabric_drop_tpg = vhost_scsi_drop_tpg,
|
|
|
+ .fabric_post_link = vhost_scsi_port_link,
|
|
|
+ .fabric_pre_unlink = vhost_scsi_port_unlink,
|
|
|
.fabric_make_np = NULL,
|
|
|
.fabric_drop_np = NULL,
|
|
|
- .fabric_make_nodeacl = tcm_vhost_make_nodeacl,
|
|
|
- .fabric_drop_nodeacl = tcm_vhost_drop_nodeacl,
|
|
|
+ .fabric_make_nodeacl = vhost_scsi_make_nodeacl,
|
|
|
+ .fabric_drop_nodeacl = vhost_scsi_drop_nodeacl,
|
|
|
};
|
|
|
|
|
|
-static int tcm_vhost_register_configfs(void)
|
|
|
+static int vhost_scsi_register_configfs(void)
|
|
|
{
|
|
|
struct target_fabric_configfs *fabric;
|
|
|
int ret;
|
|
|
|
|
|
- pr_debug("TCM_VHOST fabric module %s on %s/%s"
|
|
|
- " on "UTS_RELEASE"\n", TCM_VHOST_VERSION, utsname()->sysname,
|
|
|
+ pr_debug("vhost-scsi fabric module %s on %s/%s"
|
|
|
+ " on "UTS_RELEASE"\n", VHOST_SCSI_VERSION, utsname()->sysname,
|
|
|
utsname()->machine);
|
|
|
/*
|
|
|
* Register the top level struct config_item_type with TCM core
|
|
@@ -2326,14 +2340,14 @@ static int tcm_vhost_register_configfs(void)
|
|
|
return PTR_ERR(fabric);
|
|
|
}
|
|
|
/*
|
|
|
- * Setup fabric->tf_ops from our local tcm_vhost_ops
|
|
|
+ * Setup fabric->tf_ops from our local vhost_scsi_ops
|
|
|
*/
|
|
|
- fabric->tf_ops = tcm_vhost_ops;
|
|
|
+ fabric->tf_ops = vhost_scsi_ops;
|
|
|
/*
|
|
|
* Setup default attribute lists for various fabric->tf_cit_tmpl
|
|
|
*/
|
|
|
- fabric->tf_cit_tmpl.tfc_wwn_cit.ct_attrs = tcm_vhost_wwn_attrs;
|
|
|
- fabric->tf_cit_tmpl.tfc_tpg_base_cit.ct_attrs = tcm_vhost_tpg_attrs;
|
|
|
+ fabric->tf_cit_tmpl.tfc_wwn_cit.ct_attrs = vhost_scsi_wwn_attrs;
|
|
|
+ fabric->tf_cit_tmpl.tfc_tpg_base_cit.ct_attrs = vhost_scsi_tpg_attrs;
|
|
|
fabric->tf_cit_tmpl.tfc_tpg_attrib_cit.ct_attrs = NULL;
|
|
|
fabric->tf_cit_tmpl.tfc_tpg_param_cit.ct_attrs = NULL;
|
|
|
fabric->tf_cit_tmpl.tfc_tpg_np_base_cit.ct_attrs = NULL;
|
|
@@ -2353,37 +2367,37 @@ static int tcm_vhost_register_configfs(void)
|
|
|
/*
|
|
|
* Setup our local pointer to *fabric
|
|
|
*/
|
|
|
- tcm_vhost_fabric_configfs = fabric;
|
|
|
- pr_debug("TCM_VHOST[0] - Set fabric -> tcm_vhost_fabric_configfs\n");
|
|
|
+ vhost_scsi_fabric_configfs = fabric;
|
|
|
+ pr_debug("TCM_VHOST[0] - Set fabric -> vhost_scsi_fabric_configfs\n");
|
|
|
return 0;
|
|
|
};
|
|
|
|
|
|
-static void tcm_vhost_deregister_configfs(void)
|
|
|
+static void vhost_scsi_deregister_configfs(void)
|
|
|
{
|
|
|
- if (!tcm_vhost_fabric_configfs)
|
|
|
+ if (!vhost_scsi_fabric_configfs)
|
|
|
return;
|
|
|
|
|
|
- target_fabric_configfs_deregister(tcm_vhost_fabric_configfs);
|
|
|
- tcm_vhost_fabric_configfs = NULL;
|
|
|
- pr_debug("TCM_VHOST[0] - Cleared tcm_vhost_fabric_configfs\n");
|
|
|
+ target_fabric_configfs_deregister(vhost_scsi_fabric_configfs);
|
|
|
+ vhost_scsi_fabric_configfs = NULL;
|
|
|
+ pr_debug("TCM_VHOST[0] - Cleared vhost_scsi_fabric_configfs\n");
|
|
|
};
|
|
|
|
|
|
-static int __init tcm_vhost_init(void)
|
|
|
+static int __init vhost_scsi_init(void)
|
|
|
{
|
|
|
int ret = -ENOMEM;
|
|
|
/*
|
|
|
* Use our own dedicated workqueue for submitting I/O into
|
|
|
* target core to avoid contention within system_wq.
|
|
|
*/
|
|
|
- tcm_vhost_workqueue = alloc_workqueue("tcm_vhost", 0, 0);
|
|
|
- if (!tcm_vhost_workqueue)
|
|
|
+ vhost_scsi_workqueue = alloc_workqueue("vhost_scsi", 0, 0);
|
|
|
+ if (!vhost_scsi_workqueue)
|
|
|
goto out;
|
|
|
|
|
|
ret = vhost_scsi_register();
|
|
|
if (ret < 0)
|
|
|
goto out_destroy_workqueue;
|
|
|
|
|
|
- ret = tcm_vhost_register_configfs();
|
|
|
+ ret = vhost_scsi_register_configfs();
|
|
|
if (ret < 0)
|
|
|
goto out_vhost_scsi_deregister;
|
|
|
|
|
@@ -2392,20 +2406,20 @@ static int __init tcm_vhost_init(void)
|
|
|
out_vhost_scsi_deregister:
|
|
|
vhost_scsi_deregister();
|
|
|
out_destroy_workqueue:
|
|
|
- destroy_workqueue(tcm_vhost_workqueue);
|
|
|
+ destroy_workqueue(vhost_scsi_workqueue);
|
|
|
out:
|
|
|
return ret;
|
|
|
};
|
|
|
|
|
|
-static void tcm_vhost_exit(void)
|
|
|
+static void vhost_scsi_exit(void)
|
|
|
{
|
|
|
- tcm_vhost_deregister_configfs();
|
|
|
+ vhost_scsi_deregister_configfs();
|
|
|
vhost_scsi_deregister();
|
|
|
- destroy_workqueue(tcm_vhost_workqueue);
|
|
|
+ destroy_workqueue(vhost_scsi_workqueue);
|
|
|
};
|
|
|
|
|
|
MODULE_DESCRIPTION("VHOST_SCSI series fabric driver");
|
|
|
MODULE_ALIAS("tcm_vhost");
|
|
|
MODULE_LICENSE("GPL");
|
|
|
-module_init(tcm_vhost_init);
|
|
|
-module_exit(tcm_vhost_exit);
|
|
|
+module_init(vhost_scsi_init);
|
|
|
+module_exit(vhost_scsi_exit);
|