]> git.karo-electronics.de Git - karo-tx-linux.git/blobdiff - drivers/target/target_core_configfs.c
Merge branch 'irq-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel...
[karo-tx-linux.git] / drivers / target / target_core_configfs.c
index e7b0430a0575d0403dbb38b0fd4d41df1ccce79d..c2e9fea90b4a4bc16a0384d79fa9684c9f4176e0 100644 (file)
@@ -41,7 +41,6 @@
 #include <target/target_core_backend.h>
 #include <target/target_core_fabric.h>
 #include <target/target_core_fabric_configfs.h>
-#include <target/target_core_configfs.h>
 #include <target/configfs_macros.h>
 
 #include "target_core_internal.h"
 #include "target_core_xcopy.h"
 
 #define TB_CIT_SETUP(_name, _item_ops, _group_ops, _attrs)             \
-static void target_core_setup_##_name##_cit(struct se_subsystem_api *sa) \
+static void target_core_setup_##_name##_cit(struct target_backend *tb) \
 {                                                                      \
-       struct target_backend_cits *tbc = &sa->tb_cits;                 \
-       struct config_item_type *cit = &tbc->tb_##_name##_cit;          \
+       struct config_item_type *cit = &tb->tb_##_name##_cit;           \
                                                                        \
        cit->ct_item_ops = _item_ops;                                   \
        cit->ct_group_ops = _group_ops;                                 \
        cit->ct_attrs = _attrs;                                         \
-       cit->ct_owner = sa->owner;                                      \
+       cit->ct_owner = tb->ops->owner;                                 \
+       pr_debug("Setup generic %s\n", __stringify(_name));             \
+}
+
+#define TB_CIT_SETUP_DRV(_name, _item_ops, _group_ops)                 \
+static void target_core_setup_##_name##_cit(struct target_backend *tb) \
+{                                                                      \
+       struct config_item_type *cit = &tb->tb_##_name##_cit;           \
+                                                                       \
+       cit->ct_item_ops = _item_ops;                                   \
+       cit->ct_group_ops = _group_ops;                                 \
+       cit->ct_attrs = tb->ops->tb_##_name##_attrs;                    \
+       cit->ct_owner = tb->ops->owner;                                 \
        pr_debug("Setup generic %s\n", __stringify(_name));             \
 }
 
@@ -92,7 +102,7 @@ static ssize_t target_core_attr_show(struct config_item *item,
                                      char *page)
 {
        return sprintf(page, "Target Engine Core ConfigFS Infrastructure %s"
-               " on %s/%s on "UTS_RELEASE"\n", TARGET_CORE_CONFIGFS_VERSION,
+               " on %s/%s on "UTS_RELEASE"\n", TARGET_CORE_VERSION,
                utsname()->sysname, utsname()->machine);
 }
 
@@ -116,7 +126,7 @@ static struct target_fabric_configfs *target_core_get_fabric(
 
        mutex_lock(&g_tf_lock);
        list_for_each_entry(tf, &g_tf_list, tf_list) {
-               if (!strcmp(tf->tf_name, name)) {
+               if (!strcmp(tf->tf_ops->name, name)) {
                        atomic_inc(&tf->tf_access_cnt);
                        mutex_unlock(&g_tf_lock);
                        return tf;
@@ -193,29 +203,24 @@ static struct config_group *target_core_register_fabric(
                return ERR_PTR(-EINVAL);
        }
        pr_debug("Target_Core_ConfigFS: REGISTER -> Located fabric:"
-                       " %s\n", tf->tf_name);
+                       " %s\n", tf->tf_ops->name);
        /*
         * On a successful target_core_get_fabric() look, the returned
         * struct target_fabric_configfs *tf will contain a usage reference.
         */
        pr_debug("Target_Core_ConfigFS: REGISTER tfc_wwn_cit -> %p\n",
-                       &tf->tf_cit_tmpl.tfc_wwn_cit);
+                       &tf->tf_wwn_cit);
 
        tf->tf_group.default_groups = tf->tf_default_groups;
        tf->tf_group.default_groups[0] = &tf->tf_disc_group;
        tf->tf_group.default_groups[1] = NULL;
 
-       config_group_init_type_name(&tf->tf_group, name,
-                       &tf->tf_cit_tmpl.tfc_wwn_cit);
+       config_group_init_type_name(&tf->tf_group, name, &tf->tf_wwn_cit);
        config_group_init_type_name(&tf->tf_disc_group, "discovery_auth",
-                       &tf->tf_cit_tmpl.tfc_discovery_cit);
+                       &tf->tf_discovery_cit);
 
        pr_debug("Target_Core_ConfigFS: REGISTER -> Allocated Fabric:"
                        " %s\n", tf->tf_group.cg_item.ci_name);
-       tf->tf_fabric = &tf->tf_group.cg_item;
-       pr_debug("Target_Core_ConfigFS: REGISTER -> Set tf->tf_fabric"
-                       " for %s\n", name);
-
        return &tf->tf_group;
 }
 
@@ -236,13 +241,9 @@ static void target_core_deregister_fabric(
                " tf list\n", config_item_name(item));
 
        pr_debug("Target_Core_ConfigFS: DEREGISTER -> located fabric:"
-                       " %s\n", tf->tf_name);
+                       " %s\n", tf->tf_ops->name);
        atomic_dec(&tf->tf_access_cnt);
 
-       pr_debug("Target_Core_ConfigFS: DEREGISTER -> Releasing"
-                       " tf->tf_fabric for %s\n", tf->tf_name);
-       tf->tf_fabric = NULL;
-
        pr_debug("Target_Core_ConfigFS: DEREGISTER -> Releasing ci"
                        " %s\n", config_item_name(item));
 
@@ -318,10 +319,6 @@ static int target_fabric_tf_ops_check(const struct target_core_fabric_ops *tfo)
                pr_err("Missing tfo->get_fabric_name()\n");
                return -EINVAL;
        }
-       if (!tfo->get_fabric_proto_ident) {
-               pr_err("Missing tfo->get_fabric_proto_ident()\n");
-               return -EINVAL;
-       }
        if (!tfo->tpg_get_wwn) {
                pr_err("Missing tfo->tpg_get_wwn()\n");
                return -EINVAL;
@@ -330,18 +327,6 @@ static int target_fabric_tf_ops_check(const struct target_core_fabric_ops *tfo)
                pr_err("Missing tfo->tpg_get_tag()\n");
                return -EINVAL;
        }
-       if (!tfo->tpg_get_default_depth) {
-               pr_err("Missing tfo->tpg_get_default_depth()\n");
-               return -EINVAL;
-       }
-       if (!tfo->tpg_get_pr_transport_id) {
-               pr_err("Missing tfo->tpg_get_pr_transport_id()\n");
-               return -EINVAL;
-       }
-       if (!tfo->tpg_get_pr_transport_id_len) {
-               pr_err("Missing tfo->tpg_get_pr_transport_id_len()\n");
-               return -EINVAL;
-       }
        if (!tfo->tpg_check_demo_mode) {
                pr_err("Missing tfo->tpg_check_demo_mode()\n");
                return -EINVAL;
@@ -358,14 +343,6 @@ static int target_fabric_tf_ops_check(const struct target_core_fabric_ops *tfo)
                pr_err("Missing tfo->tpg_check_prod_mode_write_protect()\n");
                return -EINVAL;
        }
-       if (!tfo->tpg_alloc_fabric_acl) {
-               pr_err("Missing tfo->tpg_alloc_fabric_acl()\n");
-               return -EINVAL;
-       }
-       if (!tfo->tpg_release_fabric_acl) {
-               pr_err("Missing tfo->tpg_release_fabric_acl()\n");
-               return -EINVAL;
-       }
        if (!tfo->tpg_get_inst_index) {
                pr_err("Missing tfo->tpg_get_inst_index()\n");
                return -EINVAL;
@@ -398,10 +375,6 @@ static int target_fabric_tf_ops_check(const struct target_core_fabric_ops *tfo)
                pr_err("Missing tfo->set_default_node_attributes()\n");
                return -EINVAL;
        }
-       if (!tfo->get_task_tag) {
-               pr_err("Missing tfo->get_task_tag()\n");
-               return -EINVAL;
-       }
        if (!tfo->get_cmd_state) {
                pr_err("Missing tfo->get_cmd_state()\n");
                return -EINVAL;
@@ -464,15 +437,7 @@ int target_register_template(const struct target_core_fabric_ops *fo)
 
        INIT_LIST_HEAD(&tf->tf_list);
        atomic_set(&tf->tf_access_cnt, 0);
-
-       /*
-        * Setup the default generic struct config_item_type's (cits) in
-        * struct target_fabric_configfs->tf_cit_tmpl
-        */
-       tf->tf_module = fo->module;
-       snprintf(tf->tf_name, TARGET_FABRIC_NAME_SIZE, "%s", fo->name);
-
-       tf->tf_ops = *fo;
+       tf->tf_ops = fo;
        target_fabric_setup_cits(tf);
 
        mutex_lock(&g_tf_lock);
@@ -489,7 +454,7 @@ void target_unregister_template(const struct target_core_fabric_ops *fo)
 
        mutex_lock(&g_tf_lock);
        list_for_each_entry(t, &g_tf_list, tf_list) {
-               if (!strcmp(t->tf_name, fo->name)) {
+               if (!strcmp(t->tf_ops->name, fo->name)) {
                        BUG_ON(atomic_read(&t->tf_access_cnt));
                        list_del(&t->tf_list);
                        kfree(t);
@@ -505,16 +470,605 @@ EXPORT_SYMBOL(target_unregister_template);
 //############################################################################*/
 
 /* Start functions for struct config_item_type tb_dev_attrib_cit */
+#define DEF_TB_DEV_ATTRIB_SHOW(_name)                                  \
+static ssize_t show_##_name(struct se_dev_attrib *da, char *page)      \
+{                                                                      \
+       return snprintf(page, PAGE_SIZE, "%u\n", da->_name);            \
+}
+
+DEF_TB_DEV_ATTRIB_SHOW(emulate_model_alias);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_dpo);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_fua_write);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_fua_read);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_write_cache);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_ua_intlck_ctrl);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_tas);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_tpu);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_tpws);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_caw);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_3pc);
+DEF_TB_DEV_ATTRIB_SHOW(pi_prot_type);
+DEF_TB_DEV_ATTRIB_SHOW(hw_pi_prot_type);
+DEF_TB_DEV_ATTRIB_SHOW(pi_prot_format);
+DEF_TB_DEV_ATTRIB_SHOW(enforce_pr_isids);
+DEF_TB_DEV_ATTRIB_SHOW(is_nonrot);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_rest_reord);
+DEF_TB_DEV_ATTRIB_SHOW(force_pr_aptpl);
+DEF_TB_DEV_ATTRIB_SHOW(hw_block_size);
+DEF_TB_DEV_ATTRIB_SHOW(block_size);
+DEF_TB_DEV_ATTRIB_SHOW(hw_max_sectors);
+DEF_TB_DEV_ATTRIB_SHOW(optimal_sectors);
+DEF_TB_DEV_ATTRIB_SHOW(hw_queue_depth);
+DEF_TB_DEV_ATTRIB_SHOW(queue_depth);
+DEF_TB_DEV_ATTRIB_SHOW(max_unmap_lba_count);
+DEF_TB_DEV_ATTRIB_SHOW(max_unmap_block_desc_count);
+DEF_TB_DEV_ATTRIB_SHOW(unmap_granularity);
+DEF_TB_DEV_ATTRIB_SHOW(unmap_granularity_alignment);
+DEF_TB_DEV_ATTRIB_SHOW(max_write_same_len);
+
+#define DEF_TB_DEV_ATTRIB_STORE_U32(_name)                             \
+static ssize_t store_##_name(struct se_dev_attrib *da, const char *page,\
+               size_t count)                                           \
+{                                                                      \
+       u32 val;                                                        \
+       int ret;                                                        \
+                                                                       \
+       ret = kstrtou32(page, 0, &val);                                 \
+       if (ret < 0)                                                    \
+               return ret;                                             \
+       da->_name = val;                                                \
+       return count;                                                   \
+}
+
+DEF_TB_DEV_ATTRIB_STORE_U32(max_unmap_lba_count);
+DEF_TB_DEV_ATTRIB_STORE_U32(max_unmap_block_desc_count);
+DEF_TB_DEV_ATTRIB_STORE_U32(unmap_granularity);
+DEF_TB_DEV_ATTRIB_STORE_U32(unmap_granularity_alignment);
+DEF_TB_DEV_ATTRIB_STORE_U32(max_write_same_len);
+
+#define DEF_TB_DEV_ATTRIB_STORE_BOOL(_name)                            \
+static ssize_t store_##_name(struct se_dev_attrib *da, const char *page,\
+               size_t count)                                           \
+{                                                                      \
+       bool flag;                                                      \
+       int ret;                                                        \
+                                                                       \
+       ret = strtobool(page, &flag);                                   \
+       if (ret < 0)                                                    \
+               return ret;                                             \
+       da->_name = flag;                                               \
+       return count;                                                   \
+}
+
+DEF_TB_DEV_ATTRIB_STORE_BOOL(emulate_fua_write);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(emulate_caw);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(emulate_3pc);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(enforce_pr_isids);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(is_nonrot);
+
+#define DEF_TB_DEV_ATTRIB_STORE_STUB(_name)                            \
+static ssize_t store_##_name(struct se_dev_attrib *da, const char *page,\
+               size_t count)                                           \
+{                                                                      \
+       printk_once(KERN_WARNING                                        \
+               "ignoring deprecated ##_name## attribute\n");   \
+       return count;                                                   \
+}
+
+DEF_TB_DEV_ATTRIB_STORE_STUB(emulate_dpo);
+DEF_TB_DEV_ATTRIB_STORE_STUB(emulate_fua_read);
+
+static void dev_set_t10_wwn_model_alias(struct se_device *dev)
+{
+       const char *configname;
+
+       configname = config_item_name(&dev->dev_group.cg_item);
+       if (strlen(configname) >= 16) {
+               pr_warn("dev[%p]: Backstore name '%s' is too long for "
+                       "INQUIRY_MODEL, truncating to 16 bytes\n", dev,
+                       configname);
+       }
+       snprintf(&dev->t10_wwn.model[0], 16, "%s", configname);
+}
+
+static ssize_t store_emulate_model_alias(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       struct se_device *dev = da->da_dev;
+       bool flag;
+       int ret;
+
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to change model alias"
+                       " while export_count is %d\n",
+                       dev, dev->export_count);
+               return -EINVAL;
+       }
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag) {
+               dev_set_t10_wwn_model_alias(dev);
+       } else {
+               strncpy(&dev->t10_wwn.model[0],
+                       dev->transport->inquiry_prod, 16);
+       }
+       da->emulate_model_alias = flag;
+       return count;
+}
+
+static ssize_t store_emulate_write_cache(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag && da->da_dev->transport->get_write_cache) {
+               pr_err("emulate_write_cache not supported for this device\n");
+               return -EINVAL;
+       }
+
+       da->emulate_write_cache = flag;
+       pr_debug("dev[%p]: SE Device WRITE_CACHE_EMULATION flag: %d\n",
+                       da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_emulate_ua_intlck_ctrl(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (val != 0 && val != 1 && val != 2) {
+               pr_err("Illegal value %d\n", val);
+               return -EINVAL;
+       }
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device"
+                       " UA_INTRLCK_CTRL while export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+       da->emulate_ua_intlck_ctrl = val;
+       pr_debug("dev[%p]: SE Device UA_INTRLCK_CTRL flag: %d\n",
+               da->da_dev, val);
+       return count;
+}
+
+static ssize_t store_emulate_tas(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device TAS while"
+                       " export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+       da->emulate_tas = flag;
+       pr_debug("dev[%p]: SE Device TASK_ABORTED status bit: %s\n",
+               da->da_dev, flag ? "Enabled" : "Disabled");
+
+       return count;
+}
+
+static ssize_t store_emulate_tpu(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       /*
+        * We expect this value to be non-zero when generic Block Layer
+        * Discard supported is detected iblock_create_virtdevice().
+        */
+       if (flag && !da->max_unmap_block_desc_count) {
+               pr_err("Generic Block Discard not supported\n");
+               return -ENOSYS;
+       }
+
+       da->emulate_tpu = flag;
+       pr_debug("dev[%p]: SE Device Thin Provisioning UNMAP bit: %d\n",
+               da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_emulate_tpws(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       /*
+        * We expect this value to be non-zero when generic Block Layer
+        * Discard supported is detected iblock_create_virtdevice().
+        */
+       if (flag && !da->max_unmap_block_desc_count) {
+               pr_err("Generic Block Discard not supported\n");
+               return -ENOSYS;
+       }
+
+       da->emulate_tpws = flag;
+       pr_debug("dev[%p]: SE Device Thin Provisioning WRITE_SAME: %d\n",
+                               da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_pi_prot_type(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       int old_prot = da->pi_prot_type, ret;
+       struct se_device *dev = da->da_dev;
+       u32 flag;
+
+       ret = kstrtou32(page, 0, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag != 0 && flag != 1 && flag != 2 && flag != 3) {
+               pr_err("Illegal value %d for pi_prot_type\n", flag);
+               return -EINVAL;
+       }
+       if (flag == 2) {
+               pr_err("DIF TYPE2 protection currently not supported\n");
+               return -ENOSYS;
+       }
+       if (da->hw_pi_prot_type) {
+               pr_warn("DIF protection enabled on underlying hardware,"
+                       " ignoring\n");
+               return count;
+       }
+       if (!dev->transport->init_prot || !dev->transport->free_prot) {
+               /* 0 is only allowed value for non-supporting backends */
+               if (flag == 0)
+                       return count;
+
+               pr_err("DIF protection not supported by backend: %s\n",
+                      dev->transport->name);
+               return -ENOSYS;
+       }
+       if (!(dev->dev_flags & DF_CONFIGURED)) {
+               pr_err("DIF protection requires device to be configured\n");
+               return -ENODEV;
+       }
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device PROT type while"
+                      " export_count is %d\n", dev, dev->export_count);
+               return -EINVAL;
+       }
+
+       da->pi_prot_type = flag;
+
+       if (flag && !old_prot) {
+               ret = dev->transport->init_prot(dev);
+               if (ret) {
+                       da->pi_prot_type = old_prot;
+                       return ret;
+               }
+
+       } else if (!flag && old_prot) {
+               dev->transport->free_prot(dev);
+       }
+
+       pr_debug("dev[%p]: SE Device Protection Type: %d\n", dev, flag);
+       return count;
+}
+
+static ssize_t store_pi_prot_format(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       struct se_device *dev = da->da_dev;
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (!flag)
+               return count;
+
+       if (!dev->transport->format_prot) {
+               pr_err("DIF protection format not supported by backend %s\n",
+                      dev->transport->name);
+               return -ENOSYS;
+       }
+       if (!(dev->dev_flags & DF_CONFIGURED)) {
+               pr_err("DIF protection format requires device to be configured\n");
+               return -ENODEV;
+       }
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to format SE Device PROT type while"
+                      " export_count is %d\n", dev, dev->export_count);
+               return -EINVAL;
+       }
+
+       ret = dev->transport->format_prot(dev);
+       if (ret)
+               return ret;
+
+       pr_debug("dev[%p]: SE Device Protection Format complete\n", dev);
+       return count;
+}
+
+static ssize_t store_force_pr_aptpl(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to set force_pr_aptpl while"
+                      " export_count is %d\n",
+                      da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+
+       da->force_pr_aptpl = flag;
+       pr_debug("dev[%p]: SE Device force_pr_aptpl: %d\n", da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_emulate_rest_reord(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag != 0) {
+               printk(KERN_ERR "dev[%p]: SE Device emulation of restricted"
+                       " reordering not implemented\n", da->da_dev);
+               return -ENOSYS;
+       }
+       da->emulate_rest_reord = flag;
+       pr_debug("dev[%p]: SE Device emulate_rest_reord: %d\n",
+               da->da_dev, flag);
+       return count;
+}
+
+/*
+ * Note, this can only be called on unexported SE Device Object.
+ */
+static ssize_t store_queue_depth(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       struct se_device *dev = da->da_dev;
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device TCQ while"
+                       " export_count is %d\n",
+                       dev, dev->export_count);
+               return -EINVAL;
+       }
+       if (!val) {
+               pr_err("dev[%p]: Illegal ZERO value for queue_depth\n", dev);
+               return -EINVAL;
+       }
+
+       if (val > dev->dev_attrib.queue_depth) {
+               if (val > dev->dev_attrib.hw_queue_depth) {
+                       pr_err("dev[%p]: Passed queue_depth:"
+                               " %u exceeds TCM/SE_Device MAX"
+                               " TCQ: %u\n", dev, val,
+                               dev->dev_attrib.hw_queue_depth);
+                       return -EINVAL;
+               }
+       }
+       da->queue_depth = dev->queue_depth = val;
+       pr_debug("dev[%p]: SE Device TCQ Depth changed to: %u\n", dev, val);
+       return count;
+}
+
+static ssize_t store_optimal_sectors(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device"
+                       " optimal_sectors while export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+       if (val > da->hw_max_sectors) {
+               pr_err("dev[%p]: Passed optimal_sectors %u cannot be"
+                       " greater than hw_max_sectors: %u\n",
+                       da->da_dev, val, da->hw_max_sectors);
+               return -EINVAL;
+       }
+
+       da->optimal_sectors = val;
+       pr_debug("dev[%p]: SE Device optimal_sectors changed to %u\n",
+                       da->da_dev, val);
+       return count;
+}
+
+static ssize_t store_block_size(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device block_size"
+                       " while export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+
+       if (val != 512 && val != 1024 && val != 2048 && val != 4096) {
+               pr_err("dev[%p]: Illegal value for block_device: %u"
+                       " for SE device, must be 512, 1024, 2048 or 4096\n",
+                       da->da_dev, val);
+               return -EINVAL;
+       }
+
+       da->block_size = val;
+       if (da->max_bytes_per_io)
+               da->hw_max_sectors = da->max_bytes_per_io / val;
+
+       pr_debug("dev[%p]: SE Device block_size changed to %u\n",
+                       da->da_dev, val);
+       return count;
+}
+
+CONFIGFS_EATTR_STRUCT(target_backend_dev_attrib, se_dev_attrib);
+#define TB_DEV_ATTR(_backend, _name, _mode)                            \
+static struct target_backend_dev_attrib_attribute _backend##_dev_attrib_##_name = \
+       __CONFIGFS_EATTR(_name, _mode,                                  \
+       show_##_name,                                                   \
+       store_##_name);
+
+#define TB_DEV_ATTR_RO(_backend, _name)                                        \
+static struct target_backend_dev_attrib_attribute _backend##_dev_attrib_##_name = \
+       __CONFIGFS_EATTR_RO(_name,                                      \
+       show_##_name);
+
+TB_DEV_ATTR(target_core, emulate_model_alias, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_dpo, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_fua_write, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_fua_read, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_write_cache, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_ua_intlck_ctrl, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_tas, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_tpu, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_tpws, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_caw, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_3pc, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, pi_prot_type, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR_RO(target_core, hw_pi_prot_type);
+TB_DEV_ATTR(target_core, pi_prot_format, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, enforce_pr_isids, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, is_nonrot, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, emulate_rest_reord, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, force_pr_aptpl, S_IRUGO | S_IWUSR)
+TB_DEV_ATTR_RO(target_core, hw_block_size);
+TB_DEV_ATTR(target_core, block_size, S_IRUGO | S_IWUSR)
+TB_DEV_ATTR_RO(target_core, hw_max_sectors);
+TB_DEV_ATTR(target_core, optimal_sectors, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR_RO(target_core, hw_queue_depth);
+TB_DEV_ATTR(target_core, queue_depth, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, max_unmap_lba_count, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, max_unmap_block_desc_count, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, unmap_granularity, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, unmap_granularity_alignment, S_IRUGO | S_IWUSR);
+TB_DEV_ATTR(target_core, max_write_same_len, S_IRUGO | S_IWUSR);
 
 CONFIGFS_EATTR_STRUCT(target_core_dev_attrib, se_dev_attrib);
 CONFIGFS_EATTR_OPS(target_core_dev_attrib, se_dev_attrib, da_group);
 
+/*
+ * dev_attrib attributes for devices using the target core SBC/SPC
+ * interpreter.  Any backend using spc_parse_cdb should be using
+ * these.
+ */
+struct configfs_attribute *sbc_attrib_attrs[] = {
+       &target_core_dev_attrib_emulate_model_alias.attr,
+       &target_core_dev_attrib_emulate_dpo.attr,
+       &target_core_dev_attrib_emulate_fua_write.attr,
+       &target_core_dev_attrib_emulate_fua_read.attr,
+       &target_core_dev_attrib_emulate_write_cache.attr,
+       &target_core_dev_attrib_emulate_ua_intlck_ctrl.attr,
+       &target_core_dev_attrib_emulate_tas.attr,
+       &target_core_dev_attrib_emulate_tpu.attr,
+       &target_core_dev_attrib_emulate_tpws.attr,
+       &target_core_dev_attrib_emulate_caw.attr,
+       &target_core_dev_attrib_emulate_3pc.attr,
+       &target_core_dev_attrib_pi_prot_type.attr,
+       &target_core_dev_attrib_hw_pi_prot_type.attr,
+       &target_core_dev_attrib_pi_prot_format.attr,
+       &target_core_dev_attrib_enforce_pr_isids.attr,
+       &target_core_dev_attrib_is_nonrot.attr,
+       &target_core_dev_attrib_emulate_rest_reord.attr,
+       &target_core_dev_attrib_force_pr_aptpl.attr,
+       &target_core_dev_attrib_hw_block_size.attr,
+       &target_core_dev_attrib_block_size.attr,
+       &target_core_dev_attrib_hw_max_sectors.attr,
+       &target_core_dev_attrib_optimal_sectors.attr,
+       &target_core_dev_attrib_hw_queue_depth.attr,
+       &target_core_dev_attrib_queue_depth.attr,
+       &target_core_dev_attrib_max_unmap_lba_count.attr,
+       &target_core_dev_attrib_max_unmap_block_desc_count.attr,
+       &target_core_dev_attrib_unmap_granularity.attr,
+       &target_core_dev_attrib_unmap_granularity_alignment.attr,
+       &target_core_dev_attrib_max_write_same_len.attr,
+       NULL,
+};
+EXPORT_SYMBOL(sbc_attrib_attrs);
+
+TB_DEV_ATTR_RO(target_pt, hw_pi_prot_type);
+TB_DEV_ATTR_RO(target_pt, hw_block_size);
+TB_DEV_ATTR_RO(target_pt, hw_max_sectors);
+TB_DEV_ATTR_RO(target_pt, hw_queue_depth);
+
+/*
+ * Minimal dev_attrib attributes for devices passing through CDBs.
+ * In this case we only provide a few read-only attributes for
+ * backwards compatibility.
+ */
+struct configfs_attribute *passthrough_attrib_attrs[] = {
+       &target_pt_dev_attrib_hw_pi_prot_type.attr,
+       &target_pt_dev_attrib_hw_block_size.attr,
+       &target_pt_dev_attrib_hw_max_sectors.attr,
+       &target_pt_dev_attrib_hw_queue_depth.attr,
+       NULL,
+};
+EXPORT_SYMBOL(passthrough_attrib_attrs);
+
 static struct configfs_item_operations target_core_dev_attrib_ops = {
        .show_attribute         = target_core_dev_attrib_attr_show,
        .store_attribute        = target_core_dev_attrib_attr_store,
 };
 
-TB_CIT_SETUP(dev_attrib, &target_core_dev_attrib_ops, NULL, NULL);
+TB_CIT_SETUP_DRV(dev_attrib, &target_core_dev_attrib_ops, NULL);
 
 /* End functions for struct config_item_type tb_dev_attrib_cit */
 
@@ -862,7 +1416,6 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_holder_tg_port(
                struct se_device *dev, char *page)
 {
        struct se_node_acl *se_nacl;
-       struct se_lun *lun;
        struct se_portal_group *se_tpg;
        struct t10_pr_registration *pr_reg;
        const struct target_core_fabric_ops *tfo;
@@ -877,7 +1430,6 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_holder_tg_port(
 
        se_nacl = pr_reg->pr_reg_nacl;
        se_tpg = se_nacl->se_tpg;
-       lun = pr_reg->pr_reg_tg_pt_lun;
        tfo = se_tpg->se_tpg_tfo;
 
        len += sprintf(page+len, "SPC-3 Reservation: %s"
@@ -885,9 +1437,9 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_holder_tg_port(
                tfo->tpg_get_wwn(se_tpg));
        len += sprintf(page+len, "SPC-3 Reservation: Relative Port"
                " Identifier Tag: %hu %s Portal Group Tag: %hu"
-               " %s Logical Unit: %u\n", lun->lun_sep->sep_rtpi,
+               " %s Logical Unit: %llu\n", pr_reg->tg_pt_sep_rtpi,
                tfo->get_fabric_name(), tfo->tpg_get_tag(se_tpg),
-               tfo->get_fabric_name(), lun->unpacked_lun);
+               tfo->get_fabric_name(), pr_reg->pr_aptpl_target_lun);
 
 out_unlock:
        spin_unlock(&dev->dev_reservation_lock);
@@ -1012,12 +1564,12 @@ static match_table_t tokens = {
        {Opt_res_type, "res_type=%d"},
        {Opt_res_scope, "res_scope=%d"},
        {Opt_res_all_tg_pt, "res_all_tg_pt=%d"},
-       {Opt_mapped_lun, "mapped_lun=%d"},
+       {Opt_mapped_lun, "mapped_lun=%lld"},
        {Opt_target_fabric, "target_fabric=%s"},
        {Opt_target_node, "target_node=%s"},
        {Opt_tpgt, "tpgt=%d"},
        {Opt_port_rtpi, "port_rtpi=%d"},
-       {Opt_target_lun, "target_lun=%d"},
+       {Opt_target_lun, "target_lun=%lld"},
        {Opt_err, NULL}
 };
 
@@ -1032,15 +1584,15 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
        substring_t args[MAX_OPT_ARGS];
        unsigned long long tmp_ll;
        u64 sa_res_key = 0;
-       u32 mapped_lun = 0, target_lun = 0;
+       u64 mapped_lun = 0, target_lun = 0;
        int ret = -1, res_holder = 0, all_tg_pt = 0, arg, token;
-       u16 port_rpti = 0, tpgt = 0;
-       u8 type = 0, scope;
+       u16 tpgt = 0;
+       u8 type = 0;
 
        if (dev->transport->transport_flags & TRANSPORT_FLAG_PASSTHROUGH)
-               return 0;
+               return count;
        if (dev->dev_reservation_flags & DRF_SPC2_RESERVATIONS)
-               return 0;
+               return count;
 
        if (dev->export_count) {
                pr_debug("Unable to process APTPL metadata while"
@@ -1106,24 +1658,33 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                 * PR APTPL Metadata for Reservation
                 */
                case Opt_res_holder:
-                       match_int(args, &arg);
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
                        res_holder = arg;
                        break;
                case Opt_res_type:
-                       match_int(args, &arg);
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
                        type = (u8)arg;
                        break;
                case Opt_res_scope:
-                       match_int(args, &arg);
-                       scope = (u8)arg;
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
                        break;
                case Opt_res_all_tg_pt:
-                       match_int(args, &arg);
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
                        all_tg_pt = (int)arg;
                        break;
                case Opt_mapped_lun:
-                       match_int(args, &arg);
-                       mapped_lun = (u32)arg;
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
+                       mapped_lun = (u64)arg;
                        break;
                /*
                 * PR APTPL Metadata for Target Port
@@ -1150,16 +1711,21 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                        }
                        break;
                case Opt_tpgt:
-                       match_int(args, &arg);
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
                        tpgt = (u16)arg;
                        break;
                case Opt_port_rtpi:
-                       match_int(args, &arg);
-                       port_rpti = (u16)arg;
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
                        break;
                case Opt_target_lun:
-                       match_int(args, &arg);
-                       target_lun = (u32)arg;
+                       ret = match_int(args, &arg);
+                       if (ret)
+                               goto out;
+                       target_lun = (u64)arg;
                        break;
                default:
                        break;
@@ -1223,13 +1789,13 @@ TB_CIT_SETUP(dev_pr, &target_core_dev_pr_ops, NULL, target_core_dev_pr_attrs);
 static ssize_t target_core_show_dev_info(void *p, char *page)
 {
        struct se_device *dev = p;
-       struct se_subsystem_api *t = dev->transport;
        int bl = 0;
        ssize_t read_bytes = 0;
 
        transport_dump_dev_state(dev, page, &bl);
        read_bytes += bl;
-       read_bytes += t->show_configfs_dev_params(dev, page+read_bytes);
+       read_bytes += dev->transport->show_configfs_dev_params(dev,
+                       page+read_bytes);
        return read_bytes;
 }
 
@@ -1247,9 +1813,8 @@ static ssize_t target_core_store_dev_control(
        size_t count)
 {
        struct se_device *dev = p;
-       struct se_subsystem_api *t = dev->transport;
 
-       return t->set_configfs_dev_params(dev, page, count);
+       return dev->transport->set_configfs_dev_params(dev, page, count);
 }
 
 static struct target_core_configfs_attribute target_core_attr_dev_control = {
@@ -1436,7 +2001,7 @@ static ssize_t target_core_store_alua_lu_gp(
 
        lu_gp_mem = dev->dev_alua_lu_gp_mem;
        if (!lu_gp_mem)
-               return 0;
+               return count;
 
        if (count > LU_GROUP_NAME_BUF) {
                pr_err("ALUA LU Group Alias too large!\n");
@@ -2339,21 +2904,16 @@ static ssize_t target_core_alua_tg_pt_gp_show_attr_members(
        struct t10_alua_tg_pt_gp *tg_pt_gp,
        char *page)
 {
-       struct se_port *port;
-       struct se_portal_group *tpg;
        struct se_lun *lun;
-       struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
        ssize_t len = 0, cur_len;
        unsigned char buf[TG_PT_GROUP_NAME_BUF];
 
        memset(buf, 0, TG_PT_GROUP_NAME_BUF);
 
        spin_lock(&tg_pt_gp->tg_pt_gp_lock);
-       list_for_each_entry(tg_pt_gp_mem, &tg_pt_gp->tg_pt_gp_mem_list,
-                       tg_pt_gp_mem_list) {
-               port = tg_pt_gp_mem->tg_pt;
-               tpg = port->sep_tpg;
-               lun = port->sep_lun;
+       list_for_each_entry(lun, &tg_pt_gp->tg_pt_gp_lun_list,
+                       lun_tg_pt_gp_link) {
+               struct se_portal_group *tpg = lun->lun_tpg;
 
                cur_len = snprintf(buf, TG_PT_GROUP_NAME_BUF, "%s/%s/tpgt_%hu"
                        "/%s\n", tpg->se_tpg_tfo->get_fabric_name(),
@@ -2526,9 +3086,9 @@ static struct config_group *target_core_make_subdev(
        const char *name)
 {
        struct t10_alua_tg_pt_gp *tg_pt_gp;
-       struct se_subsystem_api *t;
        struct config_item *hba_ci = &group->cg_item;
        struct se_hba *hba = item_to_hba(hba_ci);
+       struct target_backend *tb = hba->backend;
        struct se_device *dev;
        struct config_group *dev_cg = NULL, *tg_pt_gp_cg = NULL;
        struct config_group *dev_stat_grp = NULL;
@@ -2537,10 +3097,6 @@ static struct config_group *target_core_make_subdev(
        ret = mutex_lock_interruptible(&hba->hba_access_mutex);
        if (ret)
                return ERR_PTR(ret);
-       /*
-        * Locate the struct se_subsystem_api from parent's struct se_hba.
-        */
-       t = hba->transport;
 
        dev = target_alloc_device(hba, name);
        if (!dev)
@@ -2553,17 +3109,17 @@ static struct config_group *target_core_make_subdev(
        if (!dev_cg->default_groups)
                goto out_free_device;
 
-       config_group_init_type_name(dev_cg, name, &t->tb_cits.tb_dev_cit);
+       config_group_init_type_name(dev_cg, name, &tb->tb_dev_cit);
        config_group_init_type_name(&dev->dev_attrib.da_group, "attrib",
-                       &t->tb_cits.tb_dev_attrib_cit);
+                       &tb->tb_dev_attrib_cit);
        config_group_init_type_name(&dev->dev_pr_group, "pr",
-                       &t->tb_cits.tb_dev_pr_cit);
+                       &tb->tb_dev_pr_cit);
        config_group_init_type_name(&dev->t10_wwn.t10_wwn_group, "wwn",
-                       &t->tb_cits.tb_dev_wwn_cit);
+                       &tb->tb_dev_wwn_cit);
        config_group_init_type_name(&dev->t10_alua.alua_tg_pt_gps_group,
-                       "alua", &t->tb_cits.tb_dev_alua_tg_pt_gps_cit);
+                       "alua", &tb->tb_dev_alua_tg_pt_gps_cit);
        config_group_init_type_name(&dev->dev_stat_grps.stat_group,
-                       "statistics", &t->tb_cits.tb_dev_stat_cit);
+                       "statistics", &tb->tb_dev_stat_cit);
 
        dev_cg->default_groups[0] = &dev->dev_attrib.da_group;
        dev_cg->default_groups[1] = &dev->dev_pr_group;
@@ -2693,8 +3249,8 @@ static ssize_t target_core_hba_show_attr_hba_info(
        char *page)
 {
        return sprintf(page, "HBA Index: %d plugin: %s version: %s\n",
-                       hba->hba_id, hba->transport->name,
-                       TARGET_CORE_CONFIGFS_VERSION);
+                       hba->hba_id, hba->backend->ops->name,
+                       TARGET_CORE_VERSION);
 }
 
 SE_HBA_ATTR_RO(hba_info);
@@ -2713,11 +3269,10 @@ static ssize_t target_core_hba_show_attr_hba_mode(struct se_hba *hba,
 static ssize_t target_core_hba_store_attr_hba_mode(struct se_hba *hba,
                                const char *page, size_t count)
 {
-       struct se_subsystem_api *transport = hba->transport;
        unsigned long mode_flag;
        int ret;
 
-       if (transport->pmode_enable_hba == NULL)
+       if (hba->backend->ops->pmode_enable_hba == NULL)
                return -EINVAL;
 
        ret = kstrtoul(page, 0, &mode_flag);
@@ -2731,7 +3286,7 @@ static ssize_t target_core_hba_store_attr_hba_mode(struct se_hba *hba,
                return -EINVAL;
        }
 
-       ret = transport->pmode_enable_hba(hba, mode_flag);
+       ret = hba->backend->ops->pmode_enable_hba(hba, mode_flag);
        if (ret < 0)
                return -EINVAL;
        if (ret > 0)
@@ -2857,16 +3412,15 @@ static struct config_item_type target_core_cit = {
 
 /* Stop functions for struct config_item_type target_core_hba_cit */
 
-void target_core_setup_sub_cits(struct se_subsystem_api *sa)
+void target_setup_backend_cits(struct target_backend *tb)
 {
-       target_core_setup_dev_cit(sa);
-       target_core_setup_dev_attrib_cit(sa);
-       target_core_setup_dev_pr_cit(sa);
-       target_core_setup_dev_wwn_cit(sa);
-       target_core_setup_dev_alua_tg_pt_gps_cit(sa);
-       target_core_setup_dev_stat_cit(sa);
+       target_core_setup_dev_cit(tb);
+       target_core_setup_dev_attrib_cit(tb);
+       target_core_setup_dev_pr_cit(tb);
+       target_core_setup_dev_wwn_cit(tb);
+       target_core_setup_dev_alua_tg_pt_gps_cit(tb);
+       target_core_setup_dev_stat_cit(tb);
 }
-EXPORT_SYMBOL(target_core_setup_sub_cits);
 
 static int __init target_core_init_configfs(void)
 {
@@ -2968,7 +3522,7 @@ static int __init target_core_init_configfs(void)
                goto out_global;
        }
        pr_debug("TARGET_CORE[0]: Initialized ConfigFS Fabric"
-               " Infrastructure: "TARGET_CORE_CONFIGFS_VERSION" on %s/%s"
+               " Infrastructure: "TARGET_CORE_VERSION" on %s/%s"
                " on "UTS_RELEASE"\n", utsname()->sysname, utsname()->machine);
        /*
         * Register built-in RAMDISK subsystem logic for virtual LUN 0