• Home
  • Raw
  • Download

Lines Matching refs:se_dev

113 	struct se_device se_dev;  member
166 #define TCMU_DEV(_se_dev) container_of(_se_dev, struct tcmu_dev, se_dev)
382 if (nl_cmd->udev->se_dev.dev_index == dev_id) { in tcmu_genl_cmd_done()
625 struct se_device *se_dev = se_cmd->se_dev; in tcmu_alloc_cmd() local
626 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_alloc_cmd()
1178 struct se_device *se_dev = se_cmd->se_dev; in tcmu_queue_cmd() local
1179 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_queue_cmd()
1228 tcmu_tmr_notify(struct se_device *se_dev, enum tcm_tmreq_table tmf, in tcmu_tmr_notify() argument
1237 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_tmr_notify()
1580 return &udev->se_dev; in tcmu_alloc_device()
1816 struct se_device *dev = &udev->se_dev; in tcmu_dev_kref_release()
1963 ret = nla_put_u32(skb, TCMU_ATTR_DEVICE_ID, udev->se_dev.dev_index); in tcmu_netlink_event_init()
2454 struct tcmu_dev, se_dev); in tcmu_cmd_time_out_store()
2562 if (target_dev_configured(&udev->se_dev)) { in tcmu_dev_config_store()
2624 if (target_dev_configured(&udev->se_dev)) { in tcmu_dev_size_store()
2706 if (target_dev_configured(&udev->se_dev)) { in tcmu_emulate_write_cache_store()
2754 struct se_device *se_dev = container_of(to_config_group(item), in tcmu_block_dev_show() local
2757 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_block_dev_show()
2768 struct se_device *se_dev = container_of(to_config_group(item), in tcmu_block_dev_store() local
2771 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_block_dev_store()
2775 if (!target_dev_configured(&udev->se_dev)) { in tcmu_block_dev_store()
2800 struct se_device *se_dev = container_of(to_config_group(item), in tcmu_reset_ring_store() local
2803 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_reset_ring_store()
2807 if (!target_dev_configured(&udev->se_dev)) { in tcmu_reset_ring_store()
2880 if (!target_dev_configured(&udev->se_dev)) { in find_free_blocks()