tcmu_dev
struct tcmu_dev;
struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
struct tcmu_dev *udev;
queue_tmr_ring(struct tcmu_dev *udev, struct tcmu_tmr *tmr)
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev = cmd->tcmu_dev;
static int tcmu_run_tmr_queue(struct tcmu_dev *udev)
static bool tcmu_handle_completions(struct tcmu_dev *udev)
cmd->cmd_id, cmd->tcmu_dev->name);
cmd, cmd->tcmu_dev->name);
static void tcmu_device_timedout(struct tcmu_dev *udev)
struct tcmu_dev *udev = timer_container_of(udev, t, cmd_timer);
struct tcmu_dev *udev = timer_container_of(udev, t, qfull_timer);
struct tcmu_dev *udev;
udev = kzalloc_obj(struct tcmu_dev);
struct tcmu_dev *udev = TCMU_DEV(dev);
static u32 tcmu_blocks_release(struct tcmu_dev *udev, unsigned long first,
static void tcmu_remove_all_queued_tmr(struct tcmu_dev *udev)
struct tcmu_dev *udev = container_of(kref, struct tcmu_dev, kref);
#define TCMU_DEV(_se_dev) container_of(_se_dev, struct tcmu_dev, se_dev)
static void run_qfull_queue(struct tcmu_dev *udev, bool fail)
struct tcmu_dev *tcmu_dev;
struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
struct tcmu_dev *udev = vma->vm_private_data;
static struct page *tcmu_try_get_data_page(struct tcmu_dev *udev, uint32_t dpi)
struct tcmu_dev *udev = vma->vm_private_data;
struct tcmu_dev *udev = vma->vm_private_data;
struct tcmu_dev *udev = vmf->vma->vm_private_data;
struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info);
static int tcmu_init_genl_cmd_reply(struct tcmu_dev *udev, int cmd)
static void tcmu_destroy_genl_cmd_reply(struct tcmu_dev *udev)
static int tcmu_wait_genl_cmd_reply(struct tcmu_dev *udev)
static int tcmu_netlink_event_init(struct tcmu_dev *udev,
static int tcmu_netlink_event_send(struct tcmu_dev *udev,
static int tcmu_send_dev_add_event(struct tcmu_dev *udev)
static int tcmu_send_dev_remove_event(struct tcmu_dev *udev)
static int tcmu_update_uio_info(struct tcmu_dev *udev)
struct tcmu_dev *udev = TCMU_DEV(dev);
struct tcmu_dev *udev = TCMU_DEV(dev);
struct tcmu_dev *udev = TCMU_DEV(dev);
static void tcmu_unblock_dev(struct tcmu_dev *udev)
static void tcmu_block_dev(struct tcmu_dev *udev)
static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level)
static int tcmu_set_max_blocks_param(struct tcmu_dev *udev, substring_t *arg)
static int tcmu_set_data_pages_per_blk(struct tcmu_dev *udev, substring_t *arg)
static int tcmu_set_cmd_ring_size(struct tcmu_dev *udev, substring_t *arg)
struct tcmu_dev *udev = TCMU_DEV(dev);
struct tcmu_dev *udev = TCMU_DEV(dev);
struct tcmu_dev *udev = TCMU_DEV(dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = container_of(da->da_dev,
struct tcmu_dev, se_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
static int tcmu_send_dev_config_event(struct tcmu_dev *udev,
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
static int tcmu_send_dev_size_event(struct tcmu_dev *udev, u64 size)
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
static int tcmu_send_emulate_write_cache(struct tcmu_dev *udev, u8 val)
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(da->da_dev);
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev = nl_cmd->udev;
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev;
struct tcmu_dev *udev, *tmp_dev;
struct tcmu_dev *udev = NULL;
struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
static inline int tcmu_get_empty_block(struct tcmu_dev *udev,
static int tcmu_get_empty_blocks(struct tcmu_dev *udev,
uint32_t blk_size = cmd->tcmu_dev->data_blk_size;
static int new_block_to_iov(struct tcmu_dev *udev, struct tcmu_cmd *cmd,
static void tcmu_setup_iovs(struct tcmu_dev *udev, struct tcmu_cmd *cmd,
struct tcmu_dev *udev = TCMU_DEV(se_dev);
tcmu_cmd->tcmu_dev = udev;
static inline void tcmu_copy_data(struct tcmu_dev *udev,
static void scatter_data_area(struct tcmu_dev *udev, struct tcmu_cmd *tcmu_cmd,
static void gather_data_area(struct tcmu_dev *udev, struct tcmu_cmd *tcmu_cmd,
static bool is_ring_space_avail(struct tcmu_dev *udev, size_t cmd_size)
static int tcmu_alloc_data_space(struct tcmu_dev *udev, struct tcmu_cmd *cmd,
tcmu_cmd->tcmu_dev->name, tmo / MSEC_PER_SEC);
struct tcmu_dev *udev = tcmu_cmd->tcmu_dev;
static uint32_t ring_insert_padding(struct tcmu_dev *udev, size_t cmd_size)
struct tcmu_dev *udev = TCMU_DEV(se_dev);
struct tcmu_dev *udev = TCMU_DEV(se_dev);