Lines Matching refs:tcmu_dev

100 struct tcmu_dev;
106 struct tcmu_dev *udev;
111 struct tcmu_dev { struct
174 #define TCMU_DEV(_se_dev) container_of(_se_dev, struct tcmu_dev, se_dev) argument
178 struct tcmu_dev *tcmu_dev; member
305 struct tcmu_dev *udev = nl_cmd->udev; in tcmu_fail_netlink_cmd()
374 struct tcmu_dev *udev = NULL; in tcmu_genl_cmd_done()
499 struct tcmu_dev *udev = tcmu_cmd->tcmu_dev; in tcmu_cmd_free_data()
506 static inline int tcmu_get_empty_block(struct tcmu_dev *udev, in tcmu_get_empty_block()
554 static int tcmu_get_empty_blocks(struct tcmu_dev *udev, in tcmu_get_empty_blocks()
582 uint32_t blk_size = cmd->tcmu_dev->data_blk_size; in tcmu_cmd_set_block_cnts()
596 static int new_block_to_iov(struct tcmu_dev *udev, struct tcmu_cmd *cmd, in new_block_to_iov()
622 static void tcmu_setup_iovs(struct tcmu_dev *udev, struct tcmu_cmd *cmd, in tcmu_setup_iovs()
636 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_alloc_cmd()
645 tcmu_cmd->tcmu_dev = udev; in tcmu_alloc_cmd()
702 static inline void tcmu_copy_data(struct tcmu_dev *udev, in tcmu_copy_data()
774 static void scatter_data_area(struct tcmu_dev *udev, struct tcmu_cmd *tcmu_cmd, in scatter_data_area()
783 static void gather_data_area(struct tcmu_dev *udev, struct tcmu_cmd *tcmu_cmd, in gather_data_area()
820 static bool is_ring_space_avail(struct tcmu_dev *udev, size_t cmd_size) in is_ring_space_avail()
854 static int tcmu_alloc_data_space(struct tcmu_dev *udev, struct tcmu_cmd *cmd, in tcmu_alloc_data_space()
926 tcmu_cmd->tcmu_dev->name, tmo / MSEC_PER_SEC); in tcmu_setup_cmd_timer()
931 struct tcmu_dev *udev = tcmu_cmd->tcmu_dev; in add_to_qfull_queue()
955 static uint32_t ring_insert_padding(struct tcmu_dev *udev, size_t cmd_size) in ring_insert_padding()
986 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_unplug_device()
994 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_plug_device()
1014 struct tcmu_dev *udev = tcmu_cmd->tcmu_dev; in queue_cmd_ring()
1158 queue_tmr_ring(struct tcmu_dev *udev, struct tcmu_tmr *tmr) in queue_tmr_ring()
1206 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_queue_cmd()
1263 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_tmr_notify()
1324 struct tcmu_dev *udev = cmd->tcmu_dev; in tcmu_handle_completion()
1411 static int tcmu_run_tmr_queue(struct tcmu_dev *udev) in tcmu_run_tmr_queue()
1443 static bool tcmu_handle_completions(struct tcmu_dev *udev) in tcmu_handle_completions()
1532 cmd->cmd_id, cmd->tcmu_dev->name); in tcmu_check_expired_ring_cmd()
1545 cmd, cmd->tcmu_dev->name); in tcmu_check_expired_queue_cmd()
1555 static void tcmu_device_timedout(struct tcmu_dev *udev) in tcmu_device_timedout()
1567 struct tcmu_dev *udev = timer_container_of(udev, t, cmd_timer); in tcmu_cmd_timedout()
1575 struct tcmu_dev *udev = timer_container_of(udev, t, qfull_timer); in tcmu_qfull_timedout()
1603 struct tcmu_dev *udev; in tcmu_alloc_device()
1605 udev = kzalloc_obj(struct tcmu_dev); in tcmu_alloc_device()
1645 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_dev_call_rcu()
1662 static u32 tcmu_blocks_release(struct tcmu_dev *udev, unsigned long first, in tcmu_blocks_release()
1702 static void tcmu_remove_all_queued_tmr(struct tcmu_dev *udev) in tcmu_remove_all_queued_tmr()
1714 struct tcmu_dev *udev = container_of(kref, struct tcmu_dev, kref); in tcmu_dev_kref_release()
1750 static void run_qfull_queue(struct tcmu_dev *udev, bool fail) in run_qfull_queue()
1812 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info); in tcmu_irqcontrol()
1828 struct tcmu_dev *udev = vma->vm_private_data; in tcmu_find_mem_index()
1839 static struct page *tcmu_try_get_data_page(struct tcmu_dev *udev, uint32_t dpi) in tcmu_try_get_data_page()
1865 struct tcmu_dev *udev = vma->vm_private_data; in tcmu_vma_open()
1874 struct tcmu_dev *udev = vma->vm_private_data; in tcmu_vma_close()
1884 struct tcmu_dev *udev = vmf->vma->vm_private_data; in tcmu_vma_fault()
1929 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info); in tcmu_mmap()
1947 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info); in tcmu_open()
1962 struct tcmu_dev *udev = container_of(info, struct tcmu_dev, uio_info); in tcmu_release()
2003 static int tcmu_init_genl_cmd_reply(struct tcmu_dev *udev, int cmd) in tcmu_init_genl_cmd_reply()
2041 static void tcmu_destroy_genl_cmd_reply(struct tcmu_dev *udev) in tcmu_destroy_genl_cmd_reply()
2059 static int tcmu_wait_genl_cmd_reply(struct tcmu_dev *udev) in tcmu_wait_genl_cmd_reply()
2081 static int tcmu_netlink_event_init(struct tcmu_dev *udev, in tcmu_netlink_event_init()
2118 static int tcmu_netlink_event_send(struct tcmu_dev *udev, in tcmu_netlink_event_send()
2145 static int tcmu_send_dev_add_event(struct tcmu_dev *udev) in tcmu_send_dev_add_event()
2159 static int tcmu_send_dev_remove_event(struct tcmu_dev *udev) in tcmu_send_dev_remove_event()
2173 static int tcmu_update_uio_info(struct tcmu_dev *udev) in tcmu_update_uio_info()
2199 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_configure_device()
2314 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_free_device()
2322 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_destroy_device()
2339 static void tcmu_unblock_dev(struct tcmu_dev *udev) in tcmu_unblock_dev()
2346 static void tcmu_block_dev(struct tcmu_dev *udev) in tcmu_block_dev()
2362 static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level) in tcmu_reset_ring()
2465 static int tcmu_set_max_blocks_param(struct tcmu_dev *udev, substring_t *arg) in tcmu_set_max_blocks_param()
2506 static int tcmu_set_data_pages_per_blk(struct tcmu_dev *udev, substring_t *arg) in tcmu_set_data_pages_per_blk()
2539 static int tcmu_set_cmd_ring_size(struct tcmu_dev *udev, substring_t *arg) in tcmu_set_cmd_ring_size()
2577 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_set_configfs_dev_params()
2645 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_show_configfs_dev_params()
2661 struct tcmu_dev *udev = TCMU_DEV(dev); in tcmu_get_blocks()
2677 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_cmd_time_out_show()
2687 struct tcmu_dev *udev = container_of(da->da_dev, in tcmu_cmd_time_out_store()
2688 struct tcmu_dev, se_dev); in tcmu_cmd_time_out_store()
2710 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_qfull_time_out_show()
2722 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_qfull_time_out_store()
2746 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_max_data_area_mb_show()
2757 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_data_pages_per_blk_show()
2767 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_cmd_ring_size_mb_show()
2778 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_dev_config_show()
2783 static int tcmu_send_dev_config_event(struct tcmu_dev *udev, in tcmu_send_dev_config_event()
2809 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_dev_config_store()
2840 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_dev_size_show()
2845 static int tcmu_send_dev_size_event(struct tcmu_dev *udev, u64 size) in tcmu_send_dev_size_event()
2870 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_dev_size_store()
2896 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_nl_reply_supported_show()
2906 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_nl_reply_supported_store()
2928 static int tcmu_send_emulate_write_cache(struct tcmu_dev *udev, u8 val) in tcmu_send_emulate_write_cache()
2952 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_emulate_write_cache_store()
2978 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_tmr_notification_show()
2989 struct tcmu_dev *udev = TCMU_DEV(da->da_dev); in tcmu_tmr_notification_store()
3012 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_block_dev_show()
3026 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_block_dev_store()
3058 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_reset_ring_store()
3087 struct tcmu_dev *udev = TCMU_DEV(se_dev); in tcmu_free_kept_buf_store()
3188 struct tcmu_dev *udev; in find_free_blocks()
3265 struct tcmu_dev *udev, *tmp_dev; in check_timedout_devices()