Home
last modified time | relevance | path

Searched refs:io_kiocb (Results 1 – 25 of 61) sorted by relevance

123

/linux/io_uring/
H A Dnet.h28 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
29 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags);
31 void io_sendmsg_recvmsg_cleanup(struct io_kiocb *req);
32 int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
33 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags);
35 int io_send(struct io_kiocb *req, unsigned int issue_flags);
37 int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
38 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags);
39 int io_recv(struct io_kiocb *req, unsigned int issue_flags);
41 void io_sendrecv_fail(struct io_kiocb *req);
[all …]
H A Drw.h35 int io_prep_read_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe);
36 int io_prep_write_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe);
37 int io_prep_readv_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe);
38 int io_prep_writev_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe);
39 int io_prep_readv(struct io_kiocb *req, const struct io_uring_sqe *sqe);
40 int io_prep_writev(struct io_kiocb *req, const struct io_uring_sqe *sqe);
41 int io_prep_read(struct io_kiocb *req, const struct io_uring_sqe *sqe);
42 int io_prep_write(struct io_kiocb *req, const struct io_uring_sqe *sqe);
43 int io_read(struct io_kiocb *req, unsigned int issue_flags);
44 int io_write(struct io_kiocb *req, unsigned int issue_flags);
[all …]
H A Dfs.h3 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags);
5 void io_renameat_cleanup(struct io_kiocb *req);
7 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags);
9 void io_unlinkat_cleanup(struct io_kiocb *req);
11 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags);
13 void io_mkdirat_cleanup(struct io_kiocb *req);
15 int io_symlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
[all …]
H A Dopenclose.h8 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
9 int io_openat(struct io_kiocb *req, unsigned int issue_flags);
10 void io_open_cleanup(struct io_kiocb *req);
11 void io_openat_bpf_populate(struct io_uring_bpf_ctx *bctx, struct io_kiocb *req);
13 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
14 int io_openat2(struct io_kiocb *req, unsigned int issue_flags);
16 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
17 int io_close(struct io_kiocb *req, unsigned int issue_flags);
19 int io_pipe_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
20 int io_pipe(struct io_kiocb *req, unsigned int issue_flags);
[all …]
H A Dkbuf.h64 struct io_br_sel io_buffer_select(struct io_kiocb *req, size_t *len,
66 int io_buffers_select(struct io_kiocb *req, struct buf_sel_arg *arg,
68 int io_buffers_peek(struct io_kiocb *req, struct buf_sel_arg *arg,
72 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
73 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
74 int io_manage_buffers_legacy(struct io_kiocb *req, unsigned int issue_flags);
80 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags);
81 void io_kbuf_drop_legacy(struct io_kiocb *req);
83 unsigned int __io_put_kbufs(struct io_kiocb *req, struct io_buffer_list *bl,
85 bool io_kbuf_commit(struct io_kiocb *req,
[all …]
H A Dtimeout.c23 struct io_kiocb *head;
25 struct io_kiocb *prev;
38 static struct io_kiocb *__io_disarm_linked_timeout(struct io_kiocb *req,
39 struct io_kiocb *link);
41 static inline bool io_is_timeout_noseq(struct io_kiocb *req) in io_is_timeout_noseq()
49 static inline void io_put_req(struct io_kiocb *req) in io_put_req()
73 struct io_kiocb *req = tw_req.req; in io_timeout_complete()
99 struct io_kiocb *req; in io_flush_killed_timeouts()
112 static void io_kill_timeout(struct io_kiocb *req, struct list_head *list) in io_kill_timeout()
136 struct io_kiocb *req = cmd_to_io_kiocb(timeout); in io_flush_timeouts()
[all …]
H A Dxattr.h3 void io_xattr_cleanup(struct io_kiocb *req);
5 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags);
8 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags);
11 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags);
14 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
H A Dio_uring.h125 struct io_kiocb *req;
176 void io_req_defer_failed(struct io_kiocb *req, s32 res);
179 bool io_req_post_cqe(struct io_kiocb *req, s32 res, u32 cflags);
180 bool io_req_post_cqe32(struct io_kiocb *req, struct io_uring_cqe src_cqe[2]);
183 unsigned io_linked_nr(struct io_kiocb *req);
184 void io_req_track_inflight(struct io_kiocb *req);
185 struct file *io_file_get_normal(struct io_kiocb *req, int fd);
186 struct file *io_file_get_fixed(struct io_kiocb *req, int fd,
189 void io_req_task_queue(struct io_kiocb *req);
191 void io_req_task_queue_fail(struct io_kiocb *req, int ret);
[all …]
H A Dtimeout.h4 struct io_kiocb *req;
16 void io_queue_linked_timeout(struct io_kiocb *req);
17 void io_disarm_next(struct io_kiocb *req);
19 int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
20 int io_link_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
21 int io_timeout(struct io_kiocb *req, unsigned int issue_flags);
22 int io_timeout_remove_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
23 int io_timeout_remove(struct io_kiocb *req, unsigned int issue_flags);
H A Dpoll.c36 struct io_kiocb *req;
59 static inline struct io_kiocb *wqe_to_req(struct wait_queue_entry *wqe) in wqe_to_req()
63 return (struct io_kiocb *)(priv & ~IO_WQE_F_DOUBLE); in wqe_to_req()
73 static bool io_poll_get_ownership_slowpath(struct io_kiocb *req) in io_poll_get_ownership_slowpath()
94 static inline bool io_poll_get_ownership(struct io_kiocb *req) in io_poll_get_ownership()
101 static void io_poll_mark_cancelled(struct io_kiocb *req) in io_poll_mark_cancelled()
106 static struct io_poll *io_poll_get_double(struct io_kiocb *req) in io_poll_get_double()
114 static struct io_poll *io_poll_get_single(struct io_kiocb *req) in io_poll_get_single()
121 static void io_poll_req_insert(struct io_kiocb *req) in io_poll_req_insert()
171 static void io_poll_remove_entries(struct io_kiocb *req) in io_poll_remove_entries()
[all …]
H A During_cmd.c26 static void io_req_uring_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_uring_cleanup()
44 void io_uring_cmd_cleanup(struct io_kiocb *req) in io_uring_cmd_cleanup()
53 struct io_kiocb *req; in io_uring_try_cancel_uring_cmd()
80 struct io_kiocb *req = cmd_to_io_kiocb(cmd); in io_uring_cmd_del_cancelable()
104 struct io_kiocb *req = cmd_to_io_kiocb(cmd); in io_uring_cmd_mark_cancelable()
129 struct io_kiocb *req = cmd_to_io_kiocb(ioucmd); in __io_uring_cmd_do_in_task()
139 static inline void io_req_set_cqe32_extra(struct io_kiocb *req, in io_req_set_cqe32_extra()
153 struct io_kiocb *req = cmd_to_io_kiocb(ioucmd); in __io_uring_cmd_done()
184 int io_uring_cmd_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_uring_cmd_prep()
219 static inline size_t uring_sqe_size(struct io_kiocb *req) in uring_sqe_size()
[all …]
H A Drw.c38 static bool io_file_supports_nowait(struct io_kiocb *req, __poll_t mask) in io_file_supports_nowait()
64 static int io_iov_buffer_select_prep(struct io_kiocb *req) in io_iov_buffer_select_prep()
83 static int io_import_vec(int ddir, struct io_kiocb *req, in io_import_vec()
110 static int __io_import_rw_buffer(int ddir, struct io_kiocb *req, in __io_import_rw_buffer()
132 static inline int io_import_rw_buffer(int rw, struct io_kiocb *req, in io_import_rw_buffer()
147 static bool io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle()
165 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup()
204 static int io_rw_alloc_async(struct io_kiocb *req) in io_rw_alloc_async()
232 static int io_prep_rw_pi(struct io_kiocb *req, struct io_rw *rw, int ddir, in io_prep_rw_pi()
259 static int __io_prep_rw(struct io_kiocb *req, const struct io_uring_sqe *sqe, in __io_prep_rw()
[all …]
H A Dpoll.h30 static inline void io_poll_multishot_retry(struct io_kiocb *req) in io_poll_multishot_retry()
35 int io_poll_add_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
36 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags);
38 int io_poll_remove_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
39 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags);
44 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask);
45 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
H A Dxattr.c25 void io_xattr_cleanup(struct io_kiocb *req) in io_xattr_cleanup()
34 static void io_xattr_finish(struct io_kiocb *req, int ret) in io_xattr_finish()
42 static int __io_getxattr_prep(struct io_kiocb *req, in __io_getxattr_prep()
74 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fgetxattr_prep()
79 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_getxattr_prep()
97 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr()
109 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr()
122 static int __io_setxattr_prep(struct io_kiocb *req, in __io_setxattr_prep()
151 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_setxattr_prep()
169 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fsetxattr_prep()
[all …]
H A Dsync.h3 int io_sfr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags);
6 int io_fsync_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags);
9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
10 int io_fallocate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Drefs.h15 static inline bool req_ref_inc_not_zero(struct io_kiocb *req) in req_ref_inc_not_zero()
21 static inline bool req_ref_put_and_test_atomic(struct io_kiocb *req) in req_ref_put_and_test_atomic()
28 static inline bool req_ref_put_and_test(struct io_kiocb *req) in req_ref_put_and_test()
37 static inline void req_ref_get(struct io_kiocb *req) in req_ref_get()
44 static inline void req_ref_put(struct io_kiocb *req) in req_ref_put()
51 static inline void __io_req_set_refcount(struct io_kiocb *req, int nr) in __io_req_set_refcount()
59 static inline void io_req_set_refcount(struct io_kiocb *req) in io_req_set_refcount()
H A Dcancel.h16 int io_async_cancel_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
17 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags);
23 bool io_cancel_req_match(struct io_kiocb *req, struct io_cancel_data *cd);
24 bool io_match_task_safe(struct io_kiocb *head, struct io_uring_task *tctx,
29 bool (*cancel)(struct io_kiocb *));
32 bool (*cancel)(struct io_kiocb *));
39 static inline bool io_cancel_match_sequence(struct io_kiocb *req, int sequence) in io_cancel_match_sequence()
H A Dopdef.h41 int (*issue)(struct io_kiocb *, unsigned int);
42 int (*prep)(struct io_kiocb *, const struct io_uring_sqe *);
43 void (*filter_populate)(struct io_uring_bpf_ctx *, struct io_kiocb *);
49 void (*sqe_copy)(struct io_kiocb *);
50 void (*cleanup)(struct io_kiocb *);
51 void (*fail)(struct io_kiocb *);
H A Dsplice.h3 int io_tee_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_tee(struct io_kiocb *req, unsigned int issue_flags);
6 void io_splice_cleanup(struct io_kiocb *req);
7 int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
8 int io_splice(struct io_kiocb *req, unsigned int issue_flags);
H A Dfs.c50 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_renameat_prep()
82 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat()
99 void io_renameat_cleanup(struct io_kiocb *req) in io_renameat_cleanup()
107 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_unlinkat_prep()
134 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat()
152 void io_unlinkat_cleanup(struct io_kiocb *req) in io_unlinkat_cleanup()
159 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_mkdirat_prep()
183 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat()
198 void io_mkdirat_cleanup(struct io_kiocb *req) in io_mkdirat_cleanup()
205 int io_symlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_symlinkat_prep()
[all …]
H A Dopenclose.c53 static int __io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in __io_openat_prep()
85 void io_openat_bpf_populate(struct io_uring_bpf_ctx *bctx, struct io_kiocb *req) in io_openat_bpf_populate()
94 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_openat_prep()
104 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_openat2_prep()
123 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2()
188 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat()
193 void io_open_cleanup(struct io_kiocb *req) in io_open_cleanup()
212 static inline int io_close_fixed(struct io_kiocb *req, unsigned int issue_flags) in io_close_fixed()
219 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_close_prep()
236 int io_close(struct io_kiocb *req, unsigned int issue_flags) in io_close()
[all …]
H A Dmsg_ring.c57 void io_msg_ring_cleanup(struct io_kiocb *req) in io_msg_ring_cleanup()
75 struct io_kiocb *req = tw_req.req; in io_msg_tw_complete()
83 static void io_msg_remote_post(struct io_ring_ctx *ctx, struct io_kiocb *req, in io_msg_remote_post()
99 struct io_kiocb *target; in io_msg_data_remote()
148 static int io_msg_ring_data(struct io_kiocb *req, unsigned int issue_flags) in io_msg_ring_data()
156 static int io_msg_grab_file(struct io_kiocb *req, unsigned int issue_flags) in io_msg_grab_file()
176 static int io_msg_install_complete(struct io_kiocb *req, unsigned int issue_flags) in io_msg_install_complete()
211 struct io_kiocb *req = cmd_to_io_kiocb(msg); in io_msg_tw_fd_complete()
221 static int io_msg_fd_remote(struct io_kiocb *req) in io_msg_fd_remote()
234 static int io_msg_send_fd(struct io_kiocb *req, unsigned int issue_flags) in io_msg_send_fd()
[all …]
H A Dtw.c23 struct io_kiocb *req, *tmp; in io_fallback_req_func()
62 struct io_kiocb *req = container_of(node, struct io_kiocb, in io_handle_tw_list()
91 struct io_kiocb *req; in __io_fallback_tw()
94 req = container_of(node, struct io_kiocb, io_task_work.node); in __io_fallback_tw()
170 void io_req_local_work_add(struct io_kiocb *req, unsigned flags) in io_req_local_work_add()
192 struct io_kiocb *first_req = container_of(head, in io_req_local_work_add()
193 struct io_kiocb, in io_req_local_work_add()
239 void io_req_normal_work_add(struct io_kiocb *req) in io_req_normal_work_add()
267 void io_req_task_work_add_remote(struct io_kiocb *req, unsigned flags) in io_req_task_work_add_remote()
303 struct io_kiocb *req = container_of(*node, struct io_kiocb, in __io_run_local_work_loop()
H A Dfutex.h5 int io_futex_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
6 int io_futexv_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags);
8 int io_futexv_wait(struct io_kiocb *req, unsigned int issue_flags);
9 int io_futex_wake(struct io_kiocb *req, unsigned int issue_flags);
H A Dfutex.c27 struct io_kiocb *req;
56 struct io_kiocb *req = tw_req.req; in io_futex_complete()
67 struct io_kiocb *req = tw_req.req; in io_futexv_complete()
92 static bool __io_futex_cancel(struct io_kiocb *req) in __io_futex_cancel()
127 int io_futex_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_futex_prep()
159 struct io_kiocb *req = q->wake_data; in io_futex_wakev_fn()
172 int io_futexv_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_futexv_prep()
211 struct io_kiocb *req = ifd->req; in io_futex_wake_fn()
221 int io_futexv_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futexv_wait()
274 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futex_wait()
[all …]

123