Lines Matching refs:issue_flags

112 				 unsigned int issue_flags)  in __io_import_rw_buffer()  argument
123 *sel = io_buffer_select(req, &sqe_len, io->buf_group, issue_flags); in __io_import_rw_buffer()
135 unsigned int issue_flags) in io_import_rw_buffer() argument
139 ret = __io_import_rw_buffer(rw, req, io, sel, issue_flags); in io_import_rw_buffer()
147 static bool io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle() argument
151 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) in io_rw_recycle()
165 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup() argument
196 if (!io_rw_recycle(req, issue_flags)) { in io_req_rw_cleanup()
373 static int io_init_rw_fixed(struct io_kiocb *req, unsigned int issue_flags, in io_init_rw_fixed() argument
384 issue_flags); in io_init_rw_fixed()
401 int ddir, unsigned int issue_flags) in io_rw_import_reg_vec() argument
408 uvec_segs, issue_flags); in io_rw_import_reg_vec()
650 struct io_br_sel *sel, unsigned int issue_flags) in kiocb_done() argument
669 io_req_rw_cleanup(req, issue_flags); in kiocb_done()
912 unsigned int issue_flags) in __io_read() argument
914 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in __io_read()
922 ret = io_rw_import_reg_vec(req, io, ITER_DEST, issue_flags); in __io_read()
926 ret = io_import_rw_buffer(ITER_DEST, req, io, sel, issue_flags); in __io_read()
976 (issue_flags & IO_URING_F_MULTISHOT)) { in __io_read()
1026 int io_read(struct io_kiocb *req, unsigned int issue_flags) in io_read() argument
1031 ret = __io_read(req, &sel, issue_flags); in io_read()
1033 return kiocb_done(req, ret, &sel, issue_flags); in io_read()
1036 io_kbuf_recycle(req, sel.buf_list, issue_flags); in io_read()
1040 int io_read_mshot(struct io_kiocb *req, unsigned int issue_flags) in io_read_mshot() argument
1055 ret = __io_read(req, &sel, issue_flags); in io_read_mshot()
1066 if (io_kbuf_recycle(req, sel.buf_list, issue_flags)) in io_read_mshot()
1070 io_kbuf_recycle(req, sel.buf_list, issue_flags); in io_read_mshot()
1086 if (issue_flags & IO_URING_F_MULTISHOT) in io_read_mshot()
1103 io_req_rw_cleanup(req, issue_flags); in io_read_mshot()
1126 int io_write(struct io_kiocb *req, unsigned int issue_flags) in io_write() argument
1128 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_write()
1136 ret = io_rw_import_reg_vec(req, io, ITER_SOURCE, issue_flags); in io_write()
1211 return kiocb_done(req, ret2, NULL, issue_flags); in io_write()
1222 int io_read_fixed(struct io_kiocb *req, unsigned int issue_flags) in io_read_fixed() argument
1226 ret = io_init_rw_fixed(req, issue_flags, ITER_DEST); in io_read_fixed()
1230 return io_read(req, issue_flags); in io_read_fixed()
1233 int io_write_fixed(struct io_kiocb *req, unsigned int issue_flags) in io_write_fixed() argument
1237 ret = io_init_rw_fixed(req, issue_flags, ITER_SOURCE); in io_write_fixed()
1241 return io_write(req, issue_flags); in io_write_fixed()