| /linux/io_uring/ |
| A D | napi.h | 12 void io_napi_init(struct io_ring_ctx *ctx); 13 void io_napi_free(struct io_ring_ctx *ctx); 15 int io_register_napi(struct io_ring_ctx *ctx, void __user *arg); 18 void __io_napi_add(struct io_ring_ctx *ctx, struct socket *sock); 21 int io_napi_sqpoll_busy_poll(struct io_ring_ctx *ctx); 23 static inline bool io_napi(struct io_ring_ctx *ctx) in io_napi() 28 static inline void io_napi_busy_loop(struct io_ring_ctx *ctx, in io_napi_busy_loop() 44 struct io_ring_ctx *ctx = req->ctx; in io_napi_add() 57 static inline void io_napi_init(struct io_ring_ctx *ctx) in io_napi_init() 60 static inline void io_napi_free(struct io_ring_ctx *ctx) in io_napi_free() [all …]
|
| A D | io_uring.h | 40 struct io_ring_ctx *ctx; 57 struct io_ring_ctx *ctx = iowq->ctx; in io_should_wake() 69 int io_run_task_work_sig(struct io_ring_ctx *ctx); 74 void __io_commit_cqring_flush(struct io_ring_ctx *ctx); 93 struct io_ring_ctx *ctx); 102 void __io_submit_flush_completions(struct io_ring_ctx *ctx); 110 bool __io_alloc_req_refill(struct io_ring_ctx *ctx); 115 void io_activate_pollwq(struct io_ring_ctx *ctx); 266 static inline void io_cqring_wake(struct io_ring_ctx *ctx) in io_cqring_wake() 283 static inline bool io_sqring_full(struct io_ring_ctx *ctx) in io_sqring_full() [all …]
|
| A D | rsrc.h | 26 struct io_ring_ctx *ctx; 35 struct io_ring_ctx *ctx; 71 void __io_sqe_buffers_unregister(struct io_ring_ctx *ctx); 72 int io_sqe_buffers_unregister(struct io_ring_ctx *ctx); 75 void __io_sqe_files_unregister(struct io_ring_ctx *ctx); 76 int io_sqe_files_unregister(struct io_ring_ctx *ctx); 84 int io_register_rsrc(struct io_ring_ctx *ctx, void __user *arg, 95 static inline void io_charge_rsrc_node(struct io_ring_ctx *ctx, in io_charge_rsrc_node() 102 struct io_ring_ctx *ctx) in __io_req_set_rsrc_node() 110 struct io_ring_ctx *ctx, in io_req_set_rsrc_node() [all …]
|
| A D | futex.h | 12 int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, 14 bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task, 16 bool io_futex_cache_init(struct io_ring_ctx *ctx); 17 void io_futex_cache_free(struct io_ring_ctx *ctx); 19 static inline int io_futex_cancel(struct io_ring_ctx *ctx, in io_futex_cancel() 25 static inline bool io_futex_remove_all(struct io_ring_ctx *ctx, in io_futex_remove_all() 30 static inline bool io_futex_cache_init(struct io_ring_ctx *ctx) in io_futex_cache_init() 34 static inline void io_futex_cache_free(struct io_ring_ctx *ctx) in io_futex_cache_free()
|
| A D | tctx.h | 6 struct io_ring_ctx *ctx; 10 struct io_ring_ctx *ctx); 12 int __io_uring_add_tctx_node(struct io_ring_ctx *ctx); 13 int __io_uring_add_tctx_node_from_submit(struct io_ring_ctx *ctx); 17 int io_ringfd_register(struct io_ring_ctx *ctx, void __user *__arg, 19 int io_ringfd_unregister(struct io_ring_ctx *ctx, void __user *__arg, 25 static inline int io_uring_add_tctx_node(struct io_ring_ctx *ctx) in io_uring_add_tctx_node()
|
| A D | eventfd.h | 2 struct io_ring_ctx; 3 int io_eventfd_register(struct io_ring_ctx *ctx, void __user *arg, 5 int io_eventfd_unregister(struct io_ring_ctx *ctx); 7 void io_eventfd_flush_signal(struct io_ring_ctx *ctx); 8 void io_eventfd_signal(struct io_ring_ctx *ctx);
|
| A D | napi.c | 41 void __io_napi_add(struct io_ring_ctx *ctx, struct socket *sock) in __io_napi_add() 88 static void __io_napi_remove_stale(struct io_ring_ctx *ctx) in __io_napi_remove_stale() 104 static inline void io_napi_remove_stale(struct io_ring_ctx *ctx, bool is_stale) in io_napi_remove_stale() 139 static bool __io_napi_do_busy_loop(struct io_ring_ctx *ctx, in __io_napi_do_busy_loop() 160 static void io_napi_blocking_busy_loop(struct io_ring_ctx *ctx, in io_napi_blocking_busy_loop() 188 void io_napi_init(struct io_ring_ctx *ctx) in io_napi_init() 204 void io_napi_free(struct io_ring_ctx *ctx) in io_napi_free() 224 int io_register_napi(struct io_ring_ctx *ctx, void __user *arg) in io_register_napi() 256 int io_unregister_napi(struct io_ring_ctx *ctx, void __user *arg) in io_unregister_napi() 279 void __io_napi_busy_loop(struct io_ring_ctx *ctx, struct io_wait_queue *iowq) in __io_napi_busy_loop() [all …]
|
| A D | msg_ring.c | 36 static void io_double_unlock_ctx(struct io_ring_ctx *octx) in io_double_unlock_ctx() 41 static int io_double_lock_ctx(struct io_ring_ctx *octx, in io_double_lock_ctx() 76 struct io_ring_ctx *ctx = req->ctx; in io_msg_tw_complete() 106 static struct io_kiocb *io_msg_get_kiocb(struct io_ring_ctx *ctx) in io_msg_get_kiocb() 121 struct io_ring_ctx *target_ctx = req->file->private_data; in io_msg_data_remote() 139 struct io_ring_ctx *target_ctx = req->file->private_data; in io_msg_ring_data() 172 struct io_ring_ctx *ctx = req->ctx; in io_msg_grab_file() 189 struct io_ring_ctx *target_ctx = req->file->private_data; in io_msg_install_complete() 234 struct io_ring_ctx *ctx = req->file->private_data; in io_msg_fd_remote() 250 struct io_ring_ctx *target_ctx = req->file->private_data; in io_msg_send_fd() [all …]
|
| A D | sqpoll.h | 24 int io_sq_offload_create(struct io_ring_ctx *ctx, struct io_uring_params *p); 25 void io_sq_thread_finish(struct io_ring_ctx *ctx); 30 void io_sqpoll_wait_sq(struct io_ring_ctx *ctx); 31 int io_sqpoll_wq_cpu_affinity(struct io_ring_ctx *ctx, cpumask_var_t mask);
|
| A D | io_uring.c | 240 struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); in io_ring_ctx_ref_free() 247 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, in io_fallback_req_func() 278 struct io_ring_ctx *ctx; in io_ring_ctx_alloc() 1049 struct io_ring_ctx *ctx = NULL; in io_handle_tw_list() 1159 struct io_ring_ctx *ctx, in io_req_local_work_add() 2784 struct io_ring_ctx *ctx = container_of(cb, struct io_ring_ctx, in io_activate_pollwq_cb() 2861 struct io_ring_ctx *ctx; 2890 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, exit_work); in io_ring_exit_work() 3152 struct io_ring_ctx *ctx; in io_uring_cancel_generic() 3283 struct io_ring_ctx *ctx; in SYSCALL_DEFINE6() [all …]
|
| A D | rsrc.c | 159 struct io_ring_ctx *ctx = node->ctx; in io_rsrc_node_ref_zero() 195 struct io_ring_ctx *ctx) in io_rsrc_ref_quiesce() 331 static int __io_sqe_files_update(struct io_ring_ctx *ctx, in __io_sqe_files_update() 396 static int __io_sqe_buffers_update(struct io_ring_ctx *ctx, in __io_sqe_buffers_update() 599 struct io_ring_ctx *ctx = req->ctx; in io_files_update() 627 struct io_ring_ctx *ctx = data->ctx; in io_queue_rsrc_removal() 646 void __io_sqe_files_unregister(struct io_ring_ctx *ctx) in __io_sqe_files_unregister() 666 int io_sqe_files_unregister(struct io_ring_ctx *ctx) in io_sqe_files_unregister() 759 void __io_sqe_buffers_unregister(struct io_ring_ctx *ctx) in __io_sqe_buffers_unregister() 772 int io_sqe_buffers_unregister(struct io_ring_ctx *ctx) in io_sqe_buffers_unregister() [all …]
|
| A D | filetable.h | 13 int __io_fixed_fd_install(struct io_ring_ctx *ctx, struct file *file, 15 int io_fixed_fd_remove(struct io_ring_ctx *ctx, unsigned int offset); 17 int io_register_file_alloc_range(struct io_ring_ctx *ctx, 69 static inline void io_reset_alloc_hint(struct io_ring_ctx *ctx) in io_reset_alloc_hint() 74 static inline void io_file_table_set_alloc_range(struct io_ring_ctx *ctx, in io_file_table_set_alloc_range()
|
| A D | filetable.c | 16 static int io_file_bitmap_get(struct io_ring_ctx *ctx) in io_file_bitmap_get() 63 static int io_install_fixed_file(struct io_ring_ctx *ctx, struct file *file, in io_install_fixed_file() 96 int __io_fixed_fd_install(struct io_ring_ctx *ctx, struct file *file, in __io_fixed_fd_install() 123 struct io_ring_ctx *ctx = req->ctx; in io_fixed_fd_install() 135 int io_fixed_fd_remove(struct io_ring_ctx *ctx, unsigned int offset) in io_fixed_fd_remove() 160 int io_register_file_alloc_range(struct io_ring_ctx *ctx, in io_register_file_alloc_range()
|
| A D | kbuf.c | 42 static int io_buffer_add_list(struct io_ring_ctx *ctx, in io_buffer_add_list() 57 struct io_ring_ctx *ctx = req->ctx; in io_kbuf_recycle_legacy() 88 struct io_ring_ctx *ctx = req->ctx; in __io_put_kbuf() 177 struct io_ring_ctx *ctx = req->ctx; in io_buffer_select() 290 struct io_ring_ctx *ctx = req->ctx; in io_buffers_select() 322 struct io_ring_ctx *ctx = req->ctx; in io_buffers_peek() 343 static int __io_remove_buffers(struct io_ring_ctx *ctx, in __io_remove_buffers() 395 void io_destroy_buffers(struct io_ring_ctx *ctx) in io_destroy_buffers() 443 struct io_ring_ctx *ctx = req->ctx; in io_remove_buffers() 576 struct io_ring_ctx *ctx = req->ctx; in io_provide_buffers() [all …]
|
| A D | kbuf.h | 71 void io_destroy_buffers(struct io_ring_ctx *ctx); 79 int io_register_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg); 80 int io_unregister_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg); 81 int io_register_pbuf_status(struct io_ring_ctx *ctx, void __user *arg); 87 void io_put_bl(struct io_ring_ctx *ctx, struct io_buffer_list *bl); 88 struct io_buffer_list *io_pbuf_get_bl(struct io_ring_ctx *ctx,
|
| A D | futex.c | 36 bool io_futex_cache_init(struct io_ring_ctx *ctx) in io_futex_cache_init() 42 void io_futex_cache_free(struct io_ring_ctx *ctx) in io_futex_cache_free() 57 struct io_ring_ctx *ctx = req->ctx; in io_futex_complete() 93 static bool __io_futex_cancel(struct io_ring_ctx *ctx, struct io_kiocb *req) in __io_futex_cancel() 116 int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, in io_futex_cancel() 144 bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task, in io_futex_remove_all() 254 static struct io_futex_data *io_alloc_ifd(struct io_ring_ctx *ctx) in io_alloc_ifd() 269 struct io_ring_ctx *ctx = req->ctx; in io_futexv_wait() 323 struct io_ring_ctx *ctx = req->ctx; in io_futex_wait()
|
| A D | register.c | 35 static __cold int io_probe(struct io_ring_ctx *ctx, void __user *arg, in io_probe() 74 int io_unregister_personality(struct io_ring_ctx *ctx, unsigned id) in io_unregister_personality() 88 static int io_register_personality(struct io_ring_ctx *ctx) in io_register_personality() 105 static __cold int io_register_restrictions(struct io_ring_ctx *ctx, in io_register_restrictions() 175 static int io_register_enable_rings(struct io_ring_ctx *ctx) in io_register_enable_rings() 199 static __cold int __io_register_iowq_aff(struct io_ring_ctx *ctx, in __io_register_iowq_aff() 215 static __cold int io_register_iowq_aff(struct io_ring_ctx *ctx, in io_register_iowq_aff() 247 static __cold int io_unregister_iowq_aff(struct io_ring_ctx *ctx) in io_unregister_iowq_aff() 252 static __cold int io_register_iowq_max_workers(struct io_ring_ctx *ctx, in io_register_iowq_max_workers() 338 static int io_register_clock(struct io_ring_ctx *ctx, in io_register_clock() [all …]
|
| A D | timeout.h | 24 __cold void io_flush_timeouts(struct io_ring_ctx *ctx); 26 int io_timeout_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd); 27 __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
| A D | eventfd.c | 44 void io_eventfd_signal(struct io_ring_ctx *ctx) in io_eventfd_signal() 84 void io_eventfd_flush_signal(struct io_ring_ctx *ctx) in io_eventfd_flush_signal() 107 int io_eventfd_register(struct io_ring_ctx *ctx, void __user *arg, in io_eventfd_register() 146 int io_eventfd_unregister(struct io_ring_ctx *ctx) in io_eventfd_unregister()
|
| A D | timeout.c | 72 struct io_ring_ctx *ctx = req->ctx; in io_timeout_complete() 108 __cold void io_flush_timeouts(struct io_ring_ctx *ctx) in io_flush_timeouts() 202 struct io_ring_ctx *ctx = req->ctx; in io_disarm_next() 239 struct io_ring_ctx *ctx = req->ctx; in io_timeout_fn() 257 static struct io_kiocb *io_timeout_extract(struct io_ring_ctx *ctx, in io_timeout_extract() 284 int io_timeout_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd) in io_timeout_cancel() 329 struct io_ring_ctx *ctx = req->ctx; in io_link_timeout_fn() 398 static int io_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, in io_timeout_update() 462 struct io_ring_ctx *ctx = req->ctx; in io_timeout_remove() 569 struct io_ring_ctx *ctx = req->ctx; in io_timeout() [all …]
|
| A D | register.h | 5 int io_eventfd_unregister(struct io_ring_ctx *ctx); 6 int io_unregister_personality(struct io_ring_ctx *ctx, unsigned id);
|
| A D | waitid.h | 12 int io_waitid_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, 14 bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
| A D | tctx.c | 15 static struct io_wq *io_init_wq_offload(struct io_ring_ctx *ctx, in io_init_wq_offload() 61 struct io_ring_ctx *ctx) in io_uring_alloc_task_context() 94 int __io_uring_add_tctx_node(struct io_ring_ctx *ctx) in __io_uring_add_tctx_node() 136 int __io_uring_add_tctx_node_from_submit(struct io_ring_ctx *ctx) in __io_uring_add_tctx_node_from_submit() 253 int io_ringfd_register(struct io_ring_ctx *ctx, void __user *__arg, in io_ringfd_register() 312 int io_ringfd_unregister(struct io_ring_ctx *ctx, void __user *__arg, in io_ringfd_unregister()
|
| A D | sqpoll.c | 82 struct io_ring_ctx *ctx; in io_sqd_update_thread_idle() 90 void io_sq_thread_finish(struct io_ring_ctx *ctx) in io_sq_thread_finish() 107 struct io_ring_ctx *ctx_attach; in io_attach_sq_data() 170 static int __io_sq_thread(struct io_ring_ctx *ctx, bool cap_entries) in __io_sq_thread() 271 struct io_ring_ctx *ctx; in io_sq_thread() 395 void io_sqpoll_wait_sq(struct io_ring_ctx *ctx) in io_sqpoll_wait_sq() 412 __cold int io_sq_offload_create(struct io_ring_ctx *ctx, in io_sq_offload_create() 513 __cold int io_sqpoll_wq_cpu_affinity(struct io_ring_ctx *ctx, in io_sqpoll_wq_cpu_affinity()
|
| A D | uring_cmd.c | 21 struct io_ring_ctx *ctx = req->ctx; in io_uring_async_get() 49 bool io_uring_try_cancel_uring_cmd(struct io_ring_ctx *ctx, in io_uring_try_cancel_uring_cmd() 84 struct io_ring_ctx *ctx = req->ctx; in io_uring_cmd_del_cancelable() 108 struct io_ring_ctx *ctx = req->ctx; in io_uring_cmd_mark_cancelable() 211 struct io_ring_ctx *ctx = req->ctx; in io_uring_cmd_prep() 229 struct io_ring_ctx *ctx = req->ctx; in io_uring_cmd()
|