Lines Matching refs:io_ring_ctx

145 static bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
182 struct io_ring_ctx *ctx = file->private_data; in io_uring_get_socket()
191 static inline void io_submit_flush_completions(struct io_ring_ctx *ctx) in io_submit_flush_completions()
198 static inline unsigned int __io_cqring_events(struct io_ring_ctx *ctx) in __io_cqring_events()
203 static inline unsigned int __io_cqring_events_user(struct io_ring_ctx *ctx) in __io_cqring_events_user()
234 struct io_ring_ctx *ctx = head->ctx; in io_match_task_safe()
252 static inline void io_req_add_to_cache(struct io_kiocb *req, struct io_ring_ctx *ctx) in io_req_add_to_cache()
259 struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); in io_ring_ctx_ref_free()
266 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, in io_fallback_req_func()
295 static __cold struct io_ring_ctx *io_ring_ctx_alloc(struct io_uring_params *p) in io_ring_ctx_alloc()
297 struct io_ring_ctx *ctx; in io_ring_ctx_alloc()
363 static void io_account_cq_overflow(struct io_ring_ctx *ctx) in io_account_cq_overflow()
374 struct io_ring_ctx *ctx = req->ctx; in req_need_defer()
458 struct io_ring_ctx *ctx = req->ctx; in io_prep_async_work()
494 struct io_ring_ctx *ctx = req->ctx; in io_prep_async_link()
533 static __cold void io_queue_deferred(struct io_ring_ctx *ctx) in io_queue_deferred()
566 static void io_eventfd_signal(struct io_ring_ctx *ctx) in io_eventfd_signal()
603 static void io_eventfd_flush_signal(struct io_ring_ctx *ctx) in io_eventfd_flush_signal()
626 void __io_commit_cqring_flush(struct io_ring_ctx *ctx) in __io_commit_cqring_flush()
641 static inline void __io_cq_lock(struct io_ring_ctx *ctx) in __io_cq_lock()
647 static inline void io_cq_lock(struct io_ring_ctx *ctx) in io_cq_lock()
653 static inline void __io_cq_unlock_post(struct io_ring_ctx *ctx) in __io_cq_unlock_post()
666 static void io_cq_unlock_post(struct io_ring_ctx *ctx) in io_cq_unlock_post()
676 static void io_cqring_overflow_kill(struct io_ring_ctx *ctx) in io_cqring_overflow_kill()
693 static void __io_cqring_overflow_flush(struct io_ring_ctx *ctx) in __io_cqring_overflow_flush()
724 static void io_cqring_do_overflow_flush(struct io_ring_ctx *ctx) in io_cqring_do_overflow_flush()
734 static void io_cqring_overflow_flush(struct io_ring_ctx *ctx) in io_cqring_overflow_flush()
787 static bool io_cqring_event_overflow(struct io_ring_ctx *ctx, u64 user_data, in io_cqring_event_overflow()
840 bool io_cqe_cache_refill(struct io_ring_ctx *ctx, bool overflow) in io_cqe_cache_refill()
872 static bool io_fill_cqe_aux(struct io_ring_ctx *ctx, u64 user_data, s32 res, in io_fill_cqe_aux()
900 static void __io_flush_post_cqes(struct io_ring_ctx *ctx) in __io_flush_post_cqes()
925 static bool __io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags, in __io_post_aux_cqe()
939 bool io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags) in io_post_aux_cqe()
950 struct io_ring_ctx *ctx = req->ctx; in io_fill_cqe_req_aux()
982 struct io_ring_ctx *ctx = req->ctx; in __io_req_complete_post()
1037 struct io_ring_ctx *ctx = req->ctx; in io_req_complete_post()
1063 static void io_preinit_req(struct io_kiocb *req, struct io_ring_ctx *ctx) in io_preinit_req()
1073 static void io_flush_cached_locked_reqs(struct io_ring_ctx *ctx, in io_flush_cached_locked_reqs()
1088 __cold bool __io_alloc_req_refill(struct io_ring_ctx *ctx) in __io_alloc_req_refill()
1141 struct io_ring_ctx *ctx = req->ctx; in __io_req_find_next_prep()
1165 static void ctx_flush_and_put(struct io_ring_ctx *ctx, struct io_tw_state *ts) in ctx_flush_and_put()
1180 struct io_ring_ctx **ctx, in handle_tw_list()
1248 struct io_ring_ctx *last_ctx = NULL; in io_fallback_tw()
1276 struct io_ring_ctx *ctx = NULL; in tctx_task_work()
1316 struct io_ring_ctx *ctx = req->ctx; in io_req_local_work_add()
1368 struct io_ring_ctx *ctx = req->ctx; in io_req_normal_work_add()
1394 static void __cold io_move_task_work_from_local(struct io_ring_ctx *ctx) in io_move_task_work_from_local()
1408 static int __io_run_local_work(struct io_ring_ctx *ctx, struct io_tw_state *ts) in __io_run_local_work()
1448 static inline int io_run_local_work_locked(struct io_ring_ctx *ctx) in io_run_local_work_locked()
1463 static int io_run_local_work(struct io_ring_ctx *ctx) in io_run_local_work()
1515 static void io_free_batch_list(struct io_ring_ctx *ctx, in io_free_batch_list()
1553 void __io_submit_flush_completions(struct io_ring_ctx *ctx) in __io_submit_flush_completions()
1586 static unsigned io_cqring_events(struct io_ring_ctx *ctx) in io_cqring_events()
1597 static __cold void io_iopoll_try_reap_events(struct io_ring_ctx *ctx) in io_iopoll_try_reap_events()
1621 static int io_iopoll_check(struct io_ring_ctx *ctx, long min) in io_iopoll_check()
1709 struct io_ring_ctx *ctx = req->ctx; in io_iopoll_req_issued()
1811 struct io_ring_ctx *ctx = req->ctx; in io_drain_req()
1999 struct io_ring_ctx *ctx = req->ctx; in io_file_get_fixed()
2106 static inline bool io_check_restriction(struct io_ring_ctx *ctx, in io_check_restriction()
2126 struct io_ring_ctx *ctx = req->ctx; in io_init_req_drain()
2143 static int io_init_req(struct io_ring_ctx *ctx, struct io_kiocb *req, in io_init_req()
2240 struct io_ring_ctx *ctx = req->ctx; in io_submit_fail_init()
2274 static inline int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req, in io_submit_sqe()
2330 static void io_submit_state_end(struct io_ring_ctx *ctx) in io_submit_state_end()
2355 static void io_commit_sqring(struct io_ring_ctx *ctx) in io_commit_sqring()
2375 static bool io_get_sqe(struct io_ring_ctx *ctx, const struct io_uring_sqe **sqe) in io_get_sqe()
2409 int io_submit_sqes(struct io_ring_ctx *ctx, unsigned int nr) in io_submit_sqes()
2461 struct io_ring_ctx *ctx;
2467 static inline bool io_has_work(struct io_ring_ctx *ctx) in io_has_work()
2475 struct io_ring_ctx *ctx = iowq->ctx; in io_should_wake()
2500 int io_run_task_work_sig(struct io_ring_ctx *ctx) in io_run_task_work_sig()
2524 static inline int io_cqring_wait_schedule(struct io_ring_ctx *ctx, in io_cqring_wait_schedule()
2561 static int io_cqring_wait(struct io_ring_ctx *ctx, int min_events, in io_cqring_wait()
2742 static void *io_rings_map(struct io_ring_ctx *ctx, unsigned long uaddr, in io_rings_map()
2749 static void *io_sqes_map(struct io_ring_ctx *ctx, unsigned long uaddr, in io_sqes_map()
2756 static void io_rings_free(struct io_ring_ctx *ctx) in io_rings_free()
2782 static unsigned long rings_size(struct io_ring_ctx *ctx, unsigned int sq_entries, in rings_size()
2821 static int io_eventfd_register(struct io_ring_ctx *ctx, void __user *arg, in io_eventfd_register()
2859 static int io_eventfd_unregister(struct io_ring_ctx *ctx) in io_eventfd_unregister()
2876 static void io_req_caches_free(struct io_ring_ctx *ctx) in io_req_caches_free()
2899 static __cold void io_ring_ctx_free(struct io_ring_ctx *ctx) in io_ring_ctx_free()
2957 struct io_ring_ctx *ctx = container_of(cb, struct io_ring_ctx, in io_activate_pollwq_cb()
2972 static __cold void io_activate_pollwq(struct io_ring_ctx *ctx) in io_activate_pollwq()
2996 struct io_ring_ctx *ctx = file->private_data; in io_uring_poll()
3031 static int io_unregister_personality(struct io_ring_ctx *ctx, unsigned id) in io_unregister_personality()
3047 struct io_ring_ctx *ctx;
3076 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, exit_work); in io_ring_exit_work()
3174 static __cold void io_ring_ctx_wait_and_kill(struct io_ring_ctx *ctx) in io_ring_ctx_wait_and_kill()
3208 struct io_ring_ctx *ctx = file->private_data; in io_uring_release()
3228 static __cold bool io_cancel_defer_files(struct io_ring_ctx *ctx, in io_cancel_defer_files()
3255 static __cold bool io_uring_try_cancel_iowq(struct io_ring_ctx *ctx) in io_uring_try_cancel_iowq()
3279 static __cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx, in io_uring_try_cancel_requests()
3347 struct io_ring_ctx *ctx; in io_uring_cancel_generic()
3431 struct io_ring_ctx *ctx = file->private_data; in io_uring_validate_mmap_request()
3605 struct io_ring_ctx *ctx; in SYSCALL_DEFINE6()
3755 static __cold int io_allocate_scq_urings(struct io_ring_ctx *ctx, in io_allocate_scq_urings()
3826 static struct file *io_uring_get_file(struct io_ring_ctx *ctx) in io_uring_get_file()
3854 struct io_ring_ctx *ctx; in io_uring_create()
4115 static __cold int io_probe(struct io_ring_ctx *ctx, void __user *arg, in io_probe()
4155 static int io_register_personality(struct io_ring_ctx *ctx) in io_register_personality()
4172 static __cold int io_register_restrictions(struct io_ring_ctx *ctx, in io_register_restrictions()
4242 static int io_register_enable_rings(struct io_ring_ctx *ctx) in io_register_enable_rings()
4266 static __cold int __io_register_iowq_aff(struct io_ring_ctx *ctx, in __io_register_iowq_aff()
4282 static __cold int io_register_iowq_aff(struct io_ring_ctx *ctx, in io_register_iowq_aff()
4313 static __cold int io_unregister_iowq_aff(struct io_ring_ctx *ctx) in io_unregister_iowq_aff()
4318 static __cold int io_register_iowq_max_workers(struct io_ring_ctx *ctx, in io_register_iowq_max_workers()
4401 static int __io_uring_register(struct io_ring_ctx *ctx, unsigned opcode, in __io_uring_register()
4570 struct io_ring_ctx *ctx; in SYSCALL_DEFINE4()