Lines Matching refs:user_data

872 	u64				user_data;  member
1082 static bool io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data,
1487 io_cqring_fill_event(req->ctx, req->user_data, status, 0); in io_kill_timeout()
1724 static bool io_cqring_event_overflow(struct io_ring_ctx *ctx, u64 user_data, in io_cqring_event_overflow() argument
1745 ocqe->cqe.user_data = user_data; in io_cqring_event_overflow()
1752 static inline bool __io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data, in __io_cqring_fill_event() argument
1757 trace_io_uring_complete(ctx, user_data, res, cflags); in __io_cqring_fill_event()
1766 WRITE_ONCE(cqe->user_data, user_data); in __io_cqring_fill_event()
1771 return io_cqring_event_overflow(ctx, user_data, res, cflags); in __io_cqring_fill_event()
1775 static noinline bool io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data, in io_cqring_fill_event() argument
1778 return __io_cqring_fill_event(ctx, user_data, res, cflags); in io_cqring_fill_event()
1787 __io_cqring_fill_event(ctx, req->user_data, res, cflags); in io_req_complete_post()
2013 io_cqring_fill_event(link->ctx, link->user_data, in io_kill_linked_timeout()
2038 io_cqring_fill_event(link->ctx, link->user_data, res, 0); in io_fail_links()
2055 io_cqring_fill_event(link->ctx, link->user_data, in io_disarm_next()
2328 __io_cqring_fill_event(ctx, req->user_data, req->result, in io_submit_flush_completions()
2440 __io_cqring_fill_event(ctx, req->user_data, req->result, in io_iopoll_complete()
5251 trace_io_uring_task_add(req->ctx, req->opcode, req->user_data, mask); in __io_async_wake()
5341 if (!io_cqring_fill_event(ctx, req->user_data, error, flags)) { in __io_poll_complete()
5508 trace_io_uring_task_run(req->ctx, req, req->opcode, req->user_data); in io_async_task_func()
5532 trace_io_uring_poll_wake(req->ctx, req->opcode, req->user_data, in io_async_wake()
5543 list = &ctx->cancel_hash[hash_long(req->user_data, ctx->cancel_hash_bits)]; in io_poll_req_insert()
5647 trace_io_uring_poll_arm(ctx, req, req->opcode, req->user_data, in io_arm_poll_handler()
5681 io_cqring_fill_event(req->ctx, req->user_data, -ECANCELED, 0); in io_poll_remove_one()
5726 if (sqe_addr != req->user_data) in io_poll_find()
5905 preq->user_data = req->poll_update.new_user_data; in io_poll_update()
5947 __u64 user_data) in io_timeout_extract() argument
5955 found = user_data == req->user_data; in io_timeout_extract()
5969 static int io_timeout_cancel(struct io_ring_ctx *ctx, __u64 user_data) in io_timeout_cancel() argument
5973 struct io_kiocb *req = io_timeout_extract(ctx, user_data); in io_timeout_cancel()
5979 io_cqring_fill_event(ctx, req->user_data, -ECANCELED, 0); in io_timeout_cancel()
6000 static int io_linked_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, in io_linked_timeout_update() argument
6009 found = user_data == req->user_data; in io_linked_timeout_update()
6025 static int io_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, in io_timeout_update() argument
6029 struct io_kiocb *req = io_timeout_extract(ctx, user_data); in io_timeout_update()
6218 u64 user_data; member
6226 return req->ctx == cd->ctx && req->user_data == cd->user_data; in io_cancel_cb()
6229 static int io_async_cancel_one(struct io_uring_task *tctx, u64 user_data, in io_async_cancel_one() argument
6232 struct io_cancel_data data = { .ctx = ctx, .user_data = user_data, }; in io_async_cancel_one()
6545 trace_io_uring_defer(ctx, req, req->user_data); in io_drain_req()
6890 ret = io_try_cancel_userdata(req, prev->user_data); in io_req_task_link_timeout()
7059 req->user_data = READ_ONCE(sqe->user_data); in io_init_req()
7148 trace_io_uring_submit_sqe(ctx, req, req->opcode, req->user_data, in io_submit_sqe()
10988 BUILD_BUG_SQE_ELEM(32, __u64, user_data); in io_uring_init()