Lines Matching refs:iodev_sqe

511 	void (*submit)(struct rtio_iodev_sqe *iodev_sqe);
728 struct rtio_iodev_sqe *iodev_sqe = CONTAINER_OF(node, struct rtio_iodev_sqe, q); in rtio_sqe_pool_alloc() local
732 return iodev_sqe; in rtio_sqe_pool_alloc()
735 static inline void rtio_sqe_pool_free(struct rtio_sqe_pool *pool, struct rtio_iodev_sqe *iodev_sqe) in rtio_sqe_pool_free() argument
737 mpsc_push(&pool->free_q, &iodev_sqe->q); in rtio_sqe_pool_free()
954 static inline struct rtio_iodev_sqe *rtio_txn_next(const struct rtio_iodev_sqe *iodev_sqe) in rtio_txn_next() argument
956 if (iodev_sqe->sqe.flags & RTIO_SQE_TRANSACTION) { in rtio_txn_next()
957 return iodev_sqe->next; in rtio_txn_next()
972 static inline struct rtio_iodev_sqe *rtio_chain_next(const struct rtio_iodev_sqe *iodev_sqe) in rtio_chain_next() argument
974 if (iodev_sqe->sqe.flags & RTIO_SQE_CHAINED) { in rtio_chain_next()
975 return iodev_sqe->next; in rtio_chain_next()
989 static inline struct rtio_iodev_sqe *rtio_iodev_sqe_next(const struct rtio_iodev_sqe *iodev_sqe) in rtio_iodev_sqe_next() argument
991 return iodev_sqe->next; in rtio_iodev_sqe_next()
1004 struct rtio_iodev_sqe *iodev_sqe = rtio_sqe_pool_alloc(r->sqe_pool); in rtio_sqe_acquire() local
1006 if (iodev_sqe == NULL) { in rtio_sqe_acquire()
1010 mpsc_push(&r->sq, &iodev_sqe->q); in rtio_sqe_acquire()
1012 return &iodev_sqe->sqe; in rtio_sqe_acquire()
1022 struct rtio_iodev_sqe *iodev_sqe; in rtio_sqe_drop_all() local
1026 iodev_sqe = CONTAINER_OF(node, struct rtio_iodev_sqe, q); in rtio_sqe_drop_all()
1027 rtio_sqe_pool_free(r->sqe_pool, iodev_sqe); in rtio_sqe_drop_all()
1132 static inline uint32_t rtio_cqe_compute_flags(struct rtio_iodev_sqe *iodev_sqe) in rtio_cqe_compute_flags() argument
1137 if (iodev_sqe->sqe.op == RTIO_OP_RX && iodev_sqe->sqe.flags & RTIO_SQE_MEMPOOL_BUFFER) { in rtio_cqe_compute_flags()
1138 struct rtio *r = iodev_sqe->r; in rtio_cqe_compute_flags()
1143 if (iodev_sqe->sqe.rx.buf) { in rtio_cqe_compute_flags()
1144 blk_index = (iodev_sqe->sqe.rx.buf - mem_pool->buffer) >> in rtio_cqe_compute_flags()
1146 blk_count = iodev_sqe->sqe.rx.buf_len >> mem_pool->info.blk_sz_shift; in rtio_cqe_compute_flags()
1151 ARG_UNUSED(iodev_sqe); in rtio_cqe_compute_flags()
1209 void rtio_executor_ok(struct rtio_iodev_sqe *iodev_sqe, int result);
1210 void rtio_executor_err(struct rtio_iodev_sqe *iodev_sqe, int result);
1220 static inline void rtio_iodev_sqe_ok(struct rtio_iodev_sqe *iodev_sqe, int result) in rtio_iodev_sqe_ok() argument
1222 rtio_executor_ok(iodev_sqe, result); in rtio_iodev_sqe_ok()
1233 static inline void rtio_iodev_sqe_err(struct rtio_iodev_sqe *iodev_sqe, int result) in rtio_iodev_sqe_err() argument
1235 rtio_executor_err(iodev_sqe, result); in rtio_iodev_sqe_err()
1299 static inline int rtio_sqe_rx_buf(const struct rtio_iodev_sqe *iodev_sqe, uint32_t min_buf_len, in rtio_sqe_rx_buf() argument
1302 struct rtio_sqe *sqe = (struct rtio_sqe *)&iodev_sqe->sqe; in rtio_sqe_rx_buf()
1306 struct rtio *r = iodev_sqe->r; in rtio_sqe_rx_buf()
1400 struct rtio_iodev_sqe *iodev_sqe = CONTAINER_OF(sqe, struct rtio_iodev_sqe, sqe); in z_impl_rtio_sqe_cancel() local
1403 iodev_sqe->sqe.flags |= RTIO_SQE_CANCELED; in z_impl_rtio_sqe_cancel()
1404 iodev_sqe = rtio_iodev_sqe_next(iodev_sqe); in z_impl_rtio_sqe_cancel()
1405 } while (iodev_sqe != NULL); in z_impl_rtio_sqe_cancel()