Lines Matching refs:ring_pages
125 struct page **ring_pages; member
340 page_count(ctx->ring_pages[i])); in aio_free_ring()
341 page = ctx->ring_pages[i]; in aio_free_ring()
344 ctx->ring_pages[i] = NULL; in aio_free_ring()
348 if (ctx->ring_pages && ctx->ring_pages != ctx->internal_pages) { in aio_free_ring()
349 kfree(ctx->ring_pages); in aio_free_ring()
350 ctx->ring_pages = NULL; in aio_free_ring()
441 if (ctx->ring_pages[idx] != &src->page) in aio_migrate_folio()
465 BUG_ON(ctx->ring_pages[idx] != &src->page); in aio_migrate_folio()
466 ctx->ring_pages[idx] = &dst->page; in aio_migrate_folio()
516 ctx->ring_pages = ctx->internal_pages; in aio_setup_ring()
518 ctx->ring_pages = kcalloc(nr_pages, sizeof(struct page *), in aio_setup_ring()
520 if (!ctx->ring_pages) { in aio_setup_ring()
537 ctx->ring_pages[i] = page; in aio_setup_ring()
570 ring = kmap_atomic(ctx->ring_pages[0]); in aio_setup_ring()
579 flush_dcache_page(ctx->ring_pages[0]); in aio_setup_ring()
681 ring = kmap_atomic(ctx->ring_pages[0]); in ioctx_add_table()
1024 ring = kmap_atomic(ctx->ring_pages[0]); in user_refill_reqs_available()
1132 ev_page = kmap_atomic(ctx->ring_pages[pos / AIO_EVENTS_PER_PAGE]); in aio_complete()
1138 flush_dcache_page(ctx->ring_pages[pos / AIO_EVENTS_PER_PAGE]); in aio_complete()
1151 ring = kmap_atomic(ctx->ring_pages[0]); in aio_complete()
1155 flush_dcache_page(ctx->ring_pages[0]); in aio_complete()
1214 ring = kmap_atomic(ctx->ring_pages[0]); in aio_read_events_ring()
1243 page = ctx->ring_pages[pos / AIO_EVENTS_PER_PAGE]; in aio_read_events_ring()
1264 ring = kmap_atomic(ctx->ring_pages[0]); in aio_read_events_ring()
1267 flush_dcache_page(ctx->ring_pages[0]); in aio_read_events_ring()