Lines Matching refs:t_page

654 	struct nullb_page *t_page;  in null_alloc_page()  local
656 t_page = kmalloc(sizeof(struct nullb_page), gfp_flags); in null_alloc_page()
657 if (!t_page) in null_alloc_page()
660 t_page->page = alloc_pages(gfp_flags, 0); in null_alloc_page()
661 if (!t_page->page) in null_alloc_page()
664 memset(t_page->bitmap, 0, sizeof(t_page->bitmap)); in null_alloc_page()
665 return t_page; in null_alloc_page()
667 kfree(t_page); in null_alloc_page()
672 static void null_free_page(struct nullb_page *t_page) in null_free_page() argument
674 __set_bit(NULLB_PAGE_FREE, t_page->bitmap); in null_free_page()
675 if (test_bit(NULLB_PAGE_LOCK, t_page->bitmap)) in null_free_page()
677 __free_page(t_page->page); in null_free_page()
678 kfree(t_page); in null_free_page()
693 struct nullb_page *t_page, *ret; in null_free_sector() local
700 t_page = radix_tree_lookup(root, idx); in null_free_sector()
701 if (t_page) { in null_free_sector()
702 __clear_bit(sector_bit, t_page->bitmap); in null_free_sector()
704 if (null_page_empty(t_page)) { in null_free_sector()
705 ret = radix_tree_delete_item(root, idx, t_page); in null_free_sector()
706 WARN_ON(ret != t_page); in null_free_sector()
715 struct nullb_page *t_page, bool is_cache) in null_radix_tree_insert() argument
721 if (radix_tree_insert(root, idx, t_page)) { in null_radix_tree_insert()
722 null_free_page(t_page); in null_radix_tree_insert()
723 t_page = radix_tree_lookup(root, idx); in null_radix_tree_insert()
724 WARN_ON(!t_page || t_page->page->index != idx); in null_radix_tree_insert()
728 return t_page; in null_radix_tree_insert()
765 struct nullb_page *t_page; in __null_lookup_page() local
772 t_page = radix_tree_lookup(root, idx); in __null_lookup_page()
773 WARN_ON(t_page && t_page->page->index != idx); in __null_lookup_page()
775 if (t_page && (for_write || test_bit(sector_bit, t_page->bitmap))) in __null_lookup_page()
776 return t_page; in __null_lookup_page()
799 struct nullb_page *t_page; in null_insert_page() local
801 t_page = null_lookup_page(nullb, sector, true, ignore_cache); in null_insert_page()
802 if (t_page) in null_insert_page()
803 return t_page; in null_insert_page()
807 t_page = null_alloc_page(GFP_NOIO); in null_insert_page()
808 if (!t_page) in null_insert_page()
816 t_page->page->index = idx; in null_insert_page()
817 t_page = null_radix_tree_insert(nullb, idx, t_page, !ignore_cache); in null_insert_page()
820 return t_page; in null_insert_page()
822 null_free_page(t_page); in null_insert_page()
833 struct nullb_page *t_page, *ret; in null_flush_cache_page() local
838 t_page = null_insert_page(nullb, idx << PAGE_SECTORS_SHIFT, true); in null_flush_cache_page()
843 if (t_page && null_page_empty(t_page)) { in null_flush_cache_page()
845 idx, t_page); in null_flush_cache_page()
846 null_free_page(t_page); in null_flush_cache_page()
851 if (!t_page) in null_flush_cache_page()
855 dst = kmap_atomic(t_page->page); in null_flush_cache_page()
863 __set_bit(i, t_page->bitmap); in null_flush_cache_page()
935 struct nullb_page *t_page; in copy_to_nullb() local
945 t_page = null_insert_page(nullb, sector, in copy_to_nullb()
947 if (!t_page) in copy_to_nullb()
951 dst = kmap_atomic(t_page->page); in copy_to_nullb()
956 __set_bit(sector & SECTOR_MASK, t_page->bitmap); in copy_to_nullb()
972 struct nullb_page *t_page; in copy_from_nullb() local
979 t_page = null_lookup_page(nullb, sector, false, in copy_from_nullb()
983 if (!t_page) { in copy_from_nullb()
987 src = kmap_atomic(t_page->page); in copy_from_nullb()