Lines Matching refs:t_page
734 struct nullb_page *t_page; in null_alloc_page() local
736 t_page = kmalloc(sizeof(struct nullb_page), gfp_flags); in null_alloc_page()
737 if (!t_page) in null_alloc_page()
740 t_page->page = alloc_pages(gfp_flags, 0); in null_alloc_page()
741 if (!t_page->page) in null_alloc_page()
744 memset(t_page->bitmap, 0, sizeof(t_page->bitmap)); in null_alloc_page()
745 return t_page; in null_alloc_page()
747 kfree(t_page); in null_alloc_page()
752 static void null_free_page(struct nullb_page *t_page) in null_free_page() argument
754 __set_bit(NULLB_PAGE_FREE, t_page->bitmap); in null_free_page()
755 if (test_bit(NULLB_PAGE_LOCK, t_page->bitmap)) in null_free_page()
757 __free_page(t_page->page); in null_free_page()
758 kfree(t_page); in null_free_page()
773 struct nullb_page *t_page, *ret; in null_free_sector() local
780 t_page = radix_tree_lookup(root, idx); in null_free_sector()
781 if (t_page) { in null_free_sector()
782 __clear_bit(sector_bit, t_page->bitmap); in null_free_sector()
784 if (null_page_empty(t_page)) { in null_free_sector()
785 ret = radix_tree_delete_item(root, idx, t_page); in null_free_sector()
786 WARN_ON(ret != t_page); in null_free_sector()
795 struct nullb_page *t_page, bool is_cache) in null_radix_tree_insert() argument
801 if (radix_tree_insert(root, idx, t_page)) { in null_radix_tree_insert()
802 null_free_page(t_page); in null_radix_tree_insert()
803 t_page = radix_tree_lookup(root, idx); in null_radix_tree_insert()
804 WARN_ON(!t_page || t_page->page->index != idx); in null_radix_tree_insert()
808 return t_page; in null_radix_tree_insert()
845 struct nullb_page *t_page; in __null_lookup_page() local
852 t_page = radix_tree_lookup(root, idx); in __null_lookup_page()
853 WARN_ON(t_page && t_page->page->index != idx); in __null_lookup_page()
855 if (t_page && (for_write || test_bit(sector_bit, t_page->bitmap))) in __null_lookup_page()
856 return t_page; in __null_lookup_page()
879 struct nullb_page *t_page; in null_insert_page() local
881 t_page = null_lookup_page(nullb, sector, true, ignore_cache); in null_insert_page()
882 if (t_page) in null_insert_page()
883 return t_page; in null_insert_page()
887 t_page = null_alloc_page(GFP_NOIO); in null_insert_page()
888 if (!t_page) in null_insert_page()
896 t_page->page->index = idx; in null_insert_page()
897 t_page = null_radix_tree_insert(nullb, idx, t_page, !ignore_cache); in null_insert_page()
900 return t_page; in null_insert_page()
902 null_free_page(t_page); in null_insert_page()
913 struct nullb_page *t_page, *ret; in null_flush_cache_page() local
918 t_page = null_insert_page(nullb, idx << PAGE_SECTORS_SHIFT, true); in null_flush_cache_page()
923 if (t_page && null_page_empty(t_page)) { in null_flush_cache_page()
925 idx, t_page); in null_flush_cache_page()
926 null_free_page(t_page); in null_flush_cache_page()
931 if (!t_page) in null_flush_cache_page()
935 dst = kmap_atomic(t_page->page); in null_flush_cache_page()
943 __set_bit(i, t_page->bitmap); in null_flush_cache_page()
1015 struct nullb_page *t_page; in copy_to_nullb() local
1025 t_page = null_insert_page(nullb, sector, in copy_to_nullb()
1027 if (!t_page) in copy_to_nullb()
1031 dst = kmap_atomic(t_page->page); in copy_to_nullb()
1036 __set_bit(sector & SECTOR_MASK, t_page->bitmap); in copy_to_nullb()
1052 struct nullb_page *t_page; in copy_from_nullb() local
1059 t_page = null_lookup_page(nullb, sector, false, in copy_from_nullb()
1063 if (!t_page) { in copy_from_nullb()
1067 src = kmap_atomic(t_page->page); in copy_from_nullb()