Lines Matching refs:t_page
667 struct nullb_page *t_page; in null_alloc_page() local
669 t_page = kmalloc(sizeof(struct nullb_page), gfp_flags); in null_alloc_page()
670 if (!t_page) in null_alloc_page()
673 t_page->page = alloc_pages(gfp_flags, 0); in null_alloc_page()
674 if (!t_page->page) in null_alloc_page()
677 memset(t_page->bitmap, 0, sizeof(t_page->bitmap)); in null_alloc_page()
678 return t_page; in null_alloc_page()
680 kfree(t_page); in null_alloc_page()
685 static void null_free_page(struct nullb_page *t_page) in null_free_page() argument
687 __set_bit(NULLB_PAGE_FREE, t_page->bitmap); in null_free_page()
688 if (test_bit(NULLB_PAGE_LOCK, t_page->bitmap)) in null_free_page()
690 __free_page(t_page->page); in null_free_page()
691 kfree(t_page); in null_free_page()
706 struct nullb_page *t_page, *ret; in null_free_sector() local
713 t_page = radix_tree_lookup(root, idx); in null_free_sector()
714 if (t_page) { in null_free_sector()
715 __clear_bit(sector_bit, t_page->bitmap); in null_free_sector()
717 if (null_page_empty(t_page)) { in null_free_sector()
718 ret = radix_tree_delete_item(root, idx, t_page); in null_free_sector()
719 WARN_ON(ret != t_page); in null_free_sector()
728 struct nullb_page *t_page, bool is_cache) in null_radix_tree_insert() argument
734 if (radix_tree_insert(root, idx, t_page)) { in null_radix_tree_insert()
735 null_free_page(t_page); in null_radix_tree_insert()
736 t_page = radix_tree_lookup(root, idx); in null_radix_tree_insert()
737 WARN_ON(!t_page || t_page->page->index != idx); in null_radix_tree_insert()
741 return t_page; in null_radix_tree_insert()
778 struct nullb_page *t_page; in __null_lookup_page() local
785 t_page = radix_tree_lookup(root, idx); in __null_lookup_page()
786 WARN_ON(t_page && t_page->page->index != idx); in __null_lookup_page()
788 if (t_page && (for_write || test_bit(sector_bit, t_page->bitmap))) in __null_lookup_page()
789 return t_page; in __null_lookup_page()
812 struct nullb_page *t_page; in null_insert_page() local
814 t_page = null_lookup_page(nullb, sector, true, ignore_cache); in null_insert_page()
815 if (t_page) in null_insert_page()
816 return t_page; in null_insert_page()
820 t_page = null_alloc_page(GFP_NOIO); in null_insert_page()
821 if (!t_page) in null_insert_page()
829 t_page->page->index = idx; in null_insert_page()
830 t_page = null_radix_tree_insert(nullb, idx, t_page, !ignore_cache); in null_insert_page()
833 return t_page; in null_insert_page()
835 null_free_page(t_page); in null_insert_page()
846 struct nullb_page *t_page, *ret; in null_flush_cache_page() local
851 t_page = null_insert_page(nullb, idx << PAGE_SECTORS_SHIFT, true); in null_flush_cache_page()
856 if (t_page && null_page_empty(t_page)) { in null_flush_cache_page()
858 idx, t_page); in null_flush_cache_page()
859 null_free_page(t_page); in null_flush_cache_page()
864 if (!t_page) in null_flush_cache_page()
868 dst = kmap_atomic(t_page->page); in null_flush_cache_page()
876 __set_bit(i, t_page->bitmap); in null_flush_cache_page()
948 struct nullb_page *t_page; in copy_to_nullb() local
958 t_page = null_insert_page(nullb, sector, in copy_to_nullb()
960 if (!t_page) in copy_to_nullb()
964 dst = kmap_atomic(t_page->page); in copy_to_nullb()
969 __set_bit(sector & SECTOR_MASK, t_page->bitmap); in copy_to_nullb()
985 struct nullb_page *t_page; in copy_from_nullb() local
992 t_page = null_lookup_page(nullb, sector, false, in copy_from_nullb()
996 if (!t_page) { in copy_from_nullb()
1000 src = kmap_atomic(t_page->page); in copy_from_nullb()