Lines Matching refs:tb
24 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in flush_tlb_pending() local
25 struct mm_struct *mm = tb->mm; in flush_tlb_pending()
27 if (!tb->tlb_nr) in flush_tlb_pending()
30 flush_tsb_user(tb); in flush_tlb_pending()
33 if (tb->tlb_nr == 1) { in flush_tlb_pending()
34 global_flush_tlb_page(mm, tb->vaddrs[0]); in flush_tlb_pending()
37 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, in flush_tlb_pending()
38 &tb->vaddrs[0]); in flush_tlb_pending()
40 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), in flush_tlb_pending()
41 tb->tlb_nr, &tb->vaddrs[0]); in flush_tlb_pending()
46 tb->tlb_nr = 0; in flush_tlb_pending()
54 struct tlb_batch *tb = this_cpu_ptr(&tlb_batch); in arch_enter_lazy_mmu_mode() local
56 tb->active = 1; in arch_enter_lazy_mmu_mode()
61 struct tlb_batch *tb = this_cpu_ptr(&tlb_batch); in arch_leave_lazy_mmu_mode() local
63 if (tb->tlb_nr) in arch_leave_lazy_mmu_mode()
65 tb->active = 0; in arch_leave_lazy_mmu_mode()
71 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in tlb_batch_add_one() local
78 nr = tb->tlb_nr; in tlb_batch_add_one()
80 if (unlikely(nr != 0 && mm != tb->mm)) { in tlb_batch_add_one()
85 if (!tb->active) { in tlb_batch_add_one()
92 tb->mm = mm; in tlb_batch_add_one()
93 tb->hugepage_shift = hugepage_shift; in tlb_batch_add_one()
96 if (tb->hugepage_shift != hugepage_shift) { in tlb_batch_add_one()
98 tb->hugepage_shift = hugepage_shift; in tlb_batch_add_one()
102 tb->vaddrs[nr] = vaddr; in tlb_batch_add_one()
103 tb->tlb_nr = ++nr; in tlb_batch_add_one()