Home
last modified time | relevance | path

Searched refs:need_flush (Results 1 – 17 of 17) sorted by relevance

/Linux-v5.10/arch/x86/mm/
Dtlb.c206 u16 *new_asid, bool *need_flush) in choose_new_asid() argument
212 *need_flush = true; in choose_new_asid()
225 *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < in choose_new_asid()
239 *need_flush = true; in choose_new_asid()
268 static void load_new_mm_cr3(pgd_t *pgdir, u16 new_asid, bool need_flush) in load_new_mm_cr3() argument
272 if (need_flush) { in load_new_mm_cr3()
430 bool need_flush; in switch_mm_irqs_off() local
524 need_flush = true; in switch_mm_irqs_off()
551 choose_new_asid(next, next_tlb_gen, &new_asid, &need_flush); in switch_mm_irqs_off()
558 if (need_flush) { in switch_mm_irqs_off()
/Linux-v5.10/arch/arm/mm/
Dpmsa-v7.c366 unsigned int subregions, bool need_flush) in mpu_setup_region() argument
385 if (need_flush) in mpu_setup_region()
439 bool need_flush = region == PMSAv7_RAM_REGION; in pmsav7_setup() local
446 xip[i].subreg, need_flush); in pmsav7_setup()
/Linux-v5.10/arch/x86/kernel/
Damd_gart_64.c89 static bool need_flush; /* global flush state. set for each gart wrap */ variable
106 need_flush = true; in alloc_iommu()
115 need_flush = true; in alloc_iommu()
119 need_flush = true; in alloc_iommu()
144 if (need_flush) { in flush_gart()
146 need_flush = false; in flush_gart()
/Linux-v5.10/mm/
Dhighmem.c166 int need_flush = 0; in flush_all_zero_pkmaps() local
197 need_flush = 1; in flush_all_zero_pkmaps()
199 if (need_flush) in flush_all_zero_pkmaps()
/Linux-v5.10/drivers/gpu/drm/etnaviv/
Detnaviv_buffer.c350 bool need_flush = switch_mmu_context || gpu->flush_seq != new_flush_seq; in etnaviv_buffer_queue() local
368 if (need_flush || switch_context) { in etnaviv_buffer_queue()
375 if (need_flush) { in etnaviv_buffer_queue()
405 if (need_flush) { in etnaviv_buffer_queue()
/Linux-v5.10/arch/sparc/kernel/
Diommu-common.c19 static inline bool need_flush(struct iommu_map_table *iommu) in need_flush() function
206 (n < pool->hint || need_flush(iommu))) { in iommu_tbl_range_alloc()
/Linux-v5.10/drivers/infiniband/hw/mlx4/
Dcm.c528 int need_flush = 0; in mlx4_ib_cm_paravirt_clean() local
536 need_flush |= !cancel_delayed_work(&map->timeout); in mlx4_ib_cm_paravirt_clean()
542 if (need_flush) in mlx4_ib_cm_paravirt_clean()
/Linux-v5.10/arch/arm64/kvm/hyp/
Dpgtable.c663 bool need_flush = false; in stage2_unmap_walker() local
674 need_flush = true; in stage2_unmap_walker()
686 if (need_flush) { in stage2_unmap_walker()
/Linux-v5.10/arch/x86/kvm/mmu/
Dtdp_mmu.c746 int need_flush = 0; in set_tdp_spte() local
770 need_flush = 1; in set_tdp_spte()
773 if (need_flush) in set_tdp_spte()
Dmmu.c1348 int need_flush = 0; in kvm_set_pte_rmapp() local
1361 need_flush = 1; in kvm_set_pte_rmapp()
1375 if (need_flush && kvm_available_flush_tlb_with_range()) { in kvm_set_pte_rmapp()
1380 return need_flush; in kvm_set_pte_rmapp()
/Linux-v5.10/fs/xfs/
Dxfs_log.c1678 bool need_flush) in xlog_write_iclog() argument
1718 if (need_flush) in xlog_write_iclog()
1836 bool need_flush = true, split = false; in xlog_sync() local
1896 need_flush = false; in xlog_sync()
1900 xlog_write_iclog(log, iclog, bno, count, need_flush); in xlog_sync()
/Linux-v5.10/fs/ceph/
Dsnap.c518 capsnap->need_flush = true; in ceph_queue_cap_snap()
530 capsnap->need_flush ? "" : "no_flush"); in ceph_queue_cap_snap()
Dsuper.h221 bool need_flush; member
Dcaps.c1431 if (capsnap->need_flush) { in __prep_cap()
1553 BUG_ON(!capsnap->need_flush); in __ceph_flush_snaps()
3006 if (!capsnap->need_flush && in ceph_try_drop_cap_snap()
/Linux-v5.10/drivers/md/
Ddm-writecache.c966 bool need_flush = false; in writecache_resume() local
1032 need_flush = true; in writecache_resume()
1053 need_flush = true; in writecache_resume()
1060 if (need_flush) { in writecache_resume()
Draid10.c4415 int need_flush = 0; in reshape_request() local
4456 need_flush = 1; in reshape_request()
4478 need_flush = 1; in reshape_request()
4488 if (need_flush || in reshape_request()
/Linux-v5.10/drivers/gpu/drm/i915/gem/selftests/
Di915_gem_context.c461 unsigned int n, m, need_flush; in cpu_fill() local
465 err = i915_gem_object_prepare_write(obj, &need_flush); in cpu_fill()