Lines Matching refs:mmu

255 	struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu;  in local_flush_tlb_all()  local
258 int num_tlb = mmu->sets * mmu->ways; in local_flush_tlb_all()
731 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in read_decode_mmu_bcr() local
763 mmu->ver = (tmp >> 24); in read_decode_mmu_bcr()
766 if (mmu->ver <= 2) { in read_decode_mmu_bcr()
768 mmu->pg_sz_k = TO_KB(0x2000); in read_decode_mmu_bcr()
769 mmu->sets = 1 << mmu2->sets; in read_decode_mmu_bcr()
770 mmu->ways = 1 << mmu2->ways; in read_decode_mmu_bcr()
771 mmu->u_dtlb = mmu2->u_dtlb; in read_decode_mmu_bcr()
772 mmu->u_itlb = mmu2->u_itlb; in read_decode_mmu_bcr()
775 mmu->pg_sz_k = 1 << (mmu3->pg_sz - 1); in read_decode_mmu_bcr()
776 mmu->sets = 1 << mmu3->sets; in read_decode_mmu_bcr()
777 mmu->ways = 1 << mmu3->ways; in read_decode_mmu_bcr()
778 mmu->u_dtlb = mmu3->u_dtlb; in read_decode_mmu_bcr()
779 mmu->u_itlb = mmu3->u_itlb; in read_decode_mmu_bcr()
780 mmu->sasid = mmu3->sasid; in read_decode_mmu_bcr()
784 mmu->pg_sz_k = 1 << (mmu4->sz0 - 1); in read_decode_mmu_bcr()
785 mmu->s_pg_sz_m = 1 << (mmu4->sz1 - 11); in read_decode_mmu_bcr()
786 mmu->sets = 64 << mmu4->n_entry; in read_decode_mmu_bcr()
787 mmu->ways = mmu4->n_ways * 2; in read_decode_mmu_bcr()
788 mmu->u_dtlb = mmu4->u_dtlb * 4; in read_decode_mmu_bcr()
789 mmu->u_itlb = mmu4->u_itlb * 4; in read_decode_mmu_bcr()
790 mmu->sasid = mmu4->sasid; in read_decode_mmu_bcr()
791 pae_exists = mmu->pae = mmu4->pae; in read_decode_mmu_bcr()
798 struct cpuinfo_arc_mmu *p_mmu = &cpuinfo_arc700[cpu_id].mmu; in arc_mmu_mumbojumbo()
823 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in arc_mmu_init() local
849 if (is_isa_arcompact() && mmu->ver == CONFIG_ARC_MMU_VER) in arc_mmu_init()
851 else if (is_isa_arcv2() && mmu->ver >= CONFIG_ARC_MMU_VER) in arc_mmu_init()
856 mmu->ver, CONFIG_ARC_MMU_VER); in arc_mmu_init()
859 if (mmu->pg_sz_k != TO_KB(PAGE_SIZE)) in arc_mmu_init()
863 mmu->s_pg_sz_m != TO_MB(HPAGE_PMD_SIZE)) in arc_mmu_init()
867 if (IS_ENABLED(CONFIG_ARC_HAS_PAE40) && !mmu->pae) in arc_mmu_init()
898 #define SET_WAY_TO_IDX(mmu, set, way) ((set) * mmu->ways + (way)) argument
913 struct cpuinfo_arc_mmu *mmu = &cpuinfo_arc700[smp_processor_id()].mmu; in do_tlb_overlap_fault() local
914 unsigned int pd0[mmu->ways]; in do_tlb_overlap_fault()
921 for (set = 0; set < mmu->sets; set++) { in do_tlb_overlap_fault()
926 for (way = 0, is_valid = 0; way < mmu->ways; way++) { in do_tlb_overlap_fault()
928 SET_WAY_TO_IDX(mmu, set, way)); in do_tlb_overlap_fault()
940 for (way = 0; way < mmu->ways - 1; way++) { in do_tlb_overlap_fault()
947 for (n = way + 1; n < mmu->ways; n++) { in do_tlb_overlap_fault()
961 SET_WAY_TO_IDX(mmu, set, way)); in do_tlb_overlap_fault()