Lines Matching refs:va

188 	uintptr_t va = POINTER_TO_UINT(sys_cache_cached_ptr_get(virt));  in sys_mm_drv_map_page()  local
191 CHECKIF(!sys_mm_drv_is_addr_aligned(va)) { in sys_mm_drv_map_page()
197 CHECKIF((va < UNUSED_L2_START_ALIGNED) || in sys_mm_drv_map_page()
198 (va >= (CONFIG_KERNEL_VM_BASE + CONFIG_KERNEL_VM_SIZE))) { in sys_mm_drv_map_page()
235 entry_idx = get_tlb_entry_idx(va); in sys_mm_drv_map_page()
271 arch_mem_map(virt, va, CONFIG_MM_DRV_PAGE_SIZE, flags); in sys_mm_drv_map_page()
292 uint8_t *va; in sys_mm_drv_map_region() local
301 va = (__sparse_force uint8_t *)sys_cache_cached_ptr_get(virt); in sys_mm_drv_map_region()
307 int ret2 = sys_mm_drv_map_page(va, pa, flags); in sys_mm_drv_map_region()
310 __ASSERT(false, "cannot map 0x%lx to %p\n", pa, va); in sys_mm_drv_map_region()
314 va += CONFIG_MM_DRV_PAGE_SIZE; in sys_mm_drv_map_region()
329 void *va = (__sparse_force void *)sys_cache_cached_ptr_get(virt); in sys_mm_drv_map_array() local
331 return sys_mm_drv_simple_map_array(va, phys, cnt, flags); in sys_mm_drv_map_array()
344 uintptr_t va = POINTER_TO_UINT(sys_cache_cached_ptr_get(virt)); in sys_mm_drv_unmap_page_wflush() local
347 CHECKIF((va < UNUSED_L2_START_ALIGNED) || in sys_mm_drv_unmap_page_wflush()
348 (va >= (CONFIG_KERNEL_VM_BASE + CONFIG_KERNEL_VM_SIZE))) { in sys_mm_drv_unmap_page_wflush()
354 CHECKIF(!sys_mm_drv_is_addr_aligned(va)) { in sys_mm_drv_unmap_page_wflush()
361 entry_idx = get_tlb_entry_idx(va); in sys_mm_drv_unmap_page_wflush()
421 void *va = (__sparse_force void *)sys_cache_cached_ptr_get(virt); in sys_mm_drv_unmap_region() local
423 return sys_mm_drv_simple_unmap_region(va, size); in sys_mm_drv_unmap_region()
435 uintptr_t va = POINTER_TO_UINT(sys_cache_cached_ptr_get(virt)); in sys_mm_drv_update_page_flags() local
438 CHECKIF(!sys_mm_drv_is_addr_aligned(va) || in sys_mm_drv_update_page_flags()
439 (va < UNUSED_L2_START_ALIGNED) || in sys_mm_drv_update_page_flags()
440 (va >= (CONFIG_KERNEL_VM_BASE + CONFIG_KERNEL_VM_SIZE))) { in sys_mm_drv_update_page_flags()
446 entry_idx = get_tlb_entry_idx(va); in sys_mm_drv_update_page_flags()
491 uint8_t *va = (uint8_t *)virt + offset; in sys_mm_drv_unmap_region_initial() local
493 int ret2 = sys_mm_drv_unmap_page_wflush(va, false); in sys_mm_drv_unmap_region_initial()
500 __ASSERT(false, "cannot unmap %p\n", va); in sys_mm_drv_unmap_region_initial()
520 uintptr_t va = POINTER_TO_UINT(sys_cache_cached_ptr_get(virt)); in sys_mm_drv_page_phys_get() local
522 CHECKIF(!sys_mm_drv_is_addr_aligned(va)) { in sys_mm_drv_page_phys_get()
528 CHECKIF((va < CONFIG_KERNEL_VM_BASE) || in sys_mm_drv_page_phys_get()
529 (va >= (CONFIG_KERNEL_VM_BASE + CONFIG_KERNEL_VM_SIZE))) { in sys_mm_drv_page_phys_get()
534 ent = tlb_entries[get_tlb_entry_idx(va)]; in sys_mm_drv_page_phys_get()
561 uintptr_t va = POINTER_TO_UINT(sys_cache_cached_ptr_get(virt)); in sys_mm_drv_page_flag_get() local
563 CHECKIF(!sys_mm_drv_is_addr_aligned(va)) { in sys_mm_drv_page_flag_get()
569 CHECKIF((va < CONFIG_KERNEL_VM_BASE) || in sys_mm_drv_page_flag_get()
570 (va >= (CONFIG_KERNEL_VM_BASE + CONFIG_KERNEL_VM_SIZE))) { in sys_mm_drv_page_flag_get()
575 ent = tlb_entries[get_tlb_entry_idx(va)]; in sys_mm_drv_page_flag_get()