Lines Matching refs:pos

575 	uint8_t *pos;  in k_mem_map_phys_guard()  local
629 VIRT_FOREACH(dst, size, pos) { in k_mem_map_phys_guard()
630 arch_mem_map(pos, in k_mem_map_phys_guard()
635 LOG_DBG("memory mapping anon pages %p to %p unpaged", dst, pos-1); in k_mem_map_phys_guard()
641 VIRT_FOREACH(dst, size, pos) { in k_mem_map_phys_guard()
642 ret = map_anon_page(pos, flags); in k_mem_map_phys_guard()
681 uint8_t *pos; in k_mem_unmap_phys_guard() local
693 pos = (uint8_t *)addr - CONFIG_MMU_PAGE_SIZE; in k_mem_unmap_phys_guard()
694 k_mem_assert_virtual_region(pos, size + (CONFIG_MMU_PAGE_SIZE * 2)); in k_mem_unmap_phys_guard()
702 pos = addr; in k_mem_unmap_phys_guard()
703 ret = arch_page_phys_get(pos - CONFIG_MMU_PAGE_SIZE, NULL); in k_mem_unmap_phys_guard()
711 ret = arch_page_phys_get(pos + size, NULL); in k_mem_unmap_phys_guard()
721 VIRT_FOREACH(addr, size, pos) { in k_mem_unmap_phys_guard()
726 status = arch_page_location_get(pos, &location); in k_mem_unmap_phys_guard()
734 arch_mem_unmap(pos, CONFIG_MMU_PAGE_SIZE); in k_mem_unmap_phys_guard()
749 ret = arch_page_phys_get(pos, &phys); in k_mem_unmap_phys_guard()
753 ret = arch_page_phys_get(pos, &phys); in k_mem_unmap_phys_guard()
757 __func__, pos); in k_mem_unmap_phys_guard()
785 arch_mem_unmap(pos, CONFIG_MMU_PAGE_SIZE); in k_mem_unmap_phys_guard()
810 pos = (uint8_t *)addr - CONFIG_MMU_PAGE_SIZE; in k_mem_unmap_phys_guard()
812 virt_region_free(pos, total_size); in k_mem_unmap_phys_guard()
1444 void *pos = (uint8_t *)addr + offset; in k_mem_page_out() local
1447 ret = do_mem_evict(pos); in k_mem_page_out()