Lines Matching refs:vaddr

178 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr,  in copy_user_page()  argument
181 cpu_dcache_wbinval_page((unsigned long)vaddr); in copy_user_page()
182 cpu_icache_inval_page((unsigned long)vaddr); in copy_user_page()
188 void clear_user_page(void *addr, unsigned long vaddr, struct page *page) in clear_user_page() argument
190 cpu_dcache_wbinval_page((unsigned long)vaddr); in clear_user_page()
191 cpu_icache_inval_page((unsigned long)vaddr); in clear_user_page()
198 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument
207 if (aliasing(vaddr, (unsigned long)kfrom)) in copy_user_highpage()
209 vto = kremap0(vaddr, pto); in copy_user_highpage()
210 vfrom = kremap1(vaddr, pfrom); in copy_user_highpage()
219 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
226 if (aliasing(kto, vaddr) && kto != 0) { in clear_user_highpage()
230 vto = kremap0(vaddr, page_to_phys(page)); in clear_user_highpage()
252 unsigned long vaddr, kto; in flush_dcache_page() local
254 vaddr = page->index << PAGE_SHIFT; in flush_dcache_page()
255 if (aliasing(vaddr, kaddr)) { in flush_dcache_page()
256 kto = kremap0(vaddr, page_to_phys(page)); in flush_dcache_page()
267 unsigned long vaddr, void *dst, void *src, int len) in copy_to_user_page() argument
272 vto = kremap0(vaddr, page_to_phys(page)); in copy_to_user_page()
273 dst = (void *)(vto | (vaddr & (PAGE_SIZE - 1))); in copy_to_user_page()
288 unsigned long vaddr, void *dst, void *src, int len) in copy_from_user_page() argument
293 vto = kremap0(vaddr, page_to_phys(page)); in copy_from_user_page()
294 src = (void *)(vto | (vaddr & (PAGE_SIZE - 1))); in copy_from_user_page()
301 struct page *page, unsigned long vaddr) in flush_anon_page() argument
312 cpu_icache_inval_page(vaddr & PAGE_MASK); in flush_anon_page()
314 if (aliasing(vaddr, kaddr)) { in flush_anon_page()
315 ktmp = kremap0(vaddr, page_to_phys(page)); in flush_anon_page()