Lines Matching refs:page

37 static inline void *kmap(struct page *page);
46 static inline void kunmap(struct page *page);
54 static inline struct page *kmap_to_page(void *addr);
96 static inline void *kmap_local_page(struct page *page);
180 static inline void *kmap_atomic(struct page *page);
187 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vma… in flush_anon_page() argument
203 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
205 void *addr = kmap_local_page(page); in clear_user_highpage()
206 clear_user_page(addr, vaddr, page); in clear_user_highpage()
226 static inline struct page *
230 struct page *page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, vaddr); in alloc_zeroed_user_highpage_movable() local
232 if (page) in alloc_zeroed_user_highpage_movable()
233 clear_user_highpage(page, vaddr); in alloc_zeroed_user_highpage_movable()
235 return page; in alloc_zeroed_user_highpage_movable()
239 static inline void clear_highpage(struct page *page) in clear_highpage() argument
241 void *kaddr = kmap_local_page(page); in clear_highpage()
246 static inline void clear_highpage_kasan_tagged(struct page *page) in clear_highpage_kasan_tagged() argument
250 tag = page_kasan_tag(page); in clear_highpage_kasan_tagged()
251 page_kasan_tag_reset(page); in clear_highpage_kasan_tagged()
252 clear_highpage(page); in clear_highpage_kasan_tagged()
253 page_kasan_tag_set(page, tag); in clear_highpage_kasan_tagged()
258 static inline void tag_clear_highpage(struct page *page) in tag_clear_highpage() argument
269 void zero_user_segments(struct page *page, unsigned start1, unsigned end1,
272 static inline void zero_user_segments(struct page *page, in zero_user_segments() argument
276 void *kaddr = kmap_local_page(page); in zero_user_segments()
279 BUG_ON(end1 > page_size(page) || end2 > page_size(page)); in zero_user_segments()
288 for (i = 0; i < compound_nr(page); i++) in zero_user_segments()
289 flush_dcache_page(page + i); in zero_user_segments()
293 static inline void zero_user_segment(struct page *page, in zero_user_segment() argument
296 zero_user_segments(page, start, end, 0, 0); in zero_user_segment()
299 static inline void zero_user(struct page *page, in zero_user() argument
302 zero_user_segments(page, start, start + size, 0, 0); in zero_user()
307 static inline void copy_user_highpage(struct page *to, struct page *from, in copy_user_highpage()
324 static inline void copy_highpage(struct page *to, struct page *from) in copy_highpage()
338 static inline void memcpy_page(struct page *dst_page, size_t dst_off, in memcpy_page()
339 struct page *src_page, size_t src_off, in memcpy_page()
351 static inline void memset_page(struct page *page, size_t offset, int val, in memset_page() argument
354 char *addr = kmap_local_page(page); in memset_page()
361 static inline void memcpy_from_page(char *to, struct page *page, in memcpy_from_page() argument
364 char *from = kmap_local_page(page); in memcpy_from_page()
371 static inline void memcpy_to_page(struct page *page, size_t offset, in memcpy_to_page() argument
374 char *to = kmap_local_page(page); in memcpy_to_page()
378 flush_dcache_page(page); in memcpy_to_page()
382 static inline void memzero_page(struct page *page, size_t offset, size_t len) in memzero_page() argument
384 char *addr = kmap_local_page(page); in memzero_page()
388 flush_dcache_page(page); in memzero_page()
403 zero_user_segments(&folio->page, start1, xend1, start2, xend2); in folio_zero_segments()
415 zero_user_segments(&folio->page, start, xend, 0, 0); in folio_zero_segment()
427 zero_user_segments(&folio->page, start, start + length, 0, 0); in folio_zero_range()