Lines Matching full:range

19  * @MMU_NOTIFY_UNMAP: either munmap() that unmap the range or a mremap() that
20 * move the range
25 * @MMU_NOTIFY_PROTECTION_VMA: update is due to protection change for the range
26 * ie using the vma access permission (vm_page_prot) to update the whole range
31 * pages in the range so to mirror those changes the user must inspect the CPU
40 * that the mm refcount is zero and the range is no longer accessible.
48 * exclusive range the owner will be initialised to the value provided by the
139 * the pages in the range, it has to implement the
145 * establishment of sptes is forbidden in the range passed to
150 * range are still mapped and have at least a refcount of one.
153 * range have been unmapped and the pages have been freed by
185 const struct mmu_notifier_range *range);
187 const struct mmu_notifier_range *range);
200 * external TLB range needs to be flushed. For more in depth
203 * Note that this function might be called with just a sub-range
248 * range. This function can sleep. Return false only if sleeping
249 * was required but mmu_notifier_range_blockable(range) is false.
253 const struct mmu_notifier_range *range,
341 * mmu_interval_read_retry - End a read side critical section against a VA range
404 mmu_notifier_range_update_to_read_only(const struct mmu_notifier_range *range);
407 mmu_notifier_range_blockable(const struct mmu_notifier_range *range) in mmu_notifier_range_blockable() argument
409 return (range->flags & MMU_NOTIFIER_RANGE_BLOCKABLE); in mmu_notifier_range_blockable()
452 mmu_notifier_invalidate_range_start(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_start() argument
457 if (mm_has_notifiers(range->mm)) { in mmu_notifier_invalidate_range_start()
458 range->flags |= MMU_NOTIFIER_RANGE_BLOCKABLE; in mmu_notifier_invalidate_range_start()
459 __mmu_notifier_invalidate_range_start(range); in mmu_notifier_invalidate_range_start()
465 mmu_notifier_invalidate_range_start_nonblock(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_start_nonblock() argument
470 if (mm_has_notifiers(range->mm)) { in mmu_notifier_invalidate_range_start_nonblock()
471 range->flags &= ~MMU_NOTIFIER_RANGE_BLOCKABLE; in mmu_notifier_invalidate_range_start_nonblock()
472 ret = __mmu_notifier_invalidate_range_start(range); in mmu_notifier_invalidate_range_start_nonblock()
479 mmu_notifier_invalidate_range_end(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_end() argument
481 if (mmu_notifier_range_blockable(range)) in mmu_notifier_invalidate_range_end()
484 if (mm_has_notifiers(range->mm)) in mmu_notifier_invalidate_range_end()
485 __mmu_notifier_invalidate_range_end(range, false); in mmu_notifier_invalidate_range_end()
489 mmu_notifier_invalidate_range_only_end(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_only_end() argument
491 if (mm_has_notifiers(range->mm)) in mmu_notifier_invalidate_range_only_end()
492 __mmu_notifier_invalidate_range_end(range, true); in mmu_notifier_invalidate_range_only_end()
514 static inline void mmu_notifier_range_init(struct mmu_notifier_range *range, in mmu_notifier_range_init() argument
522 range->vma = vma; in mmu_notifier_range_init()
523 range->event = event; in mmu_notifier_range_init()
524 range->mm = mm; in mmu_notifier_range_init()
525 range->start = start; in mmu_notifier_range_init()
526 range->end = end; in mmu_notifier_range_init()
527 range->flags = flags; in mmu_notifier_range_init()
531 struct mmu_notifier_range *range, in mmu_notifier_range_init_owner() argument
536 mmu_notifier_range_init(range, event, flags, vma, mm, start, end); in mmu_notifier_range_init_owner()
537 range->owner = owner; in mmu_notifier_range_init_owner()
654 static inline void _mmu_notifier_range_init(struct mmu_notifier_range *range, in _mmu_notifier_range_init() argument
658 range->start = start; in _mmu_notifier_range_init()
659 range->end = end; in _mmu_notifier_range_init()
662 #define mmu_notifier_range_init(range,event,flags,vma,mm,start,end) \ argument
663 _mmu_notifier_range_init(range, start, end)
664 #define mmu_notifier_range_init_owner(range, event, flags, vma, mm, start, \ argument
666 _mmu_notifier_range_init(range, start, end)
669 mmu_notifier_range_blockable(const struct mmu_notifier_range *range) in mmu_notifier_range_blockable() argument
702 mmu_notifier_invalidate_range_start(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_start() argument
707 mmu_notifier_invalidate_range_start_nonblock(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_start_nonblock() argument
713 void mmu_notifier_invalidate_range_end(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_end() argument
718 mmu_notifier_invalidate_range_only_end(struct mmu_notifier_range *range) in mmu_notifier_invalidate_range_only_end() argument