Lines Matching refs:last

66 svm_range_check_vm(struct kfd_process *p, uint64_t start, uint64_t last,
84 prange, prange->start, prange->last); in svm_range_unlink()
93 if (prange->it_node.start != 0 && prange->it_node.last != 0) in svm_range_unlink()
101 prange, prange->start, prange->last); in svm_range_add_notifier_locked()
120 prange, prange->start, prange->last); in svm_range_add_to_svms()
124 prange->it_node.last = prange->last; in svm_range_add_to_svms()
133 prange->notifier.interval_tree.last >> PAGE_SHIFT); in svm_range_remove_notifier()
136 prange->notifier.interval_tree.last != 0) in svm_range_remove_notifier()
271 uint64_t size = (prange->last - prange->start + 1) << PAGE_SHIFT; in svm_range_free()
275 prange->start, prange->last); in svm_range_free()
303 uint64_t last, bool update_mem_usage) in svm_range_new() argument
305 uint64_t size = last - start + 1; in svm_range_new()
324 prange->last = last; in svm_range_new()
343 pr_debug("svms 0x%p [0x%llx 0x%llx]\n", svms, start, last); in svm_range_new()
375 prange->start, prange->last); in svm_range_bo_release()
478 prange->svms, prange->start, prange->last); in svm_range_validate_svm_bo()
528 prange->start, prange->last); in svm_range_vram_node_new()
890 uint64_t start, uint64_t last) in svm_range_split_pages() argument
892 uint64_t npages = last - start + 1; in svm_range_split_pages()
908 uint64_t start, uint64_t last) in svm_range_split_nodes() argument
910 uint64_t npages = last - start + 1; in svm_range_split_nodes()
913 new->svms, new, new->start, start, last); in svm_range_split_nodes()
949 uint64_t start, uint64_t last) in svm_range_split_adjust() argument
954 new->svms, new->start, old->start, old->last, start, last); in svm_range_split_adjust()
957 new->last > old->last) { in svm_range_split_adjust()
962 r = svm_range_split_pages(new, old, start, last); in svm_range_split_adjust()
967 r = svm_range_split_nodes(new, old, start, last); in svm_range_split_adjust()
972 old->npages = last - start + 1; in svm_range_split_adjust()
974 old->last = last; in svm_range_split_adjust()
1008 svm_range_split(struct svm_range *prange, uint64_t start, uint64_t last, in svm_range_split() argument
1012 uint64_t old_last = prange->last; in svm_range_split()
1017 old_start, old_last, start, last); in svm_range_split()
1019 if (old_start != start && old_last != last) in svm_range_split()
1021 if (start < old_start || last > old_last) in svm_range_split()
1026 *new = svm_range_new(svms, last + 1, old_last, false); in svm_range_split()
1032 r = svm_range_split_adjust(*new, prange, start, last); in svm_range_split()
1035 r, old_start, old_last, start, last); in svm_range_split()
1060 int r = svm_range_split(prange, new_start, prange->last, &head); in svm_range_split_head()
1072 pchild, pchild->start, pchild->last, prange, op); in svm_range_add_child()
1102 unsigned long start, last, size; in svm_range_split_by_granularity() local
1111 last = ALIGN(addr + 1, size) - 1; in svm_range_split_by_granularity()
1114 prange->svms, prange->start, prange->last, start, last, size); in svm_range_split_by_granularity()
1117 r = svm_range_split(prange, start, prange->last, &head); in svm_range_split_by_granularity()
1123 if (last < prange->last) { in svm_range_split_by_granularity()
1124 r = svm_range_split(prange, prange->start, last, &tail); in svm_range_split_by_granularity()
1134 prange, prange->start, prange->last, in svm_range_split_by_granularity()
1211 uint64_t start, uint64_t last, in svm_range_unmap_from_gpu() argument
1216 pr_debug("[0x%llx 0x%llx]\n", start, last); in svm_range_unmap_from_gpu()
1219 last, init_pte_value, 0, 0, NULL, NULL, in svm_range_unmap_from_gpu()
1225 unsigned long last, uint32_t trigger) in svm_range_unmap_from_gpus() argument
1236 prange, prange->start, prange->last); in svm_range_unmap_from_gpus()
1240 if (prange->start == start && prange->last == last) { in svm_range_unmap_from_gpus()
1258 start, last, trigger); in svm_range_unmap_from_gpus()
1262 start, last, &fence); in svm_range_unmap_from_gpus()
1580 end = (prange->last + 1) << PAGE_SHIFT; in svm_range_validate_and_map()
1716 prange->svms, prange, prange->start, prange->last, in svm_range_restore_work()
1787 unsigned long start, unsigned long last, in svm_range_evict() argument
1798 svms, prange->start, prange->last, start, last); in svm_range_evict()
1810 if (pchild->start <= last && pchild->last >= start) { in svm_range_evict()
1812 pchild->start, pchild->last); in svm_range_evict()
1821 if (prange->start <= last && prange->last >= start) in svm_range_evict()
1829 prange->svms, prange->start, prange->last); in svm_range_evict()
1849 prange->svms, start, last); in svm_range_evict()
1853 l = min(last, pchild->last); in svm_range_evict()
1859 l = min(last, prange->last); in svm_range_evict()
1871 new = svm_range_new(old->svms, old->start, old->last, false); in svm_range_clone()
1911 svm_range_split_new(struct svm_range_list *svms, uint64_t start, uint64_t last, in svm_range_split_new() argument
1919 max_pages, start, last); in svm_range_split_new()
1921 while (last >= start) { in svm_range_split_new()
1922 l = min(last, ALIGN_DOWN(start + max_pages, max_pages) - 1); in svm_range_split_new()
1970 unsigned long last = start + size - 1UL; in svm_range_add() local
1978 pr_debug("svms 0x%p [0x%llx 0x%lx]\n", &p->svms, start, last); in svm_range_add()
1985 node = interval_tree_iter_first(&svms->objects, start, last); in svm_range_add()
1991 node->last); in svm_range_add()
1994 next = interval_tree_iter_next(node, start, last); in svm_range_add()
1995 next_start = min(node->last, last) + 1; in svm_range_add()
1999 } else if (node->start < start || node->last > last) { in svm_range_add()
2023 if (node->last > last) { in svm_range_add()
2025 r = svm_range_split_tail(prange, last, in svm_range_add()
2051 if (start <= last) in svm_range_add()
2052 r = svm_range_split_new(svms, start, last, in svm_range_add()
2074 unsigned long last; in svm_range_update_notifier_and_interval_tree() local
2077 last = prange->notifier.interval_tree.last >> PAGE_SHIFT; in svm_range_update_notifier_and_interval_tree()
2079 if (prange->start == start && prange->last == last) in svm_range_update_notifier_and_interval_tree()
2083 prange->svms, prange, start, last, prange->start, in svm_range_update_notifier_and_interval_tree()
2084 prange->last); in svm_range_update_notifier_and_interval_tree()
2086 if (start != 0 && last != 0) { in svm_range_update_notifier_and_interval_tree()
2091 prange->it_node.last = prange->last; in svm_range_update_notifier_and_interval_tree()
2104 svms, prange, prange->start, prange->last); in svm_range_handle_list_op()
2108 svms, prange, prange->start, prange->last); in svm_range_handle_list_op()
2115 svms, prange, prange->start, prange->last); in svm_range_handle_list_op()
2120 svms, prange, prange->start, prange->last); in svm_range_handle_list_op()
2126 prange->start, prange->last); in svm_range_handle_list_op()
2132 prange, prange->start, prange->last); in svm_range_handle_list_op()
2188 prange->start, prange->last, prange->work_item.op); in svm_range_deferred_list_work()
2263 prange, prange->start, prange->last, op); in svm_range_add_list_work()
2279 unsigned long last) in svm_range_unmap_split() argument
2286 prange->start, prange->last); in svm_range_unmap_split()
2289 if (start > prange->last || last < prange->start) in svm_range_unmap_split()
2295 if (last < tail->last) in svm_range_unmap_split()
2296 svm_range_split(tail, last + 1, tail->last, &head); in svm_range_unmap_split()
2312 unsigned long start, unsigned long last) in svm_range_unmap_from_cpu() argument
2327 prange, prange->start, prange->last, start, last); in svm_range_unmap_from_cpu()
2335 unmap_parent = start <= prange->start && last >= prange->last; in svm_range_unmap_from_cpu()
2340 l = min(last, pchild->last); in svm_range_unmap_from_cpu()
2343 svm_range_unmap_split(mm, prange, pchild, start, last); in svm_range_unmap_from_cpu()
2347 l = min(last, prange->last); in svm_range_unmap_from_cpu()
2350 svm_range_unmap_split(mm, prange, prange, start, last); in svm_range_unmap_from_cpu()
2389 unsigned long last; in svm_range_cpu_invalidate_pagetables() local
2397 last = mni->interval_tree.last; in svm_range_cpu_invalidate_pagetables()
2399 last = min(last, range->end - 1) >> PAGE_SHIFT; in svm_range_cpu_invalidate_pagetables()
2401 start, last, range->start >> PAGE_SHIFT, in svm_range_cpu_invalidate_pagetables()
2404 mni->interval_tree.last >> PAGE_SHIFT, range->event); in svm_range_cpu_invalidate_pagetables()
2413 svm_range_unmap_from_cpu(mni->mm, prange, start, last); in svm_range_cpu_invalidate_pagetables()
2416 svm_range_evict(prange, mni->mm, start, last, range->event); in svm_range_cpu_invalidate_pagetables()
2450 addr, prange->start, prange->last, node->start, node->last); in svm_range_from_addr()
2452 if (addr >= prange->start && addr <= prange->last) { in svm_range_from_addr()
2458 if (addr >= pchild->start && addr <= pchild->last) { in svm_range_from_addr()
2460 addr, pchild->start, pchild->last); in svm_range_from_addr()
2538 unsigned long *start, unsigned long *last, in svm_range_get_range_boundaries() argument
2575 if (node->last >= addr) { in svm_range_get_range_boundaries()
2579 start_limit = max(start_limit, node->last + 1); in svm_range_get_range_boundaries()
2583 *last = end_limit - 1; in svm_range_get_range_boundaries()
2587 *start, *last, *is_heap_stack); in svm_range_get_range_boundaries()
2593 svm_range_check_vm_userptr(struct kfd_process *p, uint64_t start, uint64_t last, in svm_range_check_vm_userptr() argument
2623 last << PAGE_SHIFT, in svm_range_check_vm_userptr()
2630 start, last); in svm_range_check_vm_userptr()
2650 unsigned long start, last; in svm_range_create_unregistered_range() local
2657 if (svm_range_get_range_boundaries(p, addr, &start, &last, in svm_range_create_unregistered_range()
2661 r = svm_range_check_vm(p, start, last, &bo_s, &bo_l); in svm_range_create_unregistered_range()
2663 r = svm_range_check_vm_userptr(p, start, last, &bo_s, &bo_l); in svm_range_create_unregistered_range()
2671 last = addr; in svm_range_create_unregistered_range()
2674 prange = svm_range_new(&p->svms, start, last, true); in svm_range_create_unregistered_range()
2720 svms, prange, prange->start, prange->last); in svm_range_skip_recover()
2726 svms, prange, prange->start, prange->last); in svm_range_skip_recover()
2866 svms, prange->start, prange->last); in svm_range_restore_pages()
2891 svms, prange->start, prange->last); in svm_range_restore_pages()
2897 svms, prange->start, prange->last, best_loc, in svm_range_restore_pages()
2928 r, svms, prange->start, prange->last); in svm_range_restore_pages()
2936 r, svms, prange->start, prange->last); in svm_range_restore_pages()
2976 size = (pchild->last - pchild->start + 1) << PAGE_SHIFT; in svm_range_switch_xnack_reserve_mem()
2989 size = (prange->last - prange->start + 1) << PAGE_SHIFT; in svm_range_switch_xnack_reserve_mem()
3093 svm_range_check_vm(struct kfd_process *p, uint64_t start, uint64_t last, in svm_range_check_vm() argument
3112 node = interval_tree_iter_first(&vm->va, start, last); in svm_range_check_vm()
3115 start, last); in svm_range_check_vm()
3120 *bo_l = mapping->last; in svm_range_check_vm()
3238 p->xnack_enabled, &p->svms, prange->start, prange->last, in svm_range_best_prefetch_location()
3340 prange->start, prange->last); in svm_range_evict_svm_bo_worker()
3431 prange->last); in svm_range_set_attr()
3503 uint64_t last = start + size - 1UL; in svm_range_get_attr() local
3566 node = interval_tree_iter_first(&svms->objects, start, last); in svm_range_get_attr()
3587 next = interval_tree_iter_next(node, start, last); in svm_range_get_attr()