| /Linux-v6.1/drivers/gpu/drm/ |
| D | drm_mm.c | 336 u64 hole_start; in find_hole_addr() local 342 hole_start = __drm_mm_hole_node_start(node); in find_hole_addr() 344 if (addr < hole_start) in find_hole_addr() 346 else if (addr > hole_start + node->hole_size) in find_hole_addr() 453 u64 hole_start, hole_end; in drm_mm_reserve_node() local 466 adj_start = hole_start = __drm_mm_hole_node_start(hole); in drm_mm_reserve_node() 467 adj_end = hole_end = hole_start + hole->hole_size; in drm_mm_reserve_node() 483 if (node->start > hole_start) in drm_mm_reserve_node() 543 u64 hole_start = __drm_mm_hole_node_start(hole); in drm_mm_insert_node_in_range() local 544 u64 hole_end = hole_start + hole->hole_size; in drm_mm_insert_node_in_range() [all …]
|
| /Linux-v6.1/drivers/gpu/drm/i915/selftests/ |
| D | i915_gem_gtt.c | 241 u64 hole_start, u64 hole_end, in lowlevel_hole() argument 255 for (size = 12; (hole_end - hole_start) >> size; size++) { in lowlevel_hole() 262 hole_size = (hole_end - hole_start) >> aligned_size; in lowlevel_hole() 268 __func__, hole_start, hole_end, size, hole_size); in lowlevel_hole() 284 GEM_BUG_ON(hole_start + count * BIT_ULL(aligned_size) > hole_end); in lowlevel_hole() 307 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 315 hole_end = hole_start; /* quit */ in lowlevel_hole() 364 u64 addr = hole_start + order[n] * BIT_ULL(aligned_size); in lowlevel_hole() 403 u64 hole_start, u64 hole_end, in fill_hole() argument 406 const u64 hole_size = hole_end - hole_start; in fill_hole() [all …]
|
| /Linux-v6.1/arch/sparc/kernel/ |
| D | adi_64.c | 163 unsigned long end_addr, hole_start, hole_end; in alloc_tag_store() local 167 hole_start = 0; in alloc_tag_store() 198 (tag_desc->end > hole_start)) in alloc_tag_store() 199 hole_start = tag_desc->end; in alloc_tag_store() 273 if (tmp_addr < hole_start) { in alloc_tag_store() 277 tmp_addr = hole_start + 1; in alloc_tag_store()
|
| /Linux-v6.1/include/drm/ |
| D | drm_mm.h | 394 #define drm_mm_for_each_hole(pos, mm, hole_start, hole_end) \ argument 398 hole_start = drm_mm_hole_node_start(pos), \ 399 hole_end = hole_start + pos->hole_size, \
|
| /Linux-v6.1/kernel/ |
| D | kexec_core.c | 441 unsigned long hole_start, hole_end, size; in kimage_alloc_crash_control_pages() local 446 hole_start = (image->control_page + (size - 1)) & ~(size - 1); in kimage_alloc_crash_control_pages() 447 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 461 if ((hole_end >= mstart) && (hole_start <= mend)) { in kimage_alloc_crash_control_pages() 463 hole_start = (mend + (size - 1)) & ~(size - 1); in kimage_alloc_crash_control_pages() 464 hole_end = hole_start + size - 1; in kimage_alloc_crash_control_pages() 470 pages = pfn_to_page(hole_start >> PAGE_SHIFT); in kimage_alloc_crash_control_pages()
|
| /Linux-v6.1/fs/btrfs/ |
| D | zoned.h | 48 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, 156 u64 hole_start, u64 hole_end, in btrfs_find_allocatable_zones() argument 159 return hole_start; in btrfs_find_allocatable_zones()
|
| D | volumes.c | 1433 u64 *hole_start, u64 *hole_size, in dev_extent_hole_check_zoned() argument 1441 ASSERT(IS_ALIGNED(*hole_start, zone_size)); in dev_extent_hole_check_zoned() 1444 pos = btrfs_find_allocatable_zones(device, *hole_start, in dev_extent_hole_check_zoned() 1445 *hole_start + *hole_size, in dev_extent_hole_check_zoned() 1447 if (pos != *hole_start) { in dev_extent_hole_check_zoned() 1448 *hole_size = *hole_start + *hole_size - pos; in dev_extent_hole_check_zoned() 1449 *hole_start = pos; in dev_extent_hole_check_zoned() 1463 *hole_start += *hole_size; in dev_extent_hole_check_zoned() 1468 *hole_start += zone_size; in dev_extent_hole_check_zoned() 1486 static bool dev_extent_hole_check(struct btrfs_device *device, u64 *hole_start, in dev_extent_hole_check() argument [all …]
|
| D | zoned.c | 1032 u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start, in btrfs_find_allocatable_zones() argument 1038 u64 pos = hole_start; in btrfs_find_allocatable_zones() 1043 ASSERT(IS_ALIGNED(hole_start, zinfo->zone_size)); in btrfs_find_allocatable_zones()
|
| D | inode.c | 5053 u64 hole_start = ALIGN(oldsize, fs_info->sectorsize); in btrfs_cont_expand() local 5069 if (size <= hole_start) in btrfs_cont_expand() 5072 btrfs_lock_and_flush_ordered_range(inode, hole_start, block_end - 1, in btrfs_cont_expand() 5074 cur_offset = hole_start; in btrfs_cont_expand() 5135 unlock_extent(io_tree, hole_start, block_end - 1, &cached_state); in btrfs_cont_expand()
|
| /Linux-v6.1/fs/hugetlbfs/ |
| D | inode.c | 722 loff_t hole_start, hole_end; in hugetlbfs_punch_hole() local 727 hole_start = round_up(offset, hpage_size); in hugetlbfs_punch_hole() 741 if (offset < hole_start) in hugetlbfs_punch_hole() 743 offset, min(offset + len, hole_start)); in hugetlbfs_punch_hole() 746 if (hole_end > hole_start) { in hugetlbfs_punch_hole() 749 hole_start >> PAGE_SHIFT, in hugetlbfs_punch_hole() 754 if ((offset + len) > hole_end && (offset + len) > hole_start) in hugetlbfs_punch_hole() 761 if (hole_end > hole_start) in hugetlbfs_punch_hole() 762 remove_inode_hugepages(inode, hole_start, hole_end); in hugetlbfs_punch_hole()
|
| /Linux-v6.1/drivers/gpu/drm/tests/ |
| D | drm_mm_test.c | 49 u64 hole_start, __always_unused hole_end; in assert_no_holes() local 53 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) in assert_no_holes() 74 u64 hole_start, hole_end; in assert_one_hole() local 82 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in assert_one_hole() 83 if (start != hole_start || end != hole_end) { in assert_one_hole() 87 hole_start, hole_end, start, end); in assert_one_hole() 1155 u64 hole_start, hole_end; in show_holes() local 1158 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) { in show_holes() 1171 hole_start, hole_end, hole_end - hole_start, node2); in show_holes()
|
| /Linux-v6.1/drivers/gpu/drm/i915/gt/ |
| D | intel_ggtt.c | 550 unsigned long hole_start, hole_end; in init_ggtt() local 615 drm_mm_for_each_hole(entry, &ggtt->vm.mm, hole_start, hole_end) { in init_ggtt() 618 hole_start, hole_end); in init_ggtt() 619 ggtt->vm.clear_range(&ggtt->vm, hole_start, in init_ggtt() 620 hole_end - hole_start); in init_ggtt()
|
| /Linux-v6.1/fs/ext4/ |
| D | extents.c | 2279 ext4_ext_put_gap_in_cache(struct inode *inode, ext4_lblk_t hole_start, in ext4_ext_put_gap_in_cache() argument 2284 ext4_es_find_extent_range(inode, &ext4_es_is_delayed, hole_start, in ext4_ext_put_gap_in_cache() 2285 hole_start + hole_len - 1, &es); in ext4_ext_put_gap_in_cache() 2288 if (es.es_lblk <= hole_start) in ext4_ext_put_gap_in_cache() 2290 hole_len = min(es.es_lblk - hole_start, hole_len); in ext4_ext_put_gap_in_cache() 2292 ext_debug(inode, " -> %u:%u\n", hole_start, hole_len); in ext4_ext_put_gap_in_cache() 2293 ext4_es_insert_extent(inode, hole_start, hole_len, ~0, in ext4_ext_put_gap_in_cache() 4184 ext4_lblk_t hole_start, hole_len; in ext4_ext_map_blocks() local 4186 hole_start = map->m_lblk; in ext4_ext_map_blocks() 4187 hole_len = ext4_ext_determine_hole(inode, path, &hole_start); in ext4_ext_map_blocks() [all …]
|