Lines Matching refs:SWAPFILE_CLUSTER
206 #define SWAPFILE_CLUSTER HPAGE_PMD_NR macro
210 #define SWAPFILE_CLUSTER 256 macro
297 ci += offset / SWAPFILE_CLUSTER; in lock_cluster()
401 memset(si->swap_map + idx * SWAPFILE_CLUSTER, in swap_cluster_schedule_discard()
402 SWAP_MAP_BAD, SWAPFILE_CLUSTER); in swap_cluster_schedule_discard()
432 discard_swap_cluster(si, idx * SWAPFILE_CLUSTER, in swap_do_scheduled_discard()
433 SWAPFILE_CLUSTER); in swap_do_scheduled_discard()
436 ci = lock_cluster(si, idx * SWAPFILE_CLUSTER); in swap_do_scheduled_discard()
438 memset(si->swap_map + idx * SWAPFILE_CLUSTER, in swap_do_scheduled_discard()
439 0, SWAPFILE_CLUSTER); in swap_do_scheduled_discard()
490 unsigned long idx = page_nr / SWAPFILE_CLUSTER; in inc_cluster_info_page()
497 VM_BUG_ON(cluster_count(&cluster_info[idx]) >= SWAPFILE_CLUSTER); in inc_cluster_info_page()
510 unsigned long idx = page_nr / SWAPFILE_CLUSTER; in dec_cluster_info_page()
534 offset /= SWAPFILE_CLUSTER; in scan_swap_map_ssd_cluster_conflict()
565 SWAPFILE_CLUSTER; in scan_swap_map_try_ssd_cluster()
586 (cluster_next(&cluster->index) + 1) * SWAPFILE_CLUSTER); in scan_swap_map_try_ssd_cluster()
721 if (si->pages - si->inuse_pages < SWAPFILE_CLUSTER) { in scan_swap_map_slots()
722 si->cluster_nr = SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
735 last_in_cluster = offset + SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
740 last_in_cluster = offset + SWAPFILE_CLUSTER; in scan_swap_map_slots()
743 offset -= SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
745 si->cluster_nr = SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
756 si->cluster_nr = SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
901 offset = idx * SWAPFILE_CLUSTER; in swap_alloc_cluster()
904 cluster_set_count_flag(ci, SWAPFILE_CLUSTER, CLUSTER_FLAG_HUGE); in swap_alloc_cluster()
907 for (i = 0; i < SWAPFILE_CLUSTER; i++) in swap_alloc_cluster()
910 swap_range_alloc(si, offset, SWAPFILE_CLUSTER); in swap_alloc_cluster()
918 unsigned long offset = idx * SWAPFILE_CLUSTER; in swap_free_cluster()
925 swap_range_free(si, offset, SWAPFILE_CLUSTER); in swap_free_cluster()
952 WARN_ON_ONCE(n_goal > 1 && size == SWAPFILE_CLUSTER); in get_swap_pages()
991 if (size == SWAPFILE_CLUSTER) { in get_swap_pages()
998 if (n_ret || size == SWAPFILE_CLUSTER) in get_swap_pages()
1214 unsigned long idx = offset / SWAPFILE_CLUSTER; in put_swap_page()
1227 if (size == SWAPFILE_CLUSTER) { in put_swap_page()
1230 for (i = 0; i < SWAPFILE_CLUSTER; i++) { in put_swap_page()
1237 if (free_entries == SWAPFILE_CLUSTER) { in put_swap_page()
1241 memset(map, 0, SWAPFILE_CLUSTER); in put_swap_page()
1243 mem_cgroup_uncharge_swap(entry, SWAPFILE_CLUSTER); in put_swap_page()
1424 unsigned long offset = round_down(roffset, SWAPFILE_CLUSTER); in swap_page_trans_huge_swapped()
1434 for (i = 0; i < SWAPFILE_CLUSTER; i++) { in swap_page_trans_huge_swapped()
2996 DIV_ROUND_UP(SWAP_ADDRESS_SPACE_PAGES, SWAPFILE_CLUSTER)
3010 unsigned long nr_clusters = DIV_ROUND_UP(maxpages, SWAPFILE_CLUSTER); in setup_swap_map_and_extents()
3011 unsigned long col = p->cluster_next / SWAPFILE_CLUSTER % SWAP_CLUSTER_COLS; in setup_swap_map_and_extents()
3035 for (i = maxpages; i < round_up(maxpages, SWAPFILE_CLUSTER); i++) in setup_swap_map_and_extents()
3194 nr_cluster = DIV_ROUND_UP(maxpages, SWAPFILE_CLUSTER); in SYSCALL_DEFINE2()