Lines Matching refs:SWAPFILE_CLUSTER
254 #define SWAPFILE_CLUSTER HPAGE_PMD_NR macro
258 #define SWAPFILE_CLUSTER 256 macro
345 ci += offset / SWAPFILE_CLUSTER; in lock_cluster()
449 memset(si->swap_map + idx * SWAPFILE_CLUSTER, in swap_cluster_schedule_discard()
450 SWAP_MAP_BAD, SWAPFILE_CLUSTER); in swap_cluster_schedule_discard()
480 discard_swap_cluster(si, idx * SWAPFILE_CLUSTER, in swap_do_scheduled_discard()
481 SWAPFILE_CLUSTER); in swap_do_scheduled_discard()
484 ci = lock_cluster(si, idx * SWAPFILE_CLUSTER); in swap_do_scheduled_discard()
486 memset(si->swap_map + idx * SWAPFILE_CLUSTER, in swap_do_scheduled_discard()
487 0, SWAPFILE_CLUSTER); in swap_do_scheduled_discard()
538 unsigned long idx = page_nr / SWAPFILE_CLUSTER; in inc_cluster_info_page()
545 VM_BUG_ON(cluster_count(&cluster_info[idx]) >= SWAPFILE_CLUSTER); in inc_cluster_info_page()
558 unsigned long idx = page_nr / SWAPFILE_CLUSTER; in dec_cluster_info_page()
582 offset /= SWAPFILE_CLUSTER; in scan_swap_map_ssd_cluster_conflict()
613 SWAPFILE_CLUSTER; in scan_swap_map_try_ssd_cluster()
634 (cluster_next(&cluster->index) + 1) * SWAPFILE_CLUSTER); in scan_swap_map_try_ssd_cluster()
769 if (si->pages - si->inuse_pages < SWAPFILE_CLUSTER) { in scan_swap_map_slots()
770 si->cluster_nr = SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
783 last_in_cluster = offset + SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
788 last_in_cluster = offset + SWAPFILE_CLUSTER; in scan_swap_map_slots()
791 offset -= SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
793 si->cluster_nr = SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
804 si->cluster_nr = SWAPFILE_CLUSTER - 1; in scan_swap_map_slots()
949 offset = idx * SWAPFILE_CLUSTER; in swap_alloc_cluster()
952 cluster_set_count_flag(ci, SWAPFILE_CLUSTER, CLUSTER_FLAG_HUGE); in swap_alloc_cluster()
955 for (i = 0; i < SWAPFILE_CLUSTER; i++) in swap_alloc_cluster()
958 swap_range_alloc(si, offset, SWAPFILE_CLUSTER); in swap_alloc_cluster()
966 unsigned long offset = idx * SWAPFILE_CLUSTER; in swap_free_cluster()
970 memset(si->swap_map + offset, 0, SWAPFILE_CLUSTER); in swap_free_cluster()
974 swap_range_free(si, offset, SWAPFILE_CLUSTER); in swap_free_cluster()
1001 WARN_ON_ONCE(n_goal > 1 && size == SWAPFILE_CLUSTER); in get_swap_pages()
1040 if (size == SWAPFILE_CLUSTER) { in get_swap_pages()
1047 if (n_ret || size == SWAPFILE_CLUSTER) in get_swap_pages()
1328 unsigned long idx = offset / SWAPFILE_CLUSTER; in put_swap_page()
1341 if (size == SWAPFILE_CLUSTER) { in put_swap_page()
1344 for (i = 0; i < SWAPFILE_CLUSTER; i++) { in put_swap_page()
1351 if (free_entries == SWAPFILE_CLUSTER) { in put_swap_page()
1354 mem_cgroup_uncharge_swap(entry, SWAPFILE_CLUSTER); in put_swap_page()
1544 unsigned long offset = round_down(roffset, SWAPFILE_CLUSTER); in swap_page_trans_huge_swapped()
1554 for (i = 0; i < SWAPFILE_CLUSTER; i++) { in swap_page_trans_huge_swapped()
3003 DIV_ROUND_UP(SWAP_ADDRESS_SPACE_PAGES, SWAPFILE_CLUSTER)
3017 unsigned long nr_clusters = DIV_ROUND_UP(maxpages, SWAPFILE_CLUSTER); in setup_swap_map_and_extents()
3018 unsigned long col = p->cluster_next / SWAPFILE_CLUSTER % SWAP_CLUSTER_COLS; in setup_swap_map_and_extents()
3042 for (i = maxpages; i < round_up(maxpages, SWAPFILE_CLUSTER); i++) in setup_swap_map_and_extents()
3201 nr_cluster = DIV_ROUND_UP(maxpages, SWAPFILE_CLUSTER); in SYSCALL_DEFINE2()