Lines Matching refs:nodes_allowed

974 static int next_node_allowed(int nid, nodemask_t *nodes_allowed)  in next_node_allowed()  argument
976 nid = next_node_in(nid, *nodes_allowed); in next_node_allowed()
982 static int get_valid_node_allowed(int nid, nodemask_t *nodes_allowed) in get_valid_node_allowed() argument
984 if (!node_isset(nid, *nodes_allowed)) in get_valid_node_allowed()
985 nid = next_node_allowed(nid, nodes_allowed); in get_valid_node_allowed()
996 nodemask_t *nodes_allowed) in hstate_next_node_to_alloc() argument
1000 VM_BUG_ON(!nodes_allowed); in hstate_next_node_to_alloc()
1002 nid = get_valid_node_allowed(h->next_nid_to_alloc, nodes_allowed); in hstate_next_node_to_alloc()
1003 h->next_nid_to_alloc = next_node_allowed(nid, nodes_allowed); in hstate_next_node_to_alloc()
1014 static int hstate_next_node_to_free(struct hstate *h, nodemask_t *nodes_allowed) in hstate_next_node_to_free() argument
1018 VM_BUG_ON(!nodes_allowed); in hstate_next_node_to_free()
1020 nid = get_valid_node_allowed(h->next_nid_to_free, nodes_allowed); in hstate_next_node_to_free()
1021 h->next_nid_to_free = next_node_allowed(nid, nodes_allowed); in hstate_next_node_to_free()
1421 static int alloc_pool_huge_page(struct hstate *h, nodemask_t *nodes_allowed) in alloc_pool_huge_page() argument
1427 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in alloc_pool_huge_page()
1428 page = alloc_fresh_huge_page(h, gfp_mask, node, nodes_allowed); in alloc_pool_huge_page()
1447 static int free_pool_huge_page(struct hstate *h, nodemask_t *nodes_allowed, in free_pool_huge_page() argument
1453 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in free_pool_huge_page()
2216 nodemask_t *nodes_allowed) in try_to_free_low() argument
2223 for_each_node_mask(i, *nodes_allowed) { in try_to_free_low()
2240 nodemask_t *nodes_allowed) in try_to_free_low() argument
2250 static int adjust_pool_surplus(struct hstate *h, nodemask_t *nodes_allowed, in adjust_pool_surplus() argument
2258 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
2263 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
2279 nodemask_t *nodes_allowed) in set_max_huge_pages() argument
2299 if (!adjust_pool_surplus(h, nodes_allowed, -1)) in set_max_huge_pages()
2314 ret = alloc_pool_huge_page(h, nodes_allowed); in set_max_huge_pages()
2341 try_to_free_low(h, min_count, nodes_allowed); in set_max_huge_pages()
2343 if (!free_pool_huge_page(h, nodes_allowed, 0)) in set_max_huge_pages()
2348 if (!adjust_pool_surplus(h, nodes_allowed, 1)) in set_max_huge_pages()
2404 NODEMASK_ALLOC(nodemask_t, nodes_allowed, GFP_KERNEL | __GFP_NORETRY); in __nr_hugepages_store_common()
2416 init_nodemask_of_mempolicy(nodes_allowed))) { in __nr_hugepages_store_common()
2417 NODEMASK_FREE(nodes_allowed); in __nr_hugepages_store_common()
2418 nodes_allowed = &node_states[N_MEMORY]; in __nr_hugepages_store_common()
2420 } else if (nodes_allowed) { in __nr_hugepages_store_common()
2426 init_nodemask_of_node(nodes_allowed, nid); in __nr_hugepages_store_common()
2428 nodes_allowed = &node_states[N_MEMORY]; in __nr_hugepages_store_common()
2430 h->max_huge_pages = set_max_huge_pages(h, count, nodes_allowed); in __nr_hugepages_store_common()
2432 if (nodes_allowed != &node_states[N_MEMORY]) in __nr_hugepages_store_common()
2433 NODEMASK_FREE(nodes_allowed); in __nr_hugepages_store_common()
2437 NODEMASK_FREE(nodes_allowed); in __nr_hugepages_store_common()