Lines Matching refs:nodes_allowed

1199 static int next_node_allowed(int nid, nodemask_t *nodes_allowed)  in next_node_allowed()  argument
1201 nid = next_node_in(nid, *nodes_allowed); in next_node_allowed()
1207 static int get_valid_node_allowed(int nid, nodemask_t *nodes_allowed) in get_valid_node_allowed() argument
1209 if (!node_isset(nid, *nodes_allowed)) in get_valid_node_allowed()
1210 nid = next_node_allowed(nid, nodes_allowed); in get_valid_node_allowed()
1221 nodemask_t *nodes_allowed) in hstate_next_node_to_alloc() argument
1225 VM_BUG_ON(!nodes_allowed); in hstate_next_node_to_alloc()
1227 nid = get_valid_node_allowed(h->next_nid_to_alloc, nodes_allowed); in hstate_next_node_to_alloc()
1228 h->next_nid_to_alloc = next_node_allowed(nid, nodes_allowed); in hstate_next_node_to_alloc()
1239 static int hstate_next_node_to_free(struct hstate *h, nodemask_t *nodes_allowed) in hstate_next_node_to_free() argument
1243 VM_BUG_ON(!nodes_allowed); in hstate_next_node_to_free()
1245 nid = get_valid_node_allowed(h->next_nid_to_free, nodes_allowed); in hstate_next_node_to_free()
1246 h->next_nid_to_free = next_node_allowed(nid, nodes_allowed); in hstate_next_node_to_free()
1884 static int alloc_pool_huge_page(struct hstate *h, nodemask_t *nodes_allowed, in alloc_pool_huge_page() argument
1891 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in alloc_pool_huge_page()
1892 page = alloc_fresh_huge_page(h, gfp_mask, node, nodes_allowed, in alloc_pool_huge_page()
1914 nodemask_t *nodes_allowed, in remove_pool_huge_page() argument
1921 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in remove_pool_huge_page()
3017 nodemask_t *nodes_allowed) in try_to_free_low() argument
3029 for_each_node_mask(i, *nodes_allowed) { in try_to_free_low()
3049 nodemask_t *nodes_allowed) in try_to_free_low() argument
3059 static int adjust_pool_surplus(struct hstate *h, nodemask_t *nodes_allowed, in adjust_pool_surplus() argument
3068 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
3073 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
3089 nodemask_t *nodes_allowed) in set_max_huge_pages() argument
3163 if (!adjust_pool_surplus(h, nodes_allowed, -1)) in set_max_huge_pages()
3178 ret = alloc_pool_huge_page(h, nodes_allowed, in set_max_huge_pages()
3206 try_to_free_low(h, min_count, nodes_allowed); in set_max_huge_pages()
3212 page = remove_pool_huge_page(h, nodes_allowed, 0); in set_max_huge_pages()
3225 if (!adjust_pool_surplus(h, nodes_allowed, 1)) in set_max_huge_pages()
3285 nodemask_t nodes_allowed, *n_mask; in __nr_hugepages_store_common() local
3295 init_nodemask_of_mempolicy(&nodes_allowed))) in __nr_hugepages_store_common()
3298 n_mask = &nodes_allowed; in __nr_hugepages_store_common()
3304 init_nodemask_of_node(&nodes_allowed, nid); in __nr_hugepages_store_common()
3305 n_mask = &nodes_allowed; in __nr_hugepages_store_common()