Lines Matching refs:node
1109 int node = NUMA_NO_NODE; in dequeue_huge_page_nodemask() local
1124 if (zone_to_nid(zone) == node) in dequeue_huge_page_nodemask()
1126 node = zone_to_nid(zone); in dequeue_huge_page_nodemask()
1128 page = dequeue_huge_page_node_exact(h, node); in dequeue_huge_page_nodemask()
1236 #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask) \ argument
1239 ((node = hstate_next_node_to_alloc(hs, mask)) || 1); \
1242 #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ argument
1245 ((node = hstate_next_node_to_free(hs, mask)) || 1); \
1294 int node; in alloc_gigantic_page() local
1305 for_each_node_mask(node, *nodemask) { in alloc_gigantic_page()
1306 if (node == nid || !hugetlb_cma[node]) in alloc_gigantic_page()
1309 page = cma_alloc(hugetlb_cma[node], nr_pages, in alloc_gigantic_page()
1512 struct llist_node *node; in free_hpage_workfn() local
1515 node = llist_del_all(&hpage_freelist); in free_hpage_workfn()
1517 while (node) { in free_hpage_workfn()
1518 page = container_of((struct address_space **)node, in free_hpage_workfn()
1520 node = node->next; in free_hpage_workfn()
1731 int nr_nodes, node; in alloc_pool_huge_page() local
1734 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in alloc_pool_huge_page()
1735 page = alloc_fresh_huge_page(h, gfp_mask, node, nodes_allowed, in alloc_pool_huge_page()
1758 int nr_nodes, node; in free_pool_huge_page() local
1761 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in free_pool_huge_page()
1766 if ((!acct_surplus || h->surplus_huge_pages_node[node]) && in free_pool_huge_page()
1767 !list_empty(&h->hugepage_freelists[node])) { in free_pool_huge_page()
1769 list_entry(h->hugepage_freelists[node].next, in free_pool_huge_page()
1773 h->free_huge_pages_node[node]--; in free_pool_huge_page()
1776 h->surplus_huge_pages_node[node]--; in free_pool_huge_page()
1997 int node; in alloc_huge_page_vma() local
2000 node = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in alloc_huge_page_vma()
2001 page = alloc_huge_page_nodemask(h, node, nodemask, gfp_mask); in alloc_huge_page_vma()
2454 int nr_nodes, node; in __alloc_bootmem_huge_page() local
2456 for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) { in __alloc_bootmem_huge_page()
2461 0, MEMBLOCK_ALLOC_ACCESSIBLE, node); in __alloc_bootmem_huge_page()
2628 int nr_nodes, node; in adjust_pool_surplus() local
2633 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
2634 if (h->surplus_huge_pages_node[node]) in adjust_pool_surplus()
2638 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
2639 if (h->surplus_huge_pages_node[node] < in adjust_pool_surplus()
2640 h->nr_huge_pages_node[node]) in adjust_pool_surplus()
2648 h->surplus_huge_pages_node[node] += delta; in adjust_pool_surplus()
3087 static void hugetlb_unregister_node(struct node *node) in hugetlb_unregister_node() argument
3090 struct node_hstate *nhs = &node_hstates[node->dev.id]; in hugetlb_unregister_node()
3112 static void hugetlb_register_node(struct node *node) in hugetlb_register_node() argument
3115 struct node_hstate *nhs = &node_hstates[node->dev.id]; in hugetlb_register_node()
3122 &node->dev.kobj); in hugetlb_register_node()
3132 h->name, node->dev.id); in hugetlb_register_node()
3133 hugetlb_unregister_node(node); in hugetlb_register_node()
3149 struct node *node = node_devices[nid]; in hugetlb_register_all_nodes() local
3150 if (node->dev.id == nid) in hugetlb_register_all_nodes()
3151 hugetlb_register_node(node); in hugetlb_register_all_nodes()
3421 int node; in allowed_mems_nr() local
3429 for_each_node_mask(node, cpuset_current_mems_allowed) { in allowed_mems_nr()
3431 (mpol_allowed && node_isset(node, *mpol_allowed))) in allowed_mems_nr()
3432 nr += array[node]; in allowed_mems_nr()