Lines Matching refs:size
39 unsigned long size, in __earlyonly_bootmem_alloc() argument
43 return memblock_virt_alloc_try_nid(size, align, goal, in __earlyonly_bootmem_alloc()
50 void * __meminit vmemmap_alloc_block(unsigned long size, int node) in vmemmap_alloc_block() argument
59 get_order(size)); in vmemmap_alloc_block()
63 get_order(size)); in vmemmap_alloc_block()
68 return __earlyonly_bootmem_alloc(node, size, size, in vmemmap_alloc_block()
73 void * __meminit vmemmap_alloc_block_buf(unsigned long size, int node) in vmemmap_alloc_block_buf() argument
78 return vmemmap_alloc_block(size, node); in vmemmap_alloc_block_buf()
81 ptr = (void *)ALIGN((unsigned long)vmemmap_buf, size); in vmemmap_alloc_block_buf()
82 if (ptr + size > vmemmap_buf_end) in vmemmap_alloc_block_buf()
83 return vmemmap_alloc_block(size, node); in vmemmap_alloc_block_buf()
85 vmemmap_buf = ptr + size; in vmemmap_alloc_block_buf()
201 unsigned long size = sizeof(struct page) * PAGES_PER_SECTION; in sparse_mem_maps_populate_node() local
204 size = ALIGN(size, PMD_SIZE); in sparse_mem_maps_populate_node()
205 vmemmap_buf_start = __earlyonly_bootmem_alloc(nodeid, size * map_count, in sparse_mem_maps_populate_node()
210 vmemmap_buf_end = vmemmap_buf_start + size * map_count; in sparse_mem_maps_populate_node()