/mm/ |
D | readahead.c | 268 static unsigned long get_init_ra_size(unsigned long size, unsigned long max) in get_init_ra_size() argument 272 if (newsize <= max / 32) in get_init_ra_size() 274 else if (newsize <= max / 4) in get_init_ra_size() 277 newsize = max; in get_init_ra_size() 287 unsigned long max) in get_next_ra_size() argument 292 if (cur < max / 16) in get_next_ra_size() 297 return min(newsize, max); in get_next_ra_size() 348 int max = ra->ra_pages; /* max readahead pages */ in ondemand_readahead() local 359 ra->size = get_next_ra_size(ra, max); in ondemand_readahead() 365 sequential = offset - prev_offset <= 1UL || req_size > max; in ondemand_readahead() [all …]
|
D | quicklist.c | 28 unsigned long node_free_pages, max; in max_pages() local 43 max = node_free_pages / FRACTION_OF_NODE_MEM; in max_pages() 46 max /= num_cpus_on_node; in max_pages() 48 return max(max, min_pages); in max_pages()
|
D | bootmem.c | 300 unsigned long max; in mark_bootmem() local 308 max = min(bdata->node_low_pfn, end); in mark_bootmem() 310 err = mark_bootmem_node(bdata, pos, max, reserve, flags); in mark_bootmem() 316 if (max == end) in mark_bootmem() 436 unsigned long min, max, start, sidx, midx, step; in alloc_bootmem_core() local 450 max = bdata->node_low_pfn; in alloc_bootmem_core() 455 if (limit && max > limit) in alloc_bootmem_core() 456 max = limit; in alloc_bootmem_core() 457 if (max <= min) in alloc_bootmem_core() 460 step = max(align >> PAGE_SHIFT, 1UL); in alloc_bootmem_core() [all …]
|
D | swapfile.c | 207 si->lowest_alloc = si->max; in scan_swap_map() 285 si->lowest_bit = si->max; in scan_swap_map() 452 if (offset >= p->max) in swap_info_get() 869 unsigned int max = si->max; in find_next_to_unuse() local 880 if (++i >= max) { in find_next_to_unuse() 889 max = prev + 1; in find_next_to_unuse() 1299 ret = add_swap_extent(sis, 0, sis->max, 0); in setup_swap_extents() 1315 page_no < sis->max) { in setup_swap_extents() 1369 sis->max = page_no; in setup_swap_extents() 1496 p->max = 0; in SYSCALL_DEFINE1() [all …]
|
D | page_alloc.c | 2515 return max(size, 4UL); in wait_table_hash_nr_entries() 2744 pcp->batch = max(1UL, 1 * batch); in setup_pageset() 2760 pcp->batch = max(1UL, high/4); in setup_pagelist_highmark() 3190 *end_pfn = max(*end_pfn, early_node_map[i].end_pfn); in get_pfn_range_for_nid() 3282 zone_start_pfn = max(zone_start_pfn, node_start_pfn); in zone_spanned_pages_in_node() 3320 prev_end_pfn = max(prev_end_pfn, range_start_pfn); in __absent_pages_in_range() 3333 max(range_start_pfn, prev_end_pfn); in __absent_pages_in_range() 3360 zone_start_pfn = max(arch_zone_lowest_possible_pfn[zone_type], in zone_absent_pages_in_node() 3915 required_kernelcore = max(required_kernelcore, corepages); in find_zone_movable_pfns_for_nodes() 3950 start_pfn = max(early_node_map[i].start_pfn, in find_zone_movable_pfns_for_nodes() [all …]
|
D | util.c | 36 char *kstrndup(const char *s, size_t max, gfp_t gfp) in kstrndup() argument 44 len = strnlen(s, max); in kstrndup()
|
D | page_cgroup.c | 319 unsigned long idx, max; in swap_cgroup_prepare() local 333 max = idx; in swap_cgroup_prepare() 334 for (idx = 0; idx < max; idx++) in swap_cgroup_prepare()
|
D | slub.c | 1823 for (order = max(min_order, in slab_order() 1905 align = max(align, ralign); in calculate_alignment() 2300 if (oo_objects(s->oo) > oo_objects(s->max)) in calculate_sizes() 2301 s->max = s->oo; in calculate_sizes() 2776 int objects = oo_objects(s->max); in kmem_cache_shrink() 3122 s->objsize = max(s->objsize, (int)size); in kmem_cache_create() 3368 unsigned long *map = kmalloc(BITS_TO_LONGS(oo_objects(s->max)) * in validate_slab_cache() 3460 unsigned long max; member 3467 if (t->max) in free_loc_track() 3469 get_order(sizeof(struct location) * t->max)); in free_loc_track() [all …]
|
D | slob.c | 465 int align = max(ARCH_KMALLOC_MINALIGN, ARCH_SLAB_MINALIGN); in __kmalloc_node() 499 int align = max(ARCH_KMALLOC_MINALIGN, ARCH_SLAB_MINALIGN); in kfree() 518 int align = max(ARCH_KMALLOC_MINALIGN, ARCH_SLAB_MINALIGN); in ksize()
|
D | memory_hotplug.c | 173 zone->spanned_pages = max(old_zone_end_pfn, end_pfn) - in grow_zone_span() 188 pgdat->node_spanned_pages = max(old_pgdat_end_pfn, end_pfn) - in grow_pgdat_span() 351 max_mapnr = max(page_to_pfn(page), max_mapnr); in online_page()
|
D | slab.c | 142 #define REDZONE_ALIGN max(BYTES_PER_WORD, __alignof__(unsigned long long)) 975 struct array_cache *from, unsigned int max) in transfer_objects() argument 978 int nr = min(min(from->avail, max), to->limit - to->avail); in transfer_objects() 3475 int max = shared_array->limit - shared_array->avail; in cache_flusharray() local 3476 if (max) { in cache_flusharray() 3477 if (batchcount > max) in cache_flusharray() 3478 batchcount = max; in cache_flusharray()
|
D | swap.c | 522 #define ACCT_THRESHOLD max(16, NR_CPUS * 2)
|
D | hugetlb.c | 202 seg_from = max(rg->from, f); in region_count() 1198 min_count = max(count, min_count); in set_max_huge_pages()
|
D | shmem.c | 2290 sbinfo = kmalloc(max((int)sizeof(struct shmem_sb_info), in shmem_fill_super()
|
D | memory.c | 975 start = max(vma->vm_start, start_addr); in unmap_vmas()
|