/mm/ |
D | cma.c | 244 phys_addr_t size, phys_addr_t limit, in cma_declare_contiguous_nid() argument 261 __func__, &size, &base, &limit, &alignment); in cma_declare_contiguous_nid() 290 limit &= ~(alignment - 1); in cma_declare_contiguous_nid() 315 if (limit == 0 || limit > memblock_end) in cma_declare_contiguous_nid() 316 limit = memblock_end; in cma_declare_contiguous_nid() 318 if (base + size > limit) { in cma_declare_contiguous_nid() 321 &size, &base, &limit); in cma_declare_contiguous_nid() 341 if (base < highmem_start && limit > highmem_start) { in cma_declare_contiguous_nid() 343 highmem_start, limit, nid, true); in cma_declare_contiguous_nid() 344 limit = highmem_start; in cma_declare_contiguous_nid() [all …]
|
D | slab.c | 187 unsigned int limit; member 385 .limit = BOOT_CPUCACHE_ENTRIES, 523 static void init_arraycache(struct array_cache *ac, int limit, int batch) in init_arraycache() argument 527 ac->limit = limit; in init_arraycache() 579 int nr = min3(from->avail, max, to->limit - to->avail); in transfer_objects() 608 int limit, gfp_t gfp) in alloc_alien_cache() argument 659 static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp) in alloc_alien_cache() argument 664 if (limit > 1) in alloc_alien_cache() 665 limit = 12; in alloc_alien_cache() 673 alc_ptr[i] = __alloc_alien_cache(node, limit, 0xbaadf00d, gfp); in alloc_alien_cache() [all …]
|
D | page-writeback.c | 497 unsigned long limit = node_dirty_limit(pgdat); in node_dirty_ok() local 503 return nr_pages <= limit; in node_dirty_ok() 802 unsigned long limit) in pos_ratio_polynom() argument 808 (limit - setpoint) | 1); in pos_ratio_polynom() 897 unsigned long limit = hard_dirty_limit(dtc_dom(dtc), dtc->thresh); in wb_position_ratio() local 908 if (unlikely(dtc->dirty >= limit)) in wb_position_ratio() 916 setpoint = (freerun + limit) / 2; in wb_position_ratio() 917 pos_ratio = pos_ratio_polynom(setpoint, dtc->dirty, limit); in wb_position_ratio() 1030 wb_thresh = max(wb_thresh, (limit - dtc->dirty) / 8); in wb_position_ratio() 1125 unsigned long limit = dom->dirty_limit; in update_dirty_limit() local [all …]
|
D | hugetlb_cgroup.c | 99 unsigned long limit; in hugetlb_cgroup_init() local 115 limit = round_down(PAGE_COUNTER_MAX, in hugetlb_cgroup_init() 120 limit); in hugetlb_cgroup_init() 124 limit); in hugetlb_cgroup_init() 461 unsigned long limit; in hugetlb_cgroup_read_u64_max() local 468 limit = round_down(PAGE_COUNTER_MAX, in hugetlb_cgroup_read_u64_max() 484 if (val == limit) in hugetlb_cgroup_read_u64_max()
|
D | memblock.c | 1658 static phys_addr_t __init_memblock __find_max_addr(phys_addr_t limit) in __find_max_addr() argument 1669 if (limit <= r->size) { in __find_max_addr() 1670 max_addr = r->base + limit; in __find_max_addr() 1673 limit -= r->size; in __find_max_addr() 1679 void __init memblock_enforce_memory_limit(phys_addr_t limit) in memblock_enforce_memory_limit() argument 1683 if (!limit) in memblock_enforce_memory_limit() 1686 max_addr = __find_max_addr(limit); in memblock_enforce_memory_limit() 1732 void __init memblock_mem_limit_remove_map(phys_addr_t limit) in memblock_mem_limit_remove_map() argument 1736 if (!limit) in memblock_mem_limit_remove_map() 1739 max_addr = __find_max_addr(limit); in memblock_mem_limit_remove_map() [all …]
|
D | sparse.c | 337 unsigned long goal, limit; in sparse_early_usemaps_alloc_pgdat_section() local 350 limit = goal + (1UL << PA_SECTION_SHIFT); in sparse_early_usemaps_alloc_pgdat_section() 353 usage = memblock_alloc_try_nid(size, SMP_CACHE_BYTES, goal, limit, nid); in sparse_early_usemaps_alloc_pgdat_section() 354 if (!usage && limit) { in sparse_early_usemaps_alloc_pgdat_section() 355 limit = 0; in sparse_early_usemaps_alloc_pgdat_section()
|
D | compaction.c | 1441 unsigned int limit = max(1U, freelist_scan_limit(cc) >> 1); in fast_isolate_freepages() local 1459 limit = pageblock_nr_pages >> 1; in fast_isolate_freepages() 1517 limit >>= 1; in fast_isolate_freepages() 1520 if (order_scanned >= limit) in fast_isolate_freepages() 1556 if (order_scanned >= limit) in fast_isolate_freepages() 1557 limit = max(1U, limit >> 1); in fast_isolate_freepages() 1807 unsigned int limit = freelist_scan_limit(cc); in fast_find_migrateblock() local 1856 order >= PAGE_ALLOC_COSTLY_ORDER && !found_block && nr_scanned < limit; in fast_find_migrateblock() 1871 if (nr_scanned++ >= limit) { in fast_find_migrateblock()
|
D | util.c | 486 unsigned long locked_vm, limit; in __account_locked_vm() local 494 limit = task_rlimit(task, RLIMIT_MEMLOCK) >> PAGE_SHIFT; in __account_locked_vm() 495 if (locked_vm + pages > limit) in __account_locked_vm()
|
D | nommu.c | 1463 static int limit; in do_munmap() local 1464 if (limit < 5) { in do_munmap() 1468 limit++; in do_munmap()
|
D | mmap.c | 2370 unsigned long limit; in acct_stack_growth() local 2372 limit = rlimit(RLIMIT_MEMLOCK); in acct_stack_growth() 2373 limit >>= PAGE_SHIFT; in acct_stack_growth() 2374 if (locked > limit && !capable(CAP_IPC_LOCK)) in acct_stack_growth() 2817 unlock_range(struct vm_area_struct *start, unsigned long limit) in unlock_range() argument 2822 while (tmp && tmp->vm_start < limit) { in unlock_range()
|
D | slab.h | 202 unsigned int limit; member
|
D | memcontrol.c | 1342 unsigned long limit; in mem_cgroup_margin() local 1345 limit = READ_ONCE(memcg->memory.max); in mem_cgroup_margin() 1346 if (count < limit) in mem_cgroup_margin() 1347 margin = limit - count; in mem_cgroup_margin() 1351 limit = READ_ONCE(memcg->memsw.max); in mem_cgroup_margin() 1352 if (count < limit) in mem_cgroup_margin() 1353 margin = min(margin, limit - count); in mem_cgroup_margin()
|
D | slab_common.c | 1094 sinfo.limit, sinfo.batchcount, sinfo.shared); in cache_show()
|
D | Kconfig | 730 arch) when the RLIMIT_STACK hard limit is unlimited.
|