Searched refs:delta (Results 1 – 9 of 9) sorted by relevance
/mm/ |
D | page_counter.c | 20 long delta; in propagate_protected_usage() local 32 delta = protected - old_protected; in propagate_protected_usage() 33 if (delta) in propagate_protected_usage() 34 atomic_long_add(delta, &c->parent->children_min_usage); in propagate_protected_usage() 44 delta = protected - old_protected; in propagate_protected_usage() 45 if (delta) in propagate_protected_usage() 46 atomic_long_add(delta, &c->parent->children_low_usage); in propagate_protected_usage()
|
D | vmstat.c | 317 long delta) in __mod_zone_page_state() argument 324 x = delta + __this_cpu_read(*p); in __mod_zone_page_state() 337 long delta) in __mod_node_page_state() argument 344 x = delta + __this_cpu_read(*p); in __mod_node_page_state() 481 enum zone_stat_item item, long delta, int overstep_mode) in mod_zone_state() argument 503 n = delta + o; in mod_zone_state() 519 long delta) in mod_zone_page_state() argument 521 mod_zone_state(zone, item, delta, 0); in mod_zone_page_state() 538 enum node_stat_item item, int delta, int overstep_mode) in mod_node_state() argument 560 n = delta + o; in mod_node_state() [all …]
|
D | hugetlb.c | 75 static int hugetlb_acct_memory(struct hstate *h, long delta); 135 long delta) in hugepage_subpool_get_pages() argument 137 long ret = delta; in hugepage_subpool_get_pages() 145 if ((spool->used_hpages + delta) <= spool->max_hpages) in hugepage_subpool_get_pages() 146 spool->used_hpages += delta; in hugepage_subpool_get_pages() 155 if (delta > spool->rsv_hpages) { in hugepage_subpool_get_pages() 160 ret = delta - spool->rsv_hpages; in hugepage_subpool_get_pages() 164 spool->rsv_hpages -= delta; in hugepage_subpool_get_pages() 180 long delta) in hugepage_subpool_put_pages() argument 182 long ret = delta; in hugepage_subpool_put_pages() [all …]
|
D | slob.c | 241 int delta = 0, units = SLOB_UNITS(size); in slob_page_alloc() local 258 delta = aligned - cur; in slob_page_alloc() 260 if (avail >= units + delta) { /* room enough? */ in slob_page_alloc() 263 if (delta) { /* need to fragment head to align? */ in slob_page_alloc() 265 set_slob(aligned, avail - delta, next); in slob_page_alloc() 266 set_slob(cur, delta, aligned); in slob_page_alloc()
|
D | vmscan.c | 466 unsigned long long delta; in do_shrink_slab() local 492 delta = freeable >> priority; in do_shrink_slab() 493 delta *= 4; in do_shrink_slab() 494 do_div(delta, shrinker->seeks); in do_shrink_slab() 501 delta = freeable / 2; in do_shrink_slab() 504 total_scan += delta; in do_shrink_slab() 525 if (delta < freeable / 4) in do_shrink_slab() 537 freeable, delta, total_scan, priority); in do_shrink_slab() 4179 unsigned long delta = 0; in node_pagecache_reclaimable() local 4194 delta += node_page_state(pgdat, NR_FILE_DIRTY); in node_pagecache_reclaimable() [all …]
|
D | mremap.c | 575 static int vma_expandable(struct vm_area_struct *vma, unsigned long delta) in vma_expandable() argument 577 unsigned long end = vma->vm_end + delta; in vma_expandable()
|
D | compaction.c | 34 static inline void count_compact_events(enum vm_event_item item, long delta) in count_compact_events() argument 36 count_vm_events(item, delta); in count_compact_events() 40 #define count_compact_events(item, delta) do { } while (0) argument
|
D | percpu.c | 2962 unsigned long delta; in setup_per_cpu_areas() local 2976 delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start; in setup_per_cpu_areas() 2978 __per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu]; in setup_per_cpu_areas()
|
D | page-writeback.c | 1125 long delta = avg - wb->avg_write_bandwidth; in wb_update_write_bandwidth() local 1126 WARN_ON_ONCE(atomic_long_add_return(delta, in wb_update_write_bandwidth()
|