Home
last modified time | relevance | path

Searched refs:delta (Results 1 – 9 of 9) sorted by relevance

/mm/
Dpage_counter.c20 long delta; in propagate_protected_usage() local
32 delta = protected - old_protected; in propagate_protected_usage()
33 if (delta) in propagate_protected_usage()
34 atomic_long_add(delta, &c->parent->children_min_usage); in propagate_protected_usage()
44 delta = protected - old_protected; in propagate_protected_usage()
45 if (delta) in propagate_protected_usage()
46 atomic_long_add(delta, &c->parent->children_low_usage); in propagate_protected_usage()
Dvmstat.c317 long delta) in __mod_zone_page_state() argument
324 x = delta + __this_cpu_read(*p); in __mod_zone_page_state()
337 long delta) in __mod_node_page_state() argument
344 x = delta + __this_cpu_read(*p); in __mod_node_page_state()
481 enum zone_stat_item item, long delta, int overstep_mode) in mod_zone_state() argument
503 n = delta + o; in mod_zone_state()
519 long delta) in mod_zone_page_state() argument
521 mod_zone_state(zone, item, delta, 0); in mod_zone_page_state()
538 enum node_stat_item item, int delta, int overstep_mode) in mod_node_state() argument
560 n = delta + o; in mod_node_state()
[all …]
Dhugetlb.c75 static int hugetlb_acct_memory(struct hstate *h, long delta);
135 long delta) in hugepage_subpool_get_pages() argument
137 long ret = delta; in hugepage_subpool_get_pages()
145 if ((spool->used_hpages + delta) <= spool->max_hpages) in hugepage_subpool_get_pages()
146 spool->used_hpages += delta; in hugepage_subpool_get_pages()
155 if (delta > spool->rsv_hpages) { in hugepage_subpool_get_pages()
160 ret = delta - spool->rsv_hpages; in hugepage_subpool_get_pages()
164 spool->rsv_hpages -= delta; in hugepage_subpool_get_pages()
180 long delta) in hugepage_subpool_put_pages() argument
182 long ret = delta; in hugepage_subpool_put_pages()
[all …]
Dslob.c241 int delta = 0, units = SLOB_UNITS(size); in slob_page_alloc() local
258 delta = aligned - cur; in slob_page_alloc()
260 if (avail >= units + delta) { /* room enough? */ in slob_page_alloc()
263 if (delta) { /* need to fragment head to align? */ in slob_page_alloc()
265 set_slob(aligned, avail - delta, next); in slob_page_alloc()
266 set_slob(cur, delta, aligned); in slob_page_alloc()
Dvmscan.c466 unsigned long long delta; in do_shrink_slab() local
492 delta = freeable >> priority; in do_shrink_slab()
493 delta *= 4; in do_shrink_slab()
494 do_div(delta, shrinker->seeks); in do_shrink_slab()
501 delta = freeable / 2; in do_shrink_slab()
504 total_scan += delta; in do_shrink_slab()
525 if (delta < freeable / 4) in do_shrink_slab()
537 freeable, delta, total_scan, priority); in do_shrink_slab()
4179 unsigned long delta = 0; in node_pagecache_reclaimable() local
4194 delta += node_page_state(pgdat, NR_FILE_DIRTY); in node_pagecache_reclaimable()
[all …]
Dmremap.c575 static int vma_expandable(struct vm_area_struct *vma, unsigned long delta) in vma_expandable() argument
577 unsigned long end = vma->vm_end + delta; in vma_expandable()
Dcompaction.c34 static inline void count_compact_events(enum vm_event_item item, long delta) in count_compact_events() argument
36 count_vm_events(item, delta); in count_compact_events()
40 #define count_compact_events(item, delta) do { } while (0) argument
Dpercpu.c2962 unsigned long delta; in setup_per_cpu_areas() local
2976 delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start; in setup_per_cpu_areas()
2978 __per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu]; in setup_per_cpu_areas()
Dpage-writeback.c1125 long delta = avg - wb->avg_write_bandwidth; in wb_update_write_bandwidth() local
1126 WARN_ON_ONCE(atomic_long_add_return(delta, in wb_update_write_bandwidth()