/mm/ |
D | swap_cgroup.c | 98 unsigned short old, unsigned short new) in swap_cgroup_cmpxchg() argument 109 if (retval == old) in swap_cgroup_cmpxchg() 131 unsigned short old; in swap_cgroup_record() local 139 old = sc->id; in swap_cgroup_record() 141 VM_BUG_ON(sc->id != old); in swap_cgroup_record() 153 return old; in swap_cgroup_record()
|
D | page_counter.c | 174 unsigned long old; in page_counter_set_max() local 193 old = xchg(&counter->max, nr_pages); in page_counter_set_max() 195 if (page_counter_read(counter) <= usage || nr_pages >= old) in page_counter_set_max() 198 counter->max = old; in page_counter_set_max()
|
D | mempolicy.c | 782 struct mempolicy *old; in vma_replace_policy() local 802 old = vma->vm_policy; in vma_replace_policy() 804 mpol_put(old); in vma_replace_policy() 864 struct mempolicy *new, *old; in do_set_mempolicy() local 885 old = current->mempolicy; in do_set_mempolicy() 890 mpol_put(old); in do_set_mempolicy() 1511 struct mempolicy *new, *old; in SYSCALL_DEFINE4() local 1547 old = vma_policy(vma); in SYSCALL_DEFINE4() 1548 if (!old) { in SYSCALL_DEFINE4() 1552 if (old->mode != MPOL_BIND && old->mode != MPOL_PREFERRED_MANY) { in SYSCALL_DEFINE4() [all …]
|
D | swap_state.c | 93 void *old; in add_to_swap_cache() local 112 old = xas_load(&xas); in add_to_swap_cache() 113 if (xa_is_value(old)) { in add_to_swap_cache() 115 *shadowp = old; in add_to_swap_cache() 250 void *old; in clear_shadow_from_swap_cache() local 260 xas_for_each(&xas, old, end) { in clear_shadow_from_swap_cache() 261 if (!xa_is_value(old)) in clear_shadow_from_swap_cache()
|
D | filemap.c | 749 errseq_t old = READ_ONCE(file->f_wb_err); in file_check_and_advance_wb_err() local 753 if (errseq_check(&mapping->wb_err, old)) { in file_check_and_advance_wb_err() 756 old = file->f_wb_err; in file_check_and_advance_wb_err() 759 trace_file_check_and_advance_wb_err(file, old); in file_check_and_advance_wb_err() 825 void replace_page_cache_folio(struct folio *old, struct folio *new) in replace_page_cache_folio() argument 827 struct address_space *mapping = old->mapping; in replace_page_cache_folio() 829 pgoff_t offset = old->index; in replace_page_cache_folio() 832 VM_BUG_ON_FOLIO(!folio_test_locked(old), old); in replace_page_cache_folio() 840 mem_cgroup_migrate(old, new); in replace_page_cache_folio() 845 old->mapping = NULL; in replace_page_cache_folio() [all …]
|
D | shrinker_debug.c | 212 const char *new, *old; in shrinker_debugfs_rename() local 225 old = shrinker->name; in shrinker_debugfs_rename() 243 kfree_const(old); in shrinker_debugfs_rename()
|
D | memcontrol.c | 2318 struct mem_cgroup *old = READ_ONCE(stock->cached); in drain_stock() local 2320 if (!old) in drain_stock() 2324 page_counter_uncharge(&old->memory, stock->nr_pages); in drain_stock() 2326 page_counter_uncharge(&old->memsw, stock->nr_pages); in drain_stock() 2330 css_put(&old->css); in drain_stock() 2337 struct obj_cgroup *old = NULL; in drain_local_stock() local 2348 old = drain_obj_stock(stock); in drain_local_stock() 2353 if (old) in drain_local_stock() 2354 obj_cgroup_put(old); in drain_local_stock() 3225 struct obj_cgroup *old = NULL; in mod_objcg_state() local [all …]
|
D | page_pinner.c | 384 struct captured_pinner *new, *old; in buffer_size_set() local 391 old = pp_buffer.buffer; in buffer_size_set() 396 kvfree(old); in buffer_size_set()
|
D | shmem.c | 863 void *old; in shmem_free_swap() local 865 old = xa_cmpxchg_irq(&mapping->i_pages, index, radswap, NULL, 0); in shmem_free_swap() 866 if (old != radswap) in shmem_free_swap() 1747 struct folio *old, *new; in shmem_replace_folio() local 1753 old = *foliop; in shmem_replace_folio() 1754 entry = old->swap; in shmem_replace_folio() 1763 VM_BUG_ON_FOLIO(folio_test_large(old), old); in shmem_replace_folio() 1769 folio_copy(new, old); in shmem_replace_folio() 1783 error = shmem_replace_entry(swap_mapping, swap_index, old, new); in shmem_replace_folio() 1785 mem_cgroup_migrate(old, new); in shmem_replace_folio() [all …]
|
D | slub.c | 519 freelist_aba_t old = { .freelist = freelist_old, .counter = counters_old }; in __update_freelist_fast() local 522 return try_cmpxchg_freelist(&slab->freelist_counter.full, &old.full, new.full); in __update_freelist_fast() 2526 struct slab old; in deactivate_slab() local 2570 old.freelist = READ_ONCE(slab->freelist); in deactivate_slab() 2571 old.counters = READ_ONCE(slab->counters); in deactivate_slab() 2572 VM_BUG_ON(!old.frozen); in deactivate_slab() 2575 new.counters = old.counters; in deactivate_slab() 2578 set_freepointer(s, freelist_tail, old.freelist); in deactivate_slab() 2581 new.freelist = old.freelist; in deactivate_slab() 2600 old.freelist, old.counters, in deactivate_slab() [all …]
|
D | pgtable-generic.c | 203 pmd_t old = pmdp_establish(vma, address, pmdp, pmd_mkinvalid(*pmdp)); in pmdp_invalidate() local 205 return old; in pmdp_invalidate()
|
D | page-writeback.c | 1209 unsigned long old = wb->write_bandwidth; in wb_update_write_bandwidth() local 1235 if (avg > old && old >= (unsigned long)bw) in wb_update_write_bandwidth() 1236 avg -= (avg - old) >> 3; in wb_update_write_bandwidth() 1238 if (avg < old && old <= (unsigned long)bw) in wb_update_write_bandwidth() 1239 avg += (old - avg) >> 3; in wb_update_write_bandwidth()
|
D | page_owner.c | 254 void __folio_copy_owner(struct folio *newfolio, struct folio *old) in __folio_copy_owner() argument 260 old_ext = page_ext_get(&old->page); in __folio_copy_owner()
|
D | vmscan.c | 229 struct shrinker_info *new, *old; in expand_one_shrinker_info() local 236 old = shrinker_info_protected(memcg, nid); in expand_one_shrinker_info() 238 if (!old) in expand_one_shrinker_info() 242 if (new_nr_max <= old->map_nr_max) in expand_one_shrinker_info() 257 memcpy(new->nr_deferred, old->nr_deferred, old_defer_size); in expand_one_shrinker_info() 262 kvfree_rcu(old, rcu); in expand_one_shrinker_info() 4906 int old, new; in lru_gen_rotate_memcg() local 4916 new = old = lruvec->lrugen.gen; in lru_gen_rotate_memcg() 4940 pgdat->memcg_lru.nr_memcgs[old]--; in lru_gen_rotate_memcg() 4943 if (!pgdat->memcg_lru.nr_memcgs[old] && old == get_memcg_gen(pgdat->memcg_lru.seq)) in lru_gen_rotate_memcg() [all …]
|
D | compaction.c | 3250 int ret, old; in proc_dointvec_minmax_warn_RT_change() local 3255 old = *(int *)table->data; in proc_dointvec_minmax_warn_RT_change() 3259 if (old != *(int *)table->data) in proc_dointvec_minmax_warn_RT_change()
|
D | mmap.c | 1101 static struct anon_vma *reusable_anon_vma(struct vm_area_struct *old, struct vm_area_struct *a, str… in reusable_anon_vma() argument 1104 struct anon_vma *anon_vma = READ_ONCE(old->anon_vma); in reusable_anon_vma() 1106 if (anon_vma && list_is_singular(&old->anon_vma_chain)) in reusable_anon_vma()
|
D | hugetlb.c | 5016 struct folio *new_folio, pte_t old, unsigned long sz) in hugetlb_install_folio() argument 5022 if (userfaultfd_wp(vma) && huge_pte_uffd_wp(old)) in hugetlb_install_folio()
|