/mm/ |
D | vmpressure.c | 70 static struct vmpressure *work_to_vmpressure(struct work_struct *work) in work_to_vmpressure() argument 72 return container_of(work, struct vmpressure, work); in work_to_vmpressure() 181 static void vmpressure_work_fn(struct work_struct *work) in vmpressure_work_fn() argument 183 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() 278 schedule_work(&vmpr->work); in vmpressure() 454 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init() 470 flush_work(&vmpr->work); in vmpressure_cleanup()
|
D | z3fold.c | 117 struct work_struct work; member 166 struct work_struct work; member 328 INIT_WORK(&zhdr->work, compact_page_work); in init_z3fold_page() 457 queue_work(pool->release_wq, &pool->work); in __release_z3fold_page() 492 struct z3fold_pool *pool = container_of(w, struct z3fold_pool, work); in free_pages_work() 504 cancel_work_sync(&zhdr->work); in free_pages_work() 649 work); in compact_page_work() 802 INIT_WORK(&pool->work, free_pages_work); in z3fold_create_pool() 925 cancel_work_sync(&zhdr->work); in z3fold_alloc() 1072 queue_work_on(zhdr->cpu, pool->compact_wq, &zhdr->work); in z3fold_free() [all …]
|
D | slab_common.c | 45 static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work); 568 static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work) in slab_caches_to_rcu_destroy_workfn() argument 709 static void kmemcg_workfn(struct work_struct *work) in kmemcg_workfn() argument 711 struct kmem_cache *s = container_of(work, struct kmem_cache, in kmemcg_workfn() 712 memcg_params.work); in kmemcg_workfn() 735 INIT_WORK(&s->memcg_params.work, kmemcg_workfn); in kmemcg_rcufn() 736 queue_work(memcg_kmem_cache_wq, &s->memcg_params.work); in kmemcg_rcufn() 755 INIT_WORK(&s->memcg_params.work, kmemcg_workfn); in kmemcg_cache_shutdown() 756 queue_work(memcg_kmem_cache_wq, &s->memcg_params.work); in kmemcg_cache_shutdown()
|
D | zswap.c | 126 struct work_struct work; member 625 static void __zswap_pool_release(struct work_struct *work) in __zswap_pool_release() argument 627 struct zswap_pool *pool = container_of(work, typeof(*pool), work); in __zswap_pool_release() 650 INIT_WORK(&pool->work, __zswap_pool_release); in __zswap_pool_empty() 651 schedule_work(&pool->work); in __zswap_pool_empty()
|
D | swap.c | 726 struct work_struct *work = &per_cpu(lru_add_drain_work, cpu); in lru_add_drain_all() local 734 INIT_WORK(work, lru_add_drain_per_cpu); in lru_add_drain_all() 735 queue_work_on(cpu, mm_percpu_wq, work); in lru_add_drain_all()
|
D | memory-failure.c | 1441 struct work_struct work; member 1474 schedule_work_on(smp_processor_id(), &mf_cpu->work); in memory_failure_queue() 1483 static void memory_failure_work_func(struct work_struct *work) in memory_failure_work_func() argument 1513 INIT_WORK(&mf_cpu->work, memory_failure_work_func); in memory_failure_init()
|
D | memcontrol.c | 2170 struct work_struct work; member 2308 drain_local_stock(&stock->work); in drain_all_stock() 2310 schedule_work_on(cpu, &stock->work); in drain_all_stock() 2377 static void high_work_func(struct work_struct *work) in high_work_func() argument 2381 memcg = container_of(work, struct mem_cgroup, high_work); in high_work_func() 2881 struct work_struct work; member 2887 container_of(w, struct memcg_kmem_cache_create_work, work); in memcg_kmem_cache_create_func() 2916 INIT_WORK(&cw->work, memcg_kmem_cache_create_func); in memcg_schedule_kmem_cache_create() 2918 queue_work(memcg_kmem_cache_wq, &cw->work); in memcg_schedule_kmem_cache_create() 4638 static void memcg_event_remove(struct work_struct *work) in memcg_event_remove() argument [all …]
|
D | backing-dev.c | 481 static void cgwb_release_workfn(struct work_struct *work) in cgwb_release_workfn() argument 483 struct bdi_writeback *wb = container_of(work, struct bdi_writeback, in cgwb_release_workfn()
|
D | slab.h | 91 struct work_struct work; member
|
D | page_alloc.c | 105 struct work_struct work; member 2929 static void drain_local_pages_wq(struct work_struct *work) in drain_local_pages_wq() argument 2933 drain = container_of(work, struct pcpu_drain, work); in drain_local_pages_wq() 3017 INIT_WORK(&drain->work, drain_local_pages_wq); in drain_all_pages() 3018 queue_work_on(cpu, mm_percpu_wq, &drain->work); in drain_all_pages() 3021 flush_work(&per_cpu_ptr(&pcpu_drain, cpu)->work); in drain_all_pages()
|
D | slab.c | 514 if (reap_work->work.func == NULL) { in start_cpu_timer() 1083 per_cpu(slab_reap_work, cpu).work.func = NULL; in slab_offline_cpu() 3984 struct delayed_work *work = to_delayed_work(w); in cache_reap() local 4032 schedule_delayed_work_on(smp_processor_id(), work, in cache_reap()
|
D | zsmalloc.c | 2172 static void async_free_zspage(struct work_struct *work) in async_free_zspage() argument 2180 struct zs_pool *pool = container_of(work, struct zs_pool, in async_free_zspage()
|
D | percpu.c | 187 static void pcpu_balance_workfn(struct work_struct *work); 1832 static void pcpu_balance_workfn(struct work_struct *work) in pcpu_balance_workfn() argument
|
D | kmemleak.c | 1868 static void kmemleak_do_cleanup(struct work_struct *work) in kmemleak_do_cleanup() argument
|
D | swapfile.c | 505 static void swap_discard_work(struct work_struct *work) in swap_discard_work() argument 509 si = container_of(work, struct swap_info_struct, discard_work); in swap_discard_work()
|
D | vmstat.c | 1752 static void refresh_vm_stats(struct work_struct *work) in refresh_vm_stats() argument
|
D | Kconfig | 397 benefit but it will work automatically for all applications.
|
D | slub.c | 5777 static void sysfs_slab_remove_workfn(struct work_struct *work) in sysfs_slab_remove_workfn() argument 5780 container_of(work, struct kmem_cache, kobj_remove_work); in sysfs_slab_remove_workfn()
|
D | hugetlb.c | 1359 static void free_hpage_workfn(struct work_struct *work) in free_hpage_workfn() argument
|