/mm/ |
D | vmpressure.c | 73 static struct vmpressure *work_to_vmpressure(struct work_struct *work) in work_to_vmpressure() argument 75 return container_of(work, struct vmpressure, work); in work_to_vmpressure() 167 static void vmpressure_work_fn(struct work_struct *work) in vmpressure_work_fn() argument 169 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() 265 schedule_work(&vmpr->work); in vmpressure() 412 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init() 428 flush_work(&vmpr->work); in vmpressure_cleanup()
|
D | zswap.c | 126 struct work_struct work; member 670 static void __zswap_pool_release(struct work_struct *work) in __zswap_pool_release() argument 672 struct zswap_pool *pool = container_of(work, typeof(*pool), work); in __zswap_pool_release() 695 INIT_WORK(&pool->work, __zswap_pool_release); in __zswap_pool_empty() 696 schedule_work(&pool->work); in __zswap_pool_empty()
|
D | swap.c | 700 struct work_struct *work = &per_cpu(lru_add_drain_work, cpu); in lru_add_drain_all() local 707 INIT_WORK(work, lru_add_drain_per_cpu); in lru_add_drain_all() 708 queue_work_on(cpu, lru_add_drain_wq, work); in lru_add_drain_all()
|
D | memory-failure.c | 1305 struct work_struct work; member 1340 schedule_work_on(smp_processor_id(), &mf_cpu->work); in memory_failure_queue() 1349 static void memory_failure_work_func(struct work_struct *work) in memory_failure_work_func() argument 1379 INIT_WORK(&mf_cpu->work, memory_failure_work_func); in memory_failure_init()
|
D | memcontrol.c | 1693 struct work_struct work; member 1810 drain_local_stock(&stock->work); in drain_all_stock() 1812 schedule_work_on(cpu, &stock->work); in drain_all_stock() 1850 static void high_work_func(struct work_struct *work) in high_work_func() argument 1854 memcg = container_of(work, struct mem_cgroup, high_work); in high_work_func() 2155 struct work_struct work; member 2163 container_of(w, struct memcg_kmem_cache_create_work, work); in memcg_kmem_cache_create_func() 2189 INIT_WORK(&cw->work, memcg_kmem_cache_create_func); in __memcg_schedule_kmem_cache_create() 2191 queue_work(memcg_kmem_cache_create_wq, &cw->work); in __memcg_schedule_kmem_cache_create() 3713 static void memcg_event_remove(struct work_struct *work) in memcg_event_remove() argument [all …]
|
D | backing-dev.c | 477 static void cgwb_release_workfn(struct work_struct *work) in cgwb_release_workfn() argument 479 struct bdi_writeback *wb = container_of(work, struct bdi_writeback, in cgwb_release_workfn()
|
D | percpu.c | 184 static void pcpu_balance_workfn(struct work_struct *work); 1109 static void pcpu_balance_workfn(struct work_struct *work) in pcpu_balance_workfn() argument
|
D | slab.c | 559 if (keventd_up() && reap_work->work.func == NULL) { in start_cpu_timer() 1128 per_cpu(slab_reap_work, cpu).work.func = NULL; in slab_offline_cpu() 4051 struct delayed_work *work = to_delayed_work(w); in cache_reap() local 4099 schedule_delayed_work_on(smp_processor_id(), work, in cache_reap()
|
D | zsmalloc.c | 2184 static void async_free_zspage(struct work_struct *work) in async_free_zspage() argument 2192 struct zs_pool *pool = container_of(work, struct zs_pool, in async_free_zspage()
|
D | swapfile.c | 351 static void swap_discard_work(struct work_struct *work) in swap_discard_work() argument 355 si = container_of(work, struct swap_info_struct, discard_work); in swap_discard_work()
|
D | vmstat.c | 1560 static void refresh_vm_stats(struct work_struct *work) in refresh_vm_stats() argument
|
D | kmemleak.c | 1840 static void kmemleak_do_cleanup(struct work_struct *work) in kmemleak_do_cleanup() argument
|
D | Kconfig | 438 benefit but it will work automatically for all applications.
|