/mm/kasan/ |
D | kasan_test_module.c | 94 static noinline void __init kasan_workqueue_work(struct work_struct *work) in kasan_workqueue_work() argument 96 kfree(work); in kasan_workqueue_work() 102 struct work_struct *work; in kasan_workqueue_uaf() local 109 work = kmalloc(sizeof(struct work_struct), GFP_KERNEL); in kasan_workqueue_uaf() 110 if (!work) { in kasan_workqueue_uaf() 115 INIT_WORK(work, kasan_workqueue_work); in kasan_workqueue_uaf() 116 queue_work(workqueue, work); in kasan_workqueue_uaf() 120 ((volatile struct work_struct *)work)->data; in kasan_workqueue_uaf()
|
/mm/ |
D | vmpressure.c | 70 static struct vmpressure *work_to_vmpressure(struct work_struct *work) in work_to_vmpressure() argument 72 return container_of(work, struct vmpressure, work); in work_to_vmpressure() 180 static void vmpressure_work_fn(struct work_struct *work) in vmpressure_work_fn() argument 182 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() 290 schedule_work(&vmpr->work); in vmpressure() 464 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init() 480 flush_work(&vmpr->work); in vmpressure_cleanup()
|
D | page_reporting.c | 52 schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY); in __page_reporting_request() 280 static void page_reporting_process(struct work_struct *work) in page_reporting_process() argument 282 struct delayed_work *d_work = to_delayed_work(work); in page_reporting_process() 284 container_of(d_work, struct page_reporting_dev_info, work); in page_reporting_process() 319 schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY); in page_reporting_process() 345 INIT_DELAYED_WORK(&prdev->work, &page_reporting_process); in page_reporting_register() 375 cancel_delayed_work_sync(&prdev->work); in page_reporting_unregister()
|
D | z3fold.c | 119 struct work_struct work; member 171 struct work_struct work; member 411 INIT_WORK(&zhdr->work, compact_page_work); in init_z3fold_page() 522 queue_work(pool->release_wq, &pool->work); in __release_z3fold_page() 557 struct z3fold_pool *pool = container_of(w, struct z3fold_pool, work); in free_pages_work() 569 cancel_work_sync(&zhdr->work); in free_pages_work() 851 work); in compact_page_work() 1011 INIT_WORK(&pool->work, free_pages_work); in z3fold_create_pool() 1125 cancel_work_sync(&zhdr->work); in z3fold_alloc() 1273 queue_work_on(zhdr->cpu, pool->compact_wq, &zhdr->work); in z3fold_free() [all …]
|
D | backing-dev.c | 267 static void wb_update_bandwidth_workfn(struct work_struct *work) in wb_update_bandwidth_workfn() argument 269 struct bdi_writeback *wb = container_of(to_delayed_work(work), in wb_update_bandwidth_workfn() 385 static void cleanup_offline_cgwbs_workfn(struct work_struct *work); 397 static void cgwb_release_workfn(struct work_struct *work) in cgwb_release_workfn() argument 399 struct bdi_writeback *wb = container_of(work, struct bdi_writeback, in cgwb_release_workfn() 667 static void cleanup_offline_cgwbs_workfn(struct work_struct *work) in cleanup_offline_cgwbs_workfn() argument
|
D | memory-failure.c | 1926 struct work_struct work; member 1959 schedule_work_on(smp_processor_id(), &mf_cpu->work); in memory_failure_queue() 1968 static void memory_failure_work_func(struct work_struct *work) in memory_failure_work_func() argument 1975 mf_cpu = container_of(work, struct memory_failure_cpu, work); in memory_failure_work_func() 1998 cancel_work_sync(&mf_cpu->work); in memory_failure_queue_kick() 1999 memory_failure_work_func(&mf_cpu->work); in memory_failure_queue_kick() 2011 INIT_WORK(&mf_cpu->work, memory_failure_work_func); in memory_failure_init()
|
D | swap.c | 865 struct work_struct *work = &per_cpu(lru_add_drain_work, cpu); in __lru_add_drain_all() local 875 INIT_WORK(work, lru_add_drain_per_cpu); in __lru_add_drain_all() 876 queue_work_on(cpu, mm_percpu_wq, work); in __lru_add_drain_all()
|
D | slab_common.c | 50 static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work); 428 static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work) in slab_caches_to_rcu_destroy_workfn() argument
|
D | zswap.c | 711 static void __zswap_pool_release(struct work_struct *work) in __zswap_pool_release() argument 713 struct zswap_pool *pool = container_of(work, typeof(*pool), in __zswap_pool_release()
|
D | memcontrol.c | 2137 struct work_struct work; member 2331 drain_local_stock(&stock->work); in drain_all_stock() 2333 schedule_work_on(cpu, &stock->work); in drain_all_stock() 2376 static void high_work_func(struct work_struct *work) in high_work_func() argument 2380 memcg = container_of(work, struct mem_cgroup, high_work); in high_work_func() 4732 static void memcg_event_remove(struct work_struct *work) in memcg_event_remove() argument 4735 container_of(work, struct mem_cgroup_event, remove); in memcg_event_remove() 7268 INIT_WORK(&per_cpu_ptr(&memcg_stock, cpu)->work, in mem_cgroup_init()
|
D | slab.c | 515 if (reap_work->work.func == NULL) { in start_cpu_timer() 1094 per_cpu(slab_reap_work, cpu).work.func = NULL; in slab_offline_cpu() 4002 struct delayed_work *work = to_delayed_work(w); in cache_reap() local 4050 schedule_delayed_work_on(smp_processor_id(), work, in cache_reap()
|
D | slub.c | 2658 struct work_struct work; member 2674 sfw = container_of(w, struct slub_flush_work, work); in flush_cpu_slab() 2709 INIT_WORK(&sfw->work, flush_cpu_slab); in flush_all_cpus_locked() 2712 queue_work_on(cpu, flushwq, &sfw->work); in flush_all_cpus_locked() 2719 flush_work(&sfw->work); in flush_all_cpus_locked()
|
D | zsmalloc.c | 2165 static void async_free_zspage(struct work_struct *work) in async_free_zspage() argument 2173 struct zs_pool *pool = container_of(work, struct zs_pool, in async_free_zspage()
|
D | percpu.c | 200 static void pcpu_balance_workfn(struct work_struct *work); 2234 static void pcpu_balance_workfn(struct work_struct *work) in pcpu_balance_workfn() argument
|
D | kmemleak.c | 1885 static void kmemleak_do_cleanup(struct work_struct *work) in kmemleak_do_cleanup() argument
|
D | vmstat.c | 1869 static void refresh_vm_stats(struct work_struct *work) in refresh_vm_stats() argument
|
D | swapfile.c | 503 static void swap_discard_work(struct work_struct *work) in swap_discard_work() argument 507 si = container_of(work, struct swap_info_struct, discard_work); in swap_discard_work()
|
D | Kconfig | 406 benefit but it will work automatically for all applications.
|
D | hugetlb.c | 1504 static void free_hpage_workfn(struct work_struct *work) in free_hpage_workfn() argument
|
/mm/damon/ |
D | Kconfig | 33 that work for virtual address spaces.
|
D | reclaim.c | 347 static void damon_reclaim_timer_fn(struct work_struct *work) in damon_reclaim_timer_fn() argument
|
/mm/kfence/ |
D | core.c | 702 static void wake_up_kfence_timer(struct irq_work *work) in wake_up_kfence_timer() argument 721 static void toggle_allocation_gate(struct work_struct *work) in toggle_allocation_gate() argument
|