| /mm/ |
| A D | vmpressure.c | 70 static struct vmpressure *work_to_vmpressure(struct work_struct *work) in work_to_vmpressure() argument 72 return container_of(work, struct vmpressure, work); in work_to_vmpressure() 180 static void vmpressure_work_fn(struct work_struct *work) in vmpressure_work_fn() argument 182 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() 290 schedule_work(&vmpr->work); in vmpressure() 464 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init() 480 flush_work(&vmpr->work); in vmpressure_cleanup()
|
| A D | page_reporting.c | 83 schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY); in __page_reporting_request() 307 static void page_reporting_process(struct work_struct *work) in page_reporting_process() argument 309 struct delayed_work *d_work = to_delayed_work(work); in page_reporting_process() 311 container_of(d_work, struct page_reporting_dev_info, work); in page_reporting_process() 346 schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY); in page_reporting_process() 381 INIT_DELAYED_WORK(&prdev->work, &page_reporting_process); in page_reporting_register() 412 cancel_delayed_work_sync(&prdev->work); in page_reporting_unregister()
|
| A D | slab_common.c | 1564 static void kfree_rcu_work(struct work_struct *work) in kfree_rcu_work() argument 1575 krwp = container_of(to_rcu_work(work), in kfree_rcu_work() 1766 static void kfree_rcu_monitor(struct work_struct *work) in kfree_rcu_monitor() argument 1768 struct kfree_rcu_cpu *krcp = container_of(work, in kfree_rcu_monitor() 1769 struct kfree_rcu_cpu, monitor_work.work); in kfree_rcu_monitor() 1786 static void fill_page_cache_func(struct work_struct *work) in fill_page_cache_func() argument 1790 container_of(work, struct kfree_rcu_cpu, in fill_page_cache_func() 1791 page_cache_work.work); in fill_page_cache_func() 2118 kfree_rcu_monitor(&krcp->monitor_work.work); in kfree_rcu_shrink_scan()
|
| A D | backing-dev.c | 502 static void wb_update_bandwidth_workfn(struct work_struct *work) in wb_update_bandwidth_workfn() argument 504 struct bdi_writeback *wb = container_of(to_delayed_work(work), in wb_update_bandwidth_workfn() 599 static void cleanup_offline_cgwbs_workfn(struct work_struct *work); 611 static void cgwb_release_workfn(struct work_struct *work) in cgwb_release_workfn() argument 613 struct bdi_writeback *wb = container_of(work, struct bdi_writeback, in cgwb_release_workfn() 879 static void cleanup_offline_cgwbs_workfn(struct work_struct *work) in cleanup_offline_cgwbs_workfn() argument
|
| A D | execmem.c | 161 static void execmem_cache_clean(struct work_struct *work) in execmem_cache_clean() argument 377 static void execmem_cache_free_slow(struct work_struct *work); 380 static void execmem_cache_free_slow(struct work_struct *work) in execmem_cache_free_slow() argument
|
| A D | swap.c | 870 struct work_struct *work = &per_cpu(lru_add_drain_work, cpu); in __lru_add_drain_all() local 873 INIT_WORK(work, lru_add_drain_per_cpu); in __lru_add_drain_all() 874 queue_work_on(cpu, mm_percpu_wq, work); in __lru_add_drain_all()
|
| A D | memory-failure.c | 2451 struct work_struct work; member 2486 schedule_work_on(smp_processor_id(), &mf_cpu->work); in memory_failure_queue() 2495 static void memory_failure_work_func(struct work_struct *work) in memory_failure_work_func() argument 2502 mf_cpu = container_of(work, struct memory_failure_cpu, work); in memory_failure_work_func() 2525 INIT_WORK(&mf_cpu->work, memory_failure_work_func); in memory_failure_init()
|
| A D | memcontrol.c | 1754 struct work_struct work; member 1770 struct work_struct work; member 2001 drain_local_memcg_stock(&memcg_st->work); in drain_all_stock() 2003 schedule_work_on(cpu, &memcg_st->work); in drain_all_stock() 2011 drain_local_obj_stock(&obj_st->work); in drain_all_stock() 2013 schedule_work_on(cpu, &obj_st->work); in drain_all_stock() 2056 static void high_work_func(struct work_struct *work) in high_work_func() argument 2060 memcg = container_of(work, struct mem_cgroup, high_work); in high_work_func() 5110 INIT_WORK(&per_cpu_ptr(&memcg_stock, cpu)->work, in mem_cgroup_init() 5112 INIT_WORK(&per_cpu_ptr(&obj_stock, cpu)->work, in mem_cgroup_init()
|
| A D | zswap.c | 375 static void __zswap_pool_release(struct work_struct *work) in __zswap_pool_release() argument 377 struct zswap_pool *pool = container_of(work, typeof(*pool), in __zswap_pool_release()
|
| A D | swapfile.c | 577 static void swap_discard_work(struct work_struct *work) in swap_discard_work() argument 581 si = container_of(work, struct swap_info_struct, discard_work); in swap_discard_work() 868 static void swap_reclaim_work(struct work_struct *work) in swap_reclaim_work() argument 872 si = container_of(work, struct swap_info_struct, reclaim_work); in swap_reclaim_work()
|
| A D | memcontrol-v1.c | 979 static void memcg_event_remove(struct work_struct *work) in memcg_event_remove() argument 982 container_of(work, struct mem_cgroup_event, remove); in memcg_event_remove()
|
| A D | zsmalloc.c | 1833 static void async_free_zspage(struct work_struct *work) in async_free_zspage() argument 1839 struct zs_pool *pool = container_of(work, struct zs_pool, in async_free_zspage()
|
| A D | vmalloc.c | 1048 static void drain_vmap_area_work(struct work_struct *work); 2257 static void purge_vmap_node(struct work_struct *work) in purge_vmap_node() argument 2259 struct vmap_node *vn = container_of(work, in purge_vmap_node() 2387 static void drain_vmap_area_work(struct work_struct *work) in drain_vmap_area_work() argument
|
| A D | slub.c | 3348 struct work_struct work; member 3364 sfw = container_of(w, struct slub_flush_work, work); in flush_cpu_slab() 3399 INIT_WORK(&sfw->work, flush_cpu_slab); in flush_all_cpus_locked() 3402 queue_work_on(cpu, flushwq, &sfw->work); in flush_all_cpus_locked() 3409 flush_work(&sfw->work); in flush_all_cpus_locked()
|
| A D | percpu.c | 196 static void pcpu_balance_workfn(struct work_struct *work); 2186 static void pcpu_balance_workfn(struct work_struct *work) in pcpu_balance_workfn() argument
|
| A D | kmemleak.c | 2208 static void kmemleak_do_cleanup(struct work_struct *work) in kmemleak_do_cleanup() argument
|
| A D | vmstat.c | 1974 static void refresh_vm_stats(struct work_struct *work) in refresh_vm_stats() argument
|
| A D | Kconfig | 859 benefit but it will work automatically for all applications.
|
| A D | hugetlb.c | 1665 static void free_hpage_workfn(struct work_struct *work) in free_hpage_workfn() argument
|
| /mm/kasan/ |
| A D | kasan_test_c.c | 907 static void workqueue_uaf_work(struct work_struct *work) in workqueue_uaf_work() argument 909 kfree(work); in workqueue_uaf_work() 915 struct work_struct *work; in workqueue_uaf() local 920 work = kmalloc(sizeof(struct work_struct), GFP_KERNEL); in workqueue_uaf() 921 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, work); in workqueue_uaf() 923 INIT_WORK(work, workqueue_uaf_work); in workqueue_uaf() 924 queue_work(workqueue, work); in workqueue_uaf() 928 ((volatile struct work_struct *)work)->data); in workqueue_uaf()
|
| /mm/damon/ |
| A D | Kconfig | 34 that work for virtual address spaces.
|
| /mm/kfence/ |
| A D | core.c | 824 static void wake_up_kfence_timer(struct irq_work *work) in wake_up_kfence_timer() argument 842 static void toggle_allocation_gate(struct work_struct *work) in toggle_allocation_gate() argument
|