| /mm/kasan/ |
| A D | shadow.c | 214 unsigned long action, void *data) in kasan_mem_notifier() argument 216 struct memory_notify *mem_data = data; in kasan_mem_notifier() 311 index = PFN_DOWN(addr - data->start); in kasan_populate_vmalloc_pte() 312 page = data->pages[index]; in kasan_populate_vmalloc_pte() 319 data->pages[index] = NULL; in kasan_populate_vmalloc_pte() 359 struct vmalloc_populate_data data; in __kasan_populate_vmalloc() local 363 if (!data.pages) in __kasan_populate_vmalloc() 372 data.start = start; in __kasan_populate_vmalloc() 374 kasan_populate_vmalloc_pte, &data); in __kasan_populate_vmalloc() 375 ___free_pages_bulk(data.pages, nr_pages); in __kasan_populate_vmalloc() [all …]
|
| /mm/ |
| A D | cma_debug.c | 24 static int cma_debugfs_get(void *data, u64 *val) in cma_debugfs_get() argument 26 unsigned long *p = data; in cma_debugfs_get() 34 static int cma_used_get(void *data, u64 *val) in cma_used_get() argument 36 struct cma *cma = data; in cma_used_get() 46 static int cma_maxchunk_get(void *data, u64 *val) in cma_maxchunk_get() argument 48 struct cma *cma = data; in cma_maxchunk_get() 120 static int cma_free_write(void *data, u64 val) in cma_free_write() argument 123 struct cma *cma = data; in cma_free_write() 152 static int cma_alloc_write(void *data, u64 val) in cma_alloc_write() argument 155 struct cma *cma = data; in cma_alloc_write()
|
| A D | page_io.c | 177 unsigned long *data; in is_folio_zero_filled() local 180 last_pos = PAGE_SIZE / sizeof(*data) - 1; in is_folio_zero_filled() 182 data = kmap_local_folio(folio, i * PAGE_SIZE); in is_folio_zero_filled() 188 if (data[last_pos]) { in is_folio_zero_filled() 189 kunmap_local(data); in is_folio_zero_filled() 193 if (data[pos]) { in is_folio_zero_filled() 194 kunmap_local(data); in is_folio_zero_filled() 198 kunmap_local(data); in is_folio_zero_filled()
|
| A D | percpu-km.c | 74 chunk->data = pages; in pcpu_create_chunk() 97 if (chunk->data) in pcpu_destroy_chunk() 98 __free_pages(chunk->data, order_base_2(nr_pages)); in pcpu_destroy_chunk()
|
| A D | hwpoison-inject.c | 14 static int hwpoison_inject(void *data, u64 val) in hwpoison_inject() argument 56 static int hwpoison_unpoison(void *data, u64 val) in hwpoison_unpoison() argument
|
| A D | percpu-vm.c | 349 chunk->data = vms; in pcpu_create_chunk() 366 if (chunk->data) in pcpu_destroy_chunk() 367 pcpu_free_vm_areas(chunk->data, pcpu_nr_groups); in pcpu_destroy_chunk()
|
| A D | util.c | 796 t.data = &new_policy; in overcommit_policy_handler() 826 .data = &sysctl_overcommit_memory, 835 .data = &sysctl_overcommit_ratio, 842 .data = &sysctl_overcommit_kbytes, 849 .data = &sysctl_user_reserve_kbytes, 856 .data = &sysctl_admin_reserve_kbytes,
|
| A D | compaction.c | 1799 struct compact_control *cc = (struct compact_control *)data; in compaction_alloc_noprof() 1845 static struct folio *compaction_alloc(struct folio *src, unsigned long data) in compaction_alloc() argument 1847 return alloc_hooks(compaction_alloc_noprof(src, data)); in compaction_alloc() 1855 static void compaction_free(struct folio *dst, unsigned long data) in compaction_free() argument 1857 struct compact_control *cc = (struct compact_control *)data; in compaction_free() 3275 old = *(int *)table->data; in proc_dointvec_minmax_warn_RT_change() 3279 if (old != *(int *)table->data) in proc_dointvec_minmax_warn_RT_change() 3289 .data = &sysctl_compact_memory, 3296 .data = &sysctl_compaction_proactiveness, 3305 .data = &sysctl_extfrag_threshold, [all …]
|
| A D | dmapool.c | 99 u8 *data = (void *)block; in pool_check_block() local 103 if (data[i] == POOL_POISON_FREED) in pool_check_block() 113 data, pool->size, 1); in pool_check_block()
|
| A D | page-writeback.c | 2272 .data = &dirty_background_ratio, 2281 .data = &dirty_background_bytes, 2289 .data = &vm_dirty_ratio, 2298 .data = &vm_dirty_bytes, 2306 .data = &dirty_writeback_interval, 2313 .data = &dirty_expire_interval, 2322 .data = &vm_highmem_is_dirtyable, 2332 .data = &laptop_mode, 2606 void *data) in write_cache_pages() argument 2612 error = writepage(folio, wbc, data); in write_cache_pages()
|
| A D | oom_kill.c | 711 .data = &sysctl_panic_on_oom, 720 .data = &sysctl_oom_kill_allocating_task, 727 .data = &sysctl_oom_dump_tasks,
|
| A D | mmap.c | 1510 .data = &sysctl_max_map_count, 1520 .data = &sysctl_legacy_va_layout, 1530 .data = &mmap_rnd_bits, 1541 .data = &mmap_rnd_compat_bits, 1628 unsigned long action, void *data) in reserve_mem_notifier() argument
|
| A D | memremap.c | 132 static void devm_memremap_pages_release(void *data) in devm_memremap_pages_release() argument 134 memunmap_pages(data); in devm_memremap_pages_release()
|
| A D | memory.c | 130 .data = &randomize_va_space, 2989 pte_fn_t fn, void *data, bool create, in apply_to_pte_range() argument 3015 err = fn(pte, addr, data); in apply_to_pte_range() 3032 pte_fn_t fn, void *data, bool create, in apply_to_pmd_range() argument 3060 fn, data, create, mask); in apply_to_pmd_range() 3070 pte_fn_t fn, void *data, bool create, in apply_to_pud_range() argument 3096 fn, data, create, mask); in apply_to_pud_range() 3106 pte_fn_t fn, void *data, bool create, in apply_to_p4d_range() argument 3132 fn, data, create, mask); in apply_to_p4d_range() 3142 void *data, bool create) in __apply_to_page_range() argument [all …]
|
| A D | cma.c | 1019 int cma_for_each_area(int (*it)(struct cma *cma, void *data), void *data) in cma_for_each_area() argument 1024 int ret = it(&cma_areas[i], data); in cma_for_each_area()
|
| A D | percpu-internal.h | 70 void *data; /* chunk data */ member
|
| A D | page_alloc.c | 5681 static void __build_all_zonelists(void *data) in __build_all_zonelists() argument 5685 pg_data_t *self = data; in __build_all_zonelists() 6601 .data = &min_free_kbytes, 6609 .data = &watermark_boost_factor, 6617 .data = &watermark_scale_factor, 6626 .data = &defrag_mode, 6635 .data = &percpu_pagelist_high_fraction, 6643 .data = &sysctl_lowmem_reserve_ratio, 6651 .data = &numa_zonelist_order, 6658 .data = &sysctl_min_unmapped_ratio, [all …]
|
| A D | page_owner.c | 939 static int page_owner_threshold_get(void *data, u64 *val) in page_owner_threshold_get() argument 945 static int page_owner_threshold_set(void *data, u64 val) in page_owner_threshold_set() argument
|
| A D | slub.c | 852 (*(int *)resource->data)++; in slab_add_kunit_errors() 1223 static void restore_bytes(struct kmem_cache *s, const char *message, u8 data, in restore_bytes() argument 1226 slab_fix(s, "Restoring %s 0x%p-0x%p=0x%x", message, from, to - 1, data); in restore_bytes() 1227 memset(from, data, to - from); in restore_bytes() 7210 int *data = kmalloc_array(nr_cpu_ids, sizeof(int), GFP_KERNEL); in show_stat() local 7212 if (!data) in show_stat() 7218 data[cpu] = x; in show_stat() 7226 if (data[cpu]) in show_stat() 7228 cpu, data[cpu]); in show_stat() 7231 kfree(data); in show_stat() [all …]
|
| /mm/damon/ |
| A D | sysfs.c | 1377 struct damon_sysfs_kdamond *kdamond = data; in damon_sysfs_upd_schemes_stats() 1418 static int damon_sysfs_commit_input(void *data) in damon_sysfs_commit_input() argument 1420 struct damon_sysfs_kdamond *kdamond = data; in damon_sysfs_commit_input() 1447 struct damon_sysfs_kdamond *sysfs_kdamond = data; in damon_sysfs_commit_schemes_quota_goals() 1474 struct damon_sysfs_kdamond *kdamond = data; in damon_sysfs_upd_schemes_effective_quotas() 1484 struct damon_sysfs_kdamond *kdamond = data; in damon_sysfs_upd_tuned_intervals() 1512 static int damon_sysfs_repeat_call_fn(void *data) in damon_sysfs_repeat_call_fn() argument 1514 struct damon_sysfs_kdamond *sysfs_kdamond = data; in damon_sysfs_repeat_call_fn() 1563 damon_sysfs_repeat_call_control.data = kdamond; in damon_sysfs_turn_damon_on() 1588 call_control.data = kdamond; in damon_sysfs_damon_call() [all …]
|
| A D | Kconfig | 33 This builds the default data access monitoring operations for DAMON 42 This builds the default data access monitoring operations for DAMON 63 the interface for arbitrary data access monitoring. 98 bool "Build data access monitoring stat (DAMON_STAT)"
|
| A D | stat.c | 125 static int damon_stat_damon_call_fn(void *data) in damon_stat_damon_call_fn() argument 127 struct damon_ctx *c = data; in damon_stat_damon_call_fn() 207 call_control.data = damon_stat_context; in damon_stat_start()
|
| A D | reclaim.c | 335 call_control.data = ctx; in damon_reclaim_init()
|
| A D | lru_sort.c | 331 call_control.data = ctx; in damon_lru_sort_init()
|
| A D | core.c | 1254 static int kdamond_fn(void *data); 1748 control->walk_fn(control->data, ctx, t, r, s, sz_filter_passed); in damos_walk_call_walk() 2459 ret = control->fn(control->data); in kdamond_call() 2531 static int kdamond_fn(void *data) in kdamond_fn() argument 2533 struct damon_ctx *ctx = data; in kdamond_fn()
|