| /mm/ |
| A D | mmap_lock.c | 26 void __mmap_lock_do_trace_start_locking(struct mm_struct *mm, bool write) in __mmap_lock_do_trace_start_locking() argument 28 trace_mmap_lock_start_locking(mm, write); in __mmap_lock_do_trace_start_locking() 32 void __mmap_lock_do_trace_acquire_returned(struct mm_struct *mm, bool write, in __mmap_lock_do_trace_acquire_returned() argument 35 trace_mmap_lock_acquire_returned(mm, write, success); in __mmap_lock_do_trace_acquire_returned() 39 void __mmap_lock_do_trace_released(struct mm_struct *mm, bool write) in __mmap_lock_do_trace_released() argument 41 trace_mmap_lock_released(mm, write); in __mmap_lock_do_trace_released()
|
| A D | util.c | 760 static int overcommit_ratio_handler(const struct ctl_table *table, int write, in overcommit_ratio_handler() argument 765 ret = proc_dointvec(table, write, buffer, lenp, ppos); in overcommit_ratio_handler() 766 if (ret == 0 && write) in overcommit_ratio_handler() 776 static int overcommit_policy_handler(const struct ctl_table *table, int write, in overcommit_policy_handler() argument 794 if (write) { in overcommit_policy_handler() 797 ret = proc_dointvec_minmax(&t, write, buffer, lenp, ppos); in overcommit_policy_handler() 806 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in overcommit_policy_handler() 812 static int overcommit_kbytes_handler(const struct ctl_table *table, int write, in overcommit_kbytes_handler() argument 817 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in overcommit_kbytes_handler() 818 if (ret == 0 && write) in overcommit_kbytes_handler()
|
| A D | memcontrol-v1.c | 2019 .write = mem_cgroup_reset, 2025 .write = mem_cgroup_write, 2031 .write = mem_cgroup_write, 2037 .write = mem_cgroup_reset, 2086 .write = mem_cgroup_write, 2097 .write = mem_cgroup_reset, 2103 .write = mem_cgroup_reset, 2115 .write = mem_cgroup_write, 2126 .write = mem_cgroup_reset, 2132 .write = mem_cgroup_reset, [all …]
|
| A D | hugetlb_cgroup.c | 716 .write = hugetlb_cgroup_write_dfl, 723 .write = hugetlb_cgroup_write_dfl, 763 .write = hugetlb_cgroup_write_legacy, 769 .write = hugetlb_cgroup_write_legacy, 784 .write = hugetlb_cgroup_reset, 790 .write = hugetlb_cgroup_reset, 796 .write = hugetlb_cgroup_reset, 802 .write = hugetlb_cgroup_reset,
|
| A D | huge_memory.c | 1395 if (write) { in insert_pmd() 1419 if (write) { in insert_pmd() 1445 bool write) in vmf_insert_pfn_pmd() argument 1490 bool write) in vmf_insert_folio_pmd() argument 1544 if (write) { in insert_pud() 1565 if (write) { in insert_pud() 1626 bool write) in vmf_insert_folio_pud() argument 1659 if (write) in touch_pmd() 1796 if (write) in touch_pud() 3004 if (write) in __split_huge_pmd_locked() [all …]
|
| A D | page-writeback.c | 492 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler() 493 if (ret == 0 && write) in dirty_background_ratio_handler() 504 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler() 505 if (ret == 0 && write) { in dirty_background_bytes_handler() 522 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler() 523 if (ret == 0 && write && vm_dirty_ratio != old_ratio) { in dirty_ratio_handler() 530 static int dirty_bytes_handler(const struct ctl_table *table, int write, in dirty_bytes_handler() argument 536 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_bytes_handler() 537 if (ret == 0 && write && vm_dirty_bytes != old_bytes) { in dirty_bytes_handler() 2184 ret = proc_dointvec(table, write, buffer, length, ppos); in dirty_writeback_centisecs_handler() [all …]
|
| A D | compaction.c | 2946 static int compaction_proactiveness_sysctl_handler(const struct ctl_table *table, int write, in compaction_proactiveness_sysctl_handler() argument 2951 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in compaction_proactiveness_sysctl_handler() 2955 if (write && sysctl_compaction_proactiveness) { in compaction_proactiveness_sysctl_handler() 2976 static int sysctl_compaction_handler(const struct ctl_table *table, int write, in sysctl_compaction_handler() argument 2981 ret = proc_dointvec(table, write, buffer, length, ppos); in sysctl_compaction_handler() 2988 if (write) in sysctl_compaction_handler() 3268 int write, void *buffer, size_t *lenp, loff_t *ppos) in proc_dointvec_minmax_warn_RT_change() argument 3272 if (!IS_ENABLED(CONFIG_PREEMPT_RT) || !write) in proc_dointvec_minmax_warn_RT_change() 3273 return proc_dointvec_minmax(table, write, buffer, lenp, ppos); in proc_dointvec_minmax_warn_RT_change() 3276 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in proc_dointvec_minmax_warn_RT_change()
|
| A D | page_alloc.c | 5478 if (write) in numa_zonelist_order_handler() 5480 return proc_dostring(table, write, buffer, length, ppos); in numa_zonelist_order_handler() 6452 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in postcore_initcall() 6456 if (write) { in postcore_initcall() 6468 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in watermark_scale_factor_sysctl_handler() 6472 if (write) in watermark_scale_factor_sysctl_handler() 6498 rc = proc_dointvec_minmax(table, write, buffer, length, ppos); in sysctl_min_unmapped_ratio_sysctl_handler() 6545 int write, void *buffer, size_t *length, loff_t *ppos) in lowmem_reserve_ratio_sysctl_handler() argument 6549 proc_dointvec_minmax(table, write, buffer, length, ppos); in lowmem_reserve_ratio_sysctl_handler() 6566 int write, void *buffer, size_t *length, loff_t *ppos) in percpu_pagelist_high_fraction_sysctl_handler() argument [all …]
|
| A D | gup.c | 1221 int write = (gup_flags & FOLL_WRITE); in check_vma_flags() local 1240 if (write) { in check_vma_flags() 1274 if (!arch_vma_access_permitted(vma, write, false, foreign)) in check_vma_flags() 1535 bool write = !!(fault_flags & FAULT_FLAG_WRITE); in vma_permits_fault() local 1537 vm_flags_t vm_flags = write ? VM_WRITE : VM_READ; in vma_permits_fault() 1549 if (!arch_vma_access_permitted(vma, write, false, foreign)) in vma_permits_fault() 1908 unsigned long end, bool write, int *locked) in faultin_page_range() argument 1929 if (write) in faultin_page_range()
|
| A D | vmstat.c | 79 static int sysctl_vm_numa_stat_handler(const struct ctl_table *table, int write, in sysctl_vm_numa_stat_handler() argument 85 if (write) in sysctl_vm_numa_stat_handler() 87 ret = proc_dointvec_minmax(table, write, buffer, length, ppos); in sysctl_vm_numa_stat_handler() 88 if (ret || !write) in sysctl_vm_numa_stat_handler() 1979 static int vmstat_refresh(const struct ctl_table *table, int write, in vmstat_refresh() argument 2030 if (write) in vmstat_refresh()
|
| A D | memcontrol.c | 4594 .write = memory_peak_write, 4600 .write = memory_min_write, 4606 .write = memory_low_write, 4612 .write = memory_high_write, 4618 .write = memory_max_write, 4651 .write = memory_reclaim, 5348 .write = swap_high_write, 5354 .write = swap_max_write, 5362 .write = swap_peak_write, 5550 .write = zswap_max_write, [all …]
|
| A D | hugetlb.c | 5095 return proc_doulongvec_minmax(&dup_table, write, buffer, length, ppos); in proc_hugetlb_doulongvec_minmax() 5099 const struct ctl_table *table, int write, in hugetlb_sysctl_handler_common() argument 5109 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos, in hugetlb_sysctl_handler_common() 5114 if (write) in hugetlb_sysctl_handler_common() 5121 static int hugetlb_sysctl_handler(const struct ctl_table *table, int write, in hugetlb_sysctl_handler() argument 5125 return hugetlb_sysctl_handler_common(false, table, write, in hugetlb_sysctl_handler() 5133 return hugetlb_sysctl_handler_common(true, table, write, in hugetlb_mempolicy_sysctl_handler() 5138 static int hugetlb_overcommit_handler(const struct ctl_table *table, int write, in hugetlb_overcommit_handler() argument 5150 if (write && hstate_is_gigantic(h)) in hugetlb_overcommit_handler() 5153 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos, in hugetlb_overcommit_handler() [all …]
|
| A D | internal.h | 963 unsigned long end, bool write, int *locked); 1408 pud_t *pud, bool write); 1410 pmd_t *pmd, bool write);
|
| A D | memory.c | 2640 bool write) in vmf_insert_page_mkwrite() argument 5198 bool write = vmf->flags & FAULT_FLAG_WRITE; in do_set_pmd() local 5245 if (write) in do_set_pmd() 5287 bool write = vmf->flags & FAULT_FLAG_WRITE; in set_pte_range() local 5299 if (write) in set_pte_range() 6633 void *buf, int len, int write) in generic_access_phys() argument 6651 if ((write & FOLL_WRITE) && !writable) in generic_access_phys() 6669 if (write) in generic_access_phys() 6690 int write = gup_flags & FOLL_WRITE; in __access_remote_vm() local 6733 len, write); in __access_remote_vm() [all …]
|
| A D | shrinker_debug.c | 159 .write = shrinker_debugfs_scan_write,
|
| A D | nommu.c | 1644 int write = gup_flags & FOLL_WRITE; in __access_remote_vm() local 1657 if (write && vma->vm_flags & VM_MAYWRITE) in __access_remote_vm() 1660 else if (!write && vma->vm_flags & VM_MAYREAD) in __access_remote_vm()
|
| A D | mmap.c | 1699 unsigned long addr, bool write) in mmap_read_lock_maybe_expand() argument 1701 if (!write || addr >= new_vma->vm_start) { in mmap_read_lock_maybe_expand()
|
| A D | madvise.c | 971 const bool write = madv_behavior->behavior == MADV_POPULATE_WRITE; in madvise_populate() local 979 pages = faultin_page_range(mm, start, end, write, &locked); in madvise_populate()
|
| A D | Kconfig | 1274 Arch has userfaultfd write protection support 1290 bool "Userfaultfd write protection support for shmem/hugetlbfs" 1295 Allows to create marker PTEs for userfaultfd write protection 1296 purposes. It is required to enable userfaultfd write protection on
|
| A D | Kconfig.debug | 128 Check that anonymous page is not being mapped twice with read write
|
| A D | kmemleak.c | 2179 .write = kmemleak_write,
|
| A D | vmscan.c | 5677 .write = lru_gen_seq_write,
|
| /mm/kasan/ |
| A D | sw_tags.c | 74 bool kasan_check_range(const void *addr, size_t size, bool write, in kasan_check_range() argument 85 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range() 111 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range() 116 return !kasan_report(addr, size, write, ret_ip); in kasan_check_range()
|
| A D | generic.c | 165 size_t size, bool write, in check_region_inline() argument 175 return !kasan_report(addr, size, write, ret_ip); in check_region_inline() 178 return !kasan_report(addr, size, write, ret_ip); in check_region_inline() 183 return !kasan_report(addr, size, write, ret_ip); in check_region_inline() 186 bool kasan_check_range(const void *addr, size_t size, bool write, in kasan_check_range() argument 189 return check_region_inline(addr, size, write, ret_ip); in kasan_check_range()
|
| A D | kasan.h | 346 bool kasan_check_range(const void *addr, size_t size, bool write,
|