| /mm/ |
| A D | gup.c | 109 if (flags & FOLL_PIN) { in gup_put_folio() 146 unsigned int flags) in try_grab_folio() argument 154 if (flags & FOLL_GET) in try_grab_folio() 544 if (flags & FOLL_GET) in try_grab_folio_fast() 653 unsigned int flags) in can_follow_write_pud() argument 699 unsigned int flags) in can_follow_write_pmd() argument 800 unsigned int flags) in can_follow_write_pte() argument 1695 if (flags & FOLL_PIN) in __get_user_pages_locked() 1708 flags |= FOLL_GET; in __get_user_pages_locked() 2484 unsigned int flags; in __gup_longterm_locked() local [all …]
|
| A D | kmemleak.c | 471 unsigned long flags; in mem_pool_alloc() local 505 unsigned long flags; in mem_pool_free() local 571 unsigned long flags; in __find_and_get_object() local 626 unsigned long flags; in find_and_remove_object() local 764 unsigned long flags; in __create_object() local 804 unsigned long flags; in __delete_object() local 904 unsigned long flags; in paint_it() local 951 unsigned long flags; in reset_checksum() local 973 unsigned long flags; in add_scan_area() local 1027 unsigned long flags; in object_set_excess_ref() local [all …]
|
| A D | slub.c | 245 gfp_t flags; member 1724 *flags = 0; in parse_slub_debug_flags() 1730 *flags = 0; in parse_slub_debug_flags() 1933 return flags; in kmem_cache_flags() 2705 flags = kmalloc_fix_flags(flags); in new_slab() 4301 flags = kmalloc_fix_flags(flags); in ___kmalloc_large_node() 5024 return flags; in kmalloc_gfp_adjust() 5722 slab_flags_t flags = s->flags; in calculate_sizes() local 6403 s->flags = kmem_cache_flags(flags, s->name); in do_kmem_cache_create() 7154 WRITE_ONCE(s->flags, s->flags | SLAB_FAILSLAB); in failslab_store() [all …]
|
| A D | slab_common.c | 186 flags = kmem_cache_flags(flags, name); in find_mergeable() 202 if ((flags & SLAB_MERGE_SAME) != (s->flags & SLAB_MERGE_SAME)) in find_mergeable() 1022 invalid_mask, &invalid_mask, flags, &flags); in kmalloc_fix_flags() 1025 return flags; in kmalloc_fix_flags() 1481 unsigned long flags; in drain_page_cache() local 1505 unsigned long flags; in kvfree_rcu_bulk() local 1566 unsigned long flags; in kfree_rcu_work() local 1661 unsigned long flags; in schedule_delayed_monitor_work() local 1674 unsigned long flags; in kvfree_rcu_drain_ready() local 1712 unsigned long flags; in kvfree_rcu_queue_batch() local [all …]
|
| A D | balloon_compaction.c | 45 unsigned long flags; in balloon_page_list_enqueue() local 48 spin_lock_irqsave(&b_dev_info->pages_lock, flags); in balloon_page_list_enqueue() 81 unsigned long flags; in balloon_page_list_dequeue() local 84 spin_lock_irqsave(&b_dev_info->pages_lock, flags); in balloon_page_list_dequeue() 146 unsigned long flags; in balloon_page_enqueue() local 148 spin_lock_irqsave(&b_dev_info->pages_lock, flags); in balloon_page_enqueue() 176 unsigned long flags; in balloon_page_dequeue() local 207 unsigned long flags; in balloon_page_isolate() local 212 spin_lock_irqsave(&b_dev_info->pages_lock, flags); in balloon_page_isolate() 223 unsigned long flags; in balloon_page_putback() local [all …]
|
| A D | slab.h | 53 unsigned long flags; member 102 SLAB_MATCH(flags, flags); 239 slab_flags_t flags; member 403 gfp_t kmalloc_fix_flags(gfp_t flags); 408 slab_flags_t flags); 426 return (s->flags & SLAB_KMALLOC); in is_kmalloc_cache() 504 return s->flags & flags; in kmem_cache_debug_flags() 575 if (s->flags & SLAB_KASAN) in slab_ksize() 620 return flags & __GFP_ZERO; in slab_want_init_on_alloc() 623 return flags & __GFP_ZERO; in slab_want_init_on_alloc() [all …]
|
| A D | mmap.c | 362 flags |= MAP_FIXED; in do_mmap() 364 if (!(flags & MAP_FIXED)) in do_mmap() 415 if (flags & MAP_LOCKED) in do_mmap() 446 if (flags & ~flags_mask) in do_mmap() 623 unsigned long flags; member 700 if (flags & MAP_FIXED) in generic_get_unmapped_area() 750 if (flags & MAP_FIXED) in generic_get_unmapped_area_topdown() 780 info.flags = 0; in generic_get_unmapped_area_topdown() 872 pgoff, flags, 0); in mm_get_unmapped_area() 1132 flags &= MAP_NONBLOCK; in SYSCALL_DEFINE5() [all …]
|
| A D | memory-failure.c | 746 int flags; member 876 int flags) in kill_accessing_process() argument 1561 if (flags & MF_UNPOISON) in get_hwpoison_page() 1755 kill_procs(to_kill, flags & MF_MUST_KILL, pfn, flags); in unmap_and_kill() 2098 flags |= MF_NO_RETRY; in try_memory_failure_hugetlb() 2132 page_flags = folio->flags; in try_memory_failure_hugetlb() 2393 page_flags = folio->flags; in memory_failure() 2444 int flags; member 2479 .flags = flags, in memory_failure_queue() 2790 put_ref_page(pfn, flags); in soft_offline_page() [all …]
|
| A D | mempool.c | 316 unsigned long flags; in mempool_resize() local 321 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 340 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 358 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 391 unsigned long flags; in mempool_alloc_noprof() local 410 spin_lock_irqsave(&pool->lock, flags); in mempool_alloc_noprof() 473 unsigned long flags; in mempool_alloc_preallocated() local 475 spin_lock_irqsave(&pool->lock, flags); in mempool_alloc_preallocated() 504 unsigned long flags; in mempool_free() local 543 spin_lock_irqsave(&pool->lock, flags); in mempool_free() [all …]
|
| A D | mempolicy.c | 477 policy->flags = flags; in mpol_new() 638 unsigned long flags = qp->flags; in queue_folio_required() local 681 unsigned long flags = qp->flags; in queue_folios_pte_range() local 766 unsigned long flags = qp->flags; in queue_folios_hugetlb() local 840 unsigned long flags = qp->flags; in queue_pages_test_walk() local 912 .flags = flags, in queue_pages_range() 1080 if (flags & in do_get_mempolicy() 2780 if (a->flags != b->flags) in __mpol_equal() 3350 if (flags) in mpol_parse_str() 3458 if (flags) in mpol_parse_str() [all …]
|
| A D | memfd.c | 326 *flags |= MFD_NOEXEC_SEAL; in check_sysctl_memfd_noexec() 328 *flags |= MFD_EXEC; in check_sysctl_memfd_noexec() 385 unsigned int flags = *flags_ptr; in sanitize_flags() local 387 if (!(flags & MFD_HUGETLB)) { in sanitize_flags() 398 if ((flags & MFD_EXEC) && (flags & MFD_NOEXEC_SEAL)) in sanitize_flags() 437 if (flags & MFD_HUGETLB) { in alloc_file() 440 (flags >> MFD_HUGE_SHIFT) & in alloc_file() 450 if (flags & MFD_NOEXEC_SEAL) { in alloc_file() 471 unsigned int, flags) in SYSCALL_DEFINE2() argument 477 error = sanitize_flags(&flags); in SYSCALL_DEFINE2() [all …]
|
| A D | hugetlb_vmemmap.c | 48 unsigned long flags; member 321 unsigned long flags) in vmemmap_remap_free() argument 328 .flags = flags, in vmemmap_remap_free() 377 .flags = 0, in vmemmap_remap_free() 431 .flags = flags, in vmemmap_remap_alloc() 466 if (flags & VMEMMAP_SYNCHRONIZE_RCU) in __hugetlb_vmemmap_restore_folio() 530 flags &= ~VMEMMAP_SYNCHRONIZE_RCU; in hugetlb_vmemmap_restore_folios() 566 unsigned long flags) in __hugetlb_vmemmap_optimize_folio() argument 606 vmemmap_pages, flags); in __hugetlb_vmemmap_optimize_folio() 713 flags &= ~VMEMMAP_SYNCHRONIZE_RCU; in __hugetlb_vmemmap_optimize_folios() [all …]
|
| A D | highmem.c | 154 #define lock_kmap_any(flags) spin_lock_irqsave(&kmap_lock, flags) argument 155 #define unlock_kmap_any(flags) spin_unlock_irqrestore(&kmap_lock, flags) argument 159 #define lock_kmap_any(flags) \ argument 339 unsigned long vaddr, flags; in kmap_high_get() local 341 lock_kmap_any(flags); in kmap_high_get() 347 unlock_kmap_any(flags); in kmap_high_get() 363 unsigned long flags; in kunmap_high() local 368 lock_kmap_any(flags); in kunmap_high() 395 unlock_kmap_any(flags); in kunmap_high() 755 unsigned long flags; in page_address() local [all …]
|
| A D | memblock.c | 548 this->flags != next->flags) { in memblock_merge_regions() 586 rgn->flags = flags; in memblock_insert_region() 626 type->regions[0].flags = flags; in memblock_add_range() 668 WARN_ON(flags != MEMBLOCK_NONE && flags != rgn->flags); in memblock_add_range() 676 flags); in memblock_add_range() 1010 r->flags |= flag; in memblock_setclr_flag() 1012 r->flags &= ~flag; in memblock_setclr_flag() 1557 flags); in memblock_alloc_range_nid() 1564 flags); in memblock_alloc_range_nid() 2131 flags = rgn->flags; in memblock_dump() [all …]
|
| A D | swap.c | 90 unsigned long flags; in page_cache_release() local 92 __page_cache_release(folio, &lruvec, &flags); in page_cache_release() 94 unlock_page_lruvec_irqrestore(lruvec, flags); in page_cache_release() 162 unsigned long flags = 0; in folio_batch_move_lru() local 174 unlock_page_lruvec_irqrestore(lruvec, flags); in folio_batch_move_lru() 182 unsigned long flags; in __folio_batch_add_and_move() local 652 unsigned long flags; in lru_add_drain_cpu() local 950 unsigned long flags = 0; in folios_put_refs() local 961 unlock_page_lruvec_irqrestore(lruvec, flags); in folios_put_refs() 982 __page_cache_release(folio, &lruvec, &flags); in folios_put_refs() [all …]
|
| A D | mmzone.c | 99 unsigned long old_flags, flags; in folio_xchg_last_cpupid() local 102 old_flags = READ_ONCE(folio->flags); in folio_xchg_last_cpupid() 104 flags = old_flags; in folio_xchg_last_cpupid() 105 last_cpupid = (flags >> LAST_CPUPID_PGSHIFT) & LAST_CPUPID_MASK; in folio_xchg_last_cpupid() 107 flags &= ~(LAST_CPUPID_MASK << LAST_CPUPID_PGSHIFT); in folio_xchg_last_cpupid() 108 flags |= (cpupid & LAST_CPUPID_MASK) << LAST_CPUPID_PGSHIFT; in folio_xchg_last_cpupid() 109 } while (unlikely(!try_cmpxchg(&folio->flags, &old_flags, flags))); in folio_xchg_last_cpupid()
|
| A D | mlock.c | 515 vm_flags_t flags) in apply_vma_lock_flags() argument 546 newflags |= flags; in apply_vma_lock_flags() 668 if (flags & ~MLOCK_ONFAULT) in SYSCALL_DEFINE3() 671 if (flags & MLOCK_ONFAULT) in SYSCALL_DEFINE3() 711 if (flags & MCL_FUTURE) { in apply_mlockall_flags() 714 if (flags & MCL_ONFAULT) in apply_mlockall_flags() 717 if (!(flags & MCL_CURRENT)) in apply_mlockall_flags() 721 if (flags & MCL_CURRENT) { in apply_mlockall_flags() 723 if (flags & MCL_ONFAULT) in apply_mlockall_flags() 750 if (!flags || (flags & ~(MCL_CURRENT | MCL_FUTURE | MCL_ONFAULT)) || in SYSCALL_DEFINE1() [all …]
|
| A D | process_vm_access.c | 85 unsigned int flags = 0; in process_vm_rw_single_vec() local 93 flags |= FOLL_WRITE; in process_vm_rw_single_vec() 107 flags, process_pages, in process_vm_rw_single_vec() 154 unsigned long flags, int vm_write) in process_vm_rw_core() argument 259 unsigned long flags, int vm_write) in process_vm_rw() argument 269 if (flags != 0) in process_vm_rw() 284 rc = process_vm_rw_core(pid, &iter, iov_r, riovcnt, flags, vm_write); in process_vm_rw() 294 unsigned long, riovcnt, unsigned long, flags) in SYSCALL_DEFINE6() argument 296 return process_vm_rw(pid, lvec, liovcnt, rvec, riovcnt, flags, 0); in SYSCALL_DEFINE6() 302 unsigned long, riovcnt, unsigned long, flags) in SYSCALL_DEFINE6() argument [all …]
|
| A D | swapfile.c | 478 ci->flags = new_flags; in move_cluster() 521 VM_BUG_ON(!ci->flags); in isolate_lock_cluster() 676 VM_BUG_ON(ci->flags); in inc_cluster_info_page() 1011 si->flags &= ~SWP_WRITEOK; in del_from_avail_list() 1046 si->flags |= SWP_WRITEOK; in add_to_avail_list() 1137 if (si->flags & SWP_BLKDEV) in swap_range_free() 2820 p->flags = 0; in SYSCALL_DEFINE1() 3018 p->flags = SWP_USED; in alloc_swap_info() 3043 si->flags |= SWP_BLKDEV; in claim_swapfile() 3501 si->flags = 0; in SYSCALL_DEFINE2() [all …]
|
| A D | oom_kill.c | 167 if (p->flags & PF_KTHREAD) in oom_unkillable_task() 221 test_bit(MMF_OOM_SKIP, &p->mm->flags) || in oom_badness() 527 set_bit(MMF_UNSTABLE, &mm->flags); in __oom_reap_task_mm() 622 test_bit(MMF_OOM_SKIP, &mm->flags)) in oom_reap_task() 637 set_bit(MMF_OOM_SKIP, &mm->flags); in oom_reap_task() 670 unsigned long flags; in wake_oom_reaper() local 858 if (sig->flags & SIGNAL_GROUP_EXIT) in __task_will_free_mem() 895 if (test_bit(MMF_OOM_SKIP, &mm->flags)) in task_will_free_mem() 980 set_bit(MMF_OOM_SKIP, &mm->flags); in __oom_kill_process() 990 if (unlikely(p->flags & PF_KTHREAD)) in __oom_kill_process() [all …]
|
| A D | page_alloc.c | 108 #define pcp_trylock_prepare(flags) local_irq_save(flags) argument 109 #define pcp_trylock_finish(flags) local_irq_restore(flags) argument 447 unsigned long flags; in get_pfnblock_migratetype() local 476 flags <<= bitidx; in __set_pfnblock_flags_mask() 555 unsigned long flags; in init_pageblock_migratetype() local 561 flags = migratetype; in init_pageblock_migratetype() 1062 if (unlikely(page->flags & flags)) { in page_bad_reason() 1447 unsigned long flags; in free_pcppages_bulk() local 1534 unsigned long flags; in free_one_page() local 2528 unsigned long flags; in rmqueue_bulk() local [all …]
|
| A D | page_isolation.c | 159 unsigned long flags; in set_migratetype_isolate() local 165 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 173 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 192 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 196 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 200 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 215 unsigned long flags; in unset_migratetype_isolate() local 221 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 272 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate() 600 unsigned long pfn, flags; in test_pages_isolated() local [all …]
|
| /mm/kasan/ |
| A D | common.c | 320 gfp_t flags, bool init) in unpoison_slab_object() argument 330 kasan_save_alloc_info(cache, object, flags); in unpoison_slab_object() 334 void *object, gfp_t flags, bool init) in __kasan_slab_alloc() argument 339 if (gfpflags_allow_blocking(flags)) in __kasan_slab_alloc() 393 size_t size, gfp_t flags) in __kasan_kmalloc() argument 395 if (gfpflags_allow_blocking(flags)) in __kasan_kmalloc() 413 gfp_t flags) in poison_kmalloc_large_redzone() argument 434 gfp_t flags) in __kasan_kmalloc_large() argument 436 if (gfpflags_allow_blocking(flags)) in __kasan_kmalloc_large() 443 poison_kmalloc_large_redzone(ptr, size, flags); in __kasan_kmalloc_large() [all …]
|
| A D | quarantine.c | 187 unsigned long flags; in kasan_quarantine_put() local 207 local_irq_save(flags); in kasan_quarantine_put() 211 local_irq_restore(flags); in kasan_quarantine_put() 234 local_irq_restore(flags); in kasan_quarantine_put() 242 unsigned long flags; in kasan_quarantine_reduce() local 317 unsigned long flags; in __per_cpu_remove_cache() local 321 raw_spin_lock_irqsave(&sq->lock, flags); in __per_cpu_remove_cache() 323 raw_spin_unlock_irqrestore(&sq->lock, flags); in __per_cpu_remove_cache() 344 unsigned long flags, i; in kasan_quarantine_remove_cache() local 360 raw_spin_lock_irqsave(&sq->lock, flags); in kasan_quarantine_remove_cache() [all …]
|
| /mm/kfence/ |
| A D | core.c | 335 unsigned long flags; in check_canary_byte() local 343 raw_spin_lock_irqsave(&meta->lock, flags); in check_canary_byte() 419 unsigned long flags; in kfence_guarded_alloc() local 517 unsigned long flags; in kfence_guarded_free() local 520 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free() 758 unsigned long flags; in show_object() local 1006 unsigned long flags; in kfence_shutdown_cache() local 1087 if ((flags & GFP_ZONEMASK) || in __kfence_alloc() 1098 if (s->flags & SLAB_SKIP_KFENCE) in __kfence_alloc() 1180 unsigned long flags; in __kfence_free() local [all …]
|