| /mm/ |
| A D | mempool.c | 199 spin_lock_init(&pool->lock); in mempool_init_node() 321 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 327 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 332 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 340 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 358 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 368 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 410 spin_lock_irqsave(&pool->lock, flags); in mempool_alloc_noprof() 475 spin_lock_irqsave(&pool->lock, flags); in mempool_alloc_preallocated() 543 spin_lock_irqsave(&pool->lock, flags); in mempool_free() [all …]
|
| A D | mmu_notifier.c | 44 spinlock_t lock; member 103 spin_lock(&subscriptions->lock); in mn_itree_inv_start_range() 114 spin_unlock(&subscriptions->lock); in mn_itree_inv_start_range() 136 spin_lock(&subscriptions->lock); in mn_itree_inv_end() 139 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 163 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 234 spin_lock(&subscriptions->lock); in mmu_interval_read_begin() 238 spin_unlock(&subscriptions->lock); in mmu_interval_read_begin() 323 spin_lock(&subscriptions->lock); in mn_hlist_release() 928 spin_lock(&subscriptions->lock); in __mmu_interval_notifier_insert() [all …]
|
| A D | zsmalloc.c | 161 spinlock_t lock; member 217 rwlock_t lock; member 255 spinlock_t lock; member 309 spin_lock(&zsl->lock); in zspage_read_lock() 311 spin_unlock(&zsl->lock); in zspage_read_lock() 322 spin_lock(&zsl->lock); in zspage_read_unlock() 324 spin_unlock(&zsl->lock); in zspage_read_unlock() 331 spin_lock(&zsl->lock); in zspage_write_trylock() 339 spin_unlock(&zsl->lock); in zspage_write_trylock() 350 spin_unlock(&zsl->lock); in zspage_write_unlock() [all …]
|
| A D | kmemleak.c | 136 raw_spinlock_t lock; member 669 raw_spin_lock_init(&object->lock); in __alloc_object() 1469 raw_spin_unlock(&object->lock); in pointer_update_refs() 1480 raw_spin_unlock(&object->lock); in pointer_update_refs() 1695 raw_spin_lock_irq(&object->lock); in kmemleak_scan() 1723 raw_spin_unlock_irq(&object->lock); in kmemleak_scan() 1805 raw_spin_lock_irq(&object->lock); in kmemleak_scan() 1812 raw_spin_unlock_irq(&object->lock); in kmemleak_scan() 1842 raw_spin_lock_irq(&object->lock); in kmemleak_scan() 1852 raw_spin_unlock_irq(&object->lock); in kmemleak_scan() [all …]
|
| A D | list_lru.c | 66 spin_lock_irq(&l->lock); in lock_list_lru() 68 spin_lock(&l->lock); in lock_list_lru() 71 spin_unlock_irq(&l->lock); in lock_list_lru() 73 spin_unlock(&l->lock); in lock_list_lru() 108 spin_unlock_irq(&l->lock); in unlock_list_lru() 110 spin_unlock(&l->lock); in unlock_list_lru() 144 spin_lock_irq(&l->lock); in lock_list_lru_of_memcg() 146 spin_lock(&l->lock); in lock_list_lru_of_memcg() 156 spin_unlock(&l->lock); in unlock_list_lru() 396 spin_lock_init(&l->lock); in init_one_lru() [all …]
|
| A D | swap.c | 55 local_lock_t lock; member 69 .lock = INIT_LOCAL_LOCK(lock), 192 local_lock(&cpu_fbatches.lock); in __folio_batch_add_and_move() 201 local_unlock(&cpu_fbatches.lock); in __folio_batch_add_and_move() 361 local_lock(&cpu_fbatches.lock); in __lru_cache_activate_folio() 734 local_lock(&cpu_fbatches.lock); in lru_add_drain() 748 local_lock(&cpu_fbatches.lock); in lru_add_and_bh_lrus_drain() 757 local_lock(&cpu_fbatches.lock); in lru_add_drain_cpu_zone() 809 static DEFINE_MUTEX(lock); in __lru_add_drain_all() 835 mutex_lock(&lock); in __lru_add_drain_all() [all …]
|
| A D | page_isolation.c | 165 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 173 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 192 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 196 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 200 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 221 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 272 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate() 633 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated() 635 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
|
| A D | vmalloc.c | 896 spinlock_t lock; member 1890 spin_lock(lock); in preload_this_cpu_lock() 2548 spinlock_t lock; member 2560 spinlock_t lock; member 2713 spin_lock(&vbq->lock); in new_vmap_block() 2758 spin_lock(&vbq->lock); in purge_fragmented_block() 2790 spin_lock(&vb->lock); in purge_fragmented_blocks() 2833 spin_lock(&vb->lock); in vb_alloc() 2880 spin_lock(&vb->lock); in vb_free() 2889 spin_lock(&vb->lock); in vb_free() [all …]
|
| A D | swapfile.c | 121 local_lock_t lock; member 448 spin_lock(&ci->lock); in lock_cluster() 455 spin_unlock(&ci->lock); in unlock_cluster() 467 spin_lock(&si->lock); in move_cluster() 511 spin_lock(&si->lock); in isolate_lock_cluster() 548 spin_lock(&si->lock); in swap_do_scheduled_discard() 562 spin_lock(&ci->lock); in swap_do_scheduled_discard() 571 spin_lock(&si->lock); in swap_do_scheduled_discard() 705 spin_lock(&ci->lock); in cluster_reclaim_range() 2722 spin_lock(&p->lock); in SYSCALL_DEFINE1() [all …]
|
| A D | mlock.c | 32 local_lock_t lock; member 37 .lock = INIT_LOCAL_LOCK(lock), 216 local_lock(&mlock_fbatch.lock); in mlock_drain_local() 220 local_unlock(&mlock_fbatch.lock); in mlock_drain_local() 246 local_lock(&mlock_fbatch.lock); in mlock_folio() 260 local_unlock(&mlock_fbatch.lock); in mlock_folio() 272 local_lock(&mlock_fbatch.lock); in mlock_new_folio() 283 local_unlock(&mlock_fbatch.lock); in mlock_new_folio() 294 local_lock(&mlock_fbatch.lock); in munlock_folio() 304 local_unlock(&mlock_fbatch.lock); in munlock_folio()
|
| A D | dmapool.c | 50 spinlock_t lock; member 265 spin_lock_init(&retval->lock); in dma_pool_create_node() 416 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 423 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 429 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 433 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 458 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free() 463 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free()
|
| A D | page_reporting.c | 164 spin_lock_irq(&zone->lock); in page_reporting_cycle() 222 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 234 spin_lock_irq(&zone->lock); in page_reporting_cycle() 254 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 299 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 301 spin_unlock_irq(&zone->lock); in page_reporting_process_zone()
|
| A D | slab_common.c | 1408 raw_spinlock_t lock; member 1422 .lock = __RAW_SPIN_LOCK_UNLOCKED(krc.lock), 1443 raw_spin_lock(&krcp->lock); in krc_this_cpu_lock() 1451 raw_spin_unlock_irqrestore(&krcp->lock, flags); in krc_this_cpu_unlock() 1488 raw_spin_lock_irqsave(&krcp->lock, flags); in drain_page_cache() 1528 raw_spin_lock_irqsave(&krcp->lock, flags); in kvfree_rcu_bulk() 1579 raw_spin_lock_irqsave(&krcp->lock, flags); in kfree_rcu_work() 1663 raw_spin_lock_irqsave(&krcp->lock, flags); in schedule_delayed_monitor_work() 1677 raw_spin_lock_irqsave(&krcp->lock, flags); in kvfree_rcu_drain_ready() 1716 raw_spin_lock_irqsave(&krcp->lock, flags); in kvfree_rcu_queue_batch() [all …]
|
| A D | shuffle.c | 88 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 141 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone() 143 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 146 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
|
| A D | page_alloc.c | 167 pcpu_spin_unlock(lock, ptr) 2585 spin_lock(&pcp->lock); in decay_pcp_high() 2587 spin_unlock(&pcp->lock); in decay_pcp_high() 2607 spin_lock(&pcp->lock); in drain_zone_pages() 2609 spin_unlock(&pcp->lock); in drain_zone_pages() 2623 spin_lock(&pcp->lock); in drain_pages_zone() 2632 spin_unlock(&pcp->lock); in drain_pages_zone() 6046 spin_lock(&pcp->lock); in zone_pcp_update_cacheinfo() 6051 spin_unlock(&pcp->lock); in zone_pcp_update_cacheinfo() 6372 spin_lock(&lock); in setup_per_zone_wmarks() [all …]
|
| A D | highmem.c | 739 spinlock_t lock; /* Protect this bucket's list */ member 764 spin_lock_irqsave(&pas->lock, flags); in page_address() 776 spin_unlock_irqrestore(&pas->lock, flags); in page_address() 800 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 802 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 804 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 811 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 821 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
|
| A D | cma_debug.c | 38 spin_lock_irq(&cma->lock); in cma_used_get() 40 spin_unlock_irq(&cma->lock); in cma_used_get() 55 spin_lock_irq(&cma->lock); in cma_maxchunk_get() 62 spin_unlock_irq(&cma->lock); in cma_maxchunk_get()
|
| A D | cma.c | 90 spin_lock_irqsave(&cma->lock, flags); in cma_clear_bitmap() 93 spin_unlock_irqrestore(&cma->lock, flags); in cma_clear_bitmap() 171 spin_lock_init(&cma->lock); in cma_activate_area() 760 spin_lock_irq(&cma->lock); in cma_debug_show_areas() 776 spin_unlock_irq(&cma->lock); in cma_debug_show_areas() 799 spin_lock_irq(&cma->lock); in cma_range_alloc() 805 spin_unlock_irq(&cma->lock); in cma_range_alloc() 812 spin_unlock_irq(&cma->lock); in cma_range_alloc() 822 spin_unlock_irq(&cma->lock); in cma_range_alloc()
|
| /mm/kfence/ |
| A D | core.c | 295 lockdep_assert_held(&meta->lock); in metadata_update_state() 343 raw_spin_lock_irqsave(&meta->lock, flags); in check_canary_byte() 345 raw_spin_unlock_irqrestore(&meta->lock, flags); in check_canary_byte() 483 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc() 520 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free() 549 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 646 raw_spin_lock_init(&meta->lock); in kfence_init_pool() 760 raw_spin_lock_irqsave(&meta->lock, flags); in show_object() 1029 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_shutdown_cache() 1059 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_shutdown_cache() [all …]
|
| A D | kfence_test.c | 43 spinlock_t lock; member 47 .lock = __SPIN_LOCK_UNLOCKED(observed.lock), 56 spin_lock_irqsave(&observed.lock, flags); in probe_console() 73 spin_unlock_irqrestore(&observed.lock, flags); in probe_console() 168 spin_lock_irqsave(&observed.lock, flags); in report_matches() 175 spin_unlock_irqrestore(&observed.lock, flags); in report_matches() 828 spin_lock_irqsave(&observed.lock, flags); in test_init() 832 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
|
| A D | report.c | 141 lockdep_assert_held(&meta->lock); in kfence_print_object() 211 lockdep_assert_held(&meta->lock); in kfence_report_error() 317 raw_spin_lock_irqsave(&meta->lock, flags); in __kfence_obj_info() 328 raw_spin_unlock_irqrestore(&meta->lock, flags); in __kfence_obj_info()
|
| /mm/kmsan/ |
| A D | kmsan_test.c | 32 spinlock_t lock; member 37 .lock = __SPIN_LOCK_UNLOCKED(observed.lock), 47 spin_lock_irqsave(&observed.lock, flags); in probe_console() 61 spin_unlock_irqrestore(&observed.lock, flags); in probe_console() 75 spin_lock_irqsave(&observed.lock, flags); in report_reset() 78 spin_unlock_irqrestore(&observed.lock, flags); in report_reset() 118 spin_lock_irqsave(&observed.lock, flags); in report_matches() 125 spin_unlock_irqrestore(&observed.lock, flags); in report_matches() 693 spin_lock_irqsave(&observed.lock, flags); in test_init() 697 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
|
| /mm/kasan/ |
| A D | quarantine.c | 105 raw_spinlock_t lock; member 110 .lock = __RAW_SPIN_LOCK_UNLOCKED(shrink_qlist.lock), 321 raw_spin_lock_irqsave(&sq->lock, flags); in __per_cpu_remove_cache() 323 raw_spin_unlock_irqrestore(&sq->lock, flags); in __per_cpu_remove_cache() 360 raw_spin_lock_irqsave(&sq->lock, flags); in kasan_quarantine_remove_cache() 362 raw_spin_unlock_irqrestore(&sq->lock, flags); in kasan_quarantine_remove_cache()
|
| A D | tags.c | 42 .lock = __RW_LOCK_UNLOCKED(stack_ring.lock) 113 read_lock_irqsave(&stack_ring.lock, flags); in save_stack_info() 134 read_unlock_irqrestore(&stack_ring.lock, flags); in save_stack_info()
|
| A D | report_tags.c | 42 write_lock_irqsave(&stack_ring.lock, flags); in kasan_complete_mode_report_info() 102 write_unlock_irqrestore(&stack_ring.lock, flags); in kasan_complete_mode_report_info()
|