Lines Matching refs:flags

335 	unsigned long flags;  in check_canary_byte()  local
343 raw_spin_lock_irqsave(&meta->lock, flags); in check_canary_byte()
345 raw_spin_unlock_irqrestore(&meta->lock, flags); in check_canary_byte()
419 unsigned long flags; in kfence_guarded_alloc() local
427 raw_spin_lock_irqsave(&kfence_freelist_lock, flags); in kfence_guarded_alloc()
432 raw_spin_unlock_irqrestore(&kfence_freelist_lock, flags); in kfence_guarded_alloc()
438 if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) { in kfence_guarded_alloc()
448 raw_spin_lock_irqsave(&kfence_freelist_lock, flags); in kfence_guarded_alloc()
451 raw_spin_unlock_irqrestore(&kfence_freelist_lock, flags); in kfence_guarded_alloc()
483 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc()
517 unsigned long flags; in kfence_guarded_free() local
520 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free()
527 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free()
549 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free()
570 raw_spin_lock_irqsave(&kfence_freelist_lock, flags); in kfence_guarded_free()
573 raw_spin_unlock_irqrestore(&kfence_freelist_lock, flags); in kfence_guarded_free()
758 unsigned long flags; in show_object() local
760 raw_spin_lock_irqsave(&meta->lock, flags); in show_object()
762 raw_spin_unlock_irqrestore(&meta->lock, flags); in show_object()
1006 unsigned long flags; in kfence_shutdown_cache() local
1029 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_shutdown_cache()
1031 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_shutdown_cache()
1059 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_shutdown_cache()
1062 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_shutdown_cache()
1066 void *__kfence_alloc(struct kmem_cache *s, size_t size, gfp_t flags) in __kfence_alloc() argument
1087 if ((flags & GFP_ZONEMASK) || in __kfence_alloc()
1088 ((flags & __GFP_THISNODE) && num_online_nodes() > 1) || in __kfence_alloc()
1089 (s->flags & (SLAB_CACHE_DMA | SLAB_CACHE_DMA32))) { in __kfence_alloc()
1098 if (s->flags & SLAB_SKIP_KFENCE) in __kfence_alloc()
1138 return kfence_guarded_alloc(s, size, flags, stack_entries, num_stack_entries, in __kfence_alloc()
1179 if (unlikely(meta->cache && (meta->cache->flags & SLAB_TYPESAFE_BY_RCU))) { in __kfence_free()
1180 unsigned long flags; in __kfence_free() local
1182 raw_spin_lock_irqsave(&meta->lock, flags); in __kfence_free()
1184 raw_spin_unlock_irqrestore(&meta->lock, flags); in __kfence_free()
1196 unsigned long flags; in kfence_handle_page_fault() local
1228 raw_spin_lock_irqsave(&to_report->lock, flags); in kfence_handle_page_fault()
1242 raw_spin_lock_irqsave(&to_report->lock, flags); in kfence_handle_page_fault()
1255 raw_spin_unlock_irqrestore(&to_report->lock, flags); in kfence_handle_page_fault()