Searched refs:tlb_lock (Results 1 – 3 of 3) sorted by relevance
348 mutex_lock(&l2->tlb_lock); in kvmhv_enter_nested_guest()350 mutex_unlock(&l2->tlb_lock); in kvmhv_enter_nested_guest()587 mutex_lock(&gp->tlb_lock); in kvmhv_copy_tofrom_guest_nested()618 mutex_unlock(&gp->tlb_lock); in kvmhv_copy_tofrom_guest_nested()696 mutex_init(&gp->tlb_lock); in kvmhv_alloc_nested()1115 mutex_lock(&gp->tlb_lock); in kvmhv_emulate_tlbie_tlb_addr()1125 mutex_unlock(&gp->tlb_lock); in kvmhv_emulate_tlbie_tlb_addr()1135 mutex_lock(&gp->tlb_lock); in kvmhv_emulate_tlbie_lpid()1158 mutex_unlock(&gp->tlb_lock); in kvmhv_emulate_tlbie_lpid()1659 mutex_lock(&gp->tlb_lock); in kvmhv_nested_page_fault()[all …]
215 spinlock_t tlb_lock; /* lock for tlb range flush */ member340 spin_lock_irqsave(&bank->tlb_lock, flags); in mtk_iommu_tlb_flush_all()344 spin_unlock_irqrestore(&bank->tlb_lock, flags); in mtk_iommu_tlb_flush_all()385 spin_lock_irqsave(&curbank->tlb_lock, flags); in mtk_iommu_tlb_flush_range_sync()400 spin_unlock_irqrestore(&curbank->tlb_lock, flags); in mtk_iommu_tlb_flush_range_sync()1252 spin_lock_init(&bank->tlb_lock); in mtk_iommu_probe()
31 struct mutex tlb_lock; /* serialize page faults and tlbies */ member
Completed in 10 milliseconds