Home
last modified time | relevance | path

Searched refs:tlb (Results 1 – 11 of 11) sorted by relevance

/include/asm-generic/
A Dtlb.h299 #define tlb_delay_rmap(tlb) (((tlb)->delayed_rmap = 1), true) argument
386 tlb->start = min(tlb->start, address); in __tlb_adjust_range()
387 tlb->end = max(tlb->end, address + range_size); in __tlb_adjust_range()
393 tlb->start = tlb->end = ~0; in __tlb_reset_range()
426 if (tlb->end) in tlb_flush()
440 if (tlb->fullmm || tlb->need_flush_all) { in tlb_flush()
449 flush_tlb_range(&vma, tlb->start, tlb->end); in tlb_flush()
486 if (!(tlb->freed_tables || tlb->cleared_ptes || tlb->cleared_pmds || in tlb_flush_mmu_tlbonly()
487 tlb->cleared_puds || tlb->cleared_p4ds)) in tlb_flush_mmu_tlbonly()
515 if (tlb->page_size && tlb->page_size != page_size) { in tlb_change_page_size()
[all …]
A Dpgtable-nop4d.h52 #define p4d_free_tlb(tlb, x, a) do { } while (0) argument
A Dpgtable-nopud.h60 #define pud_free_tlb(tlb, x, a) do { } while (0) argument
A Dpgtable-nopmd.h66 #define pmd_free_tlb(tlb, x, a) do { } while (0) argument
/include/linux/
A Dio-pgtable.h113 const struct iommu_flush_ops *tlb; member
275 if (iop->cfg.tlb && iop->cfg.tlb->tlb_flush_all) in io_pgtable_tlb_flush_all()
276 iop->cfg.tlb->tlb_flush_all(iop->cookie); in io_pgtable_tlb_flush_all()
283 if (iop->cfg.tlb && iop->cfg.tlb->tlb_flush_walk) in io_pgtable_tlb_flush_walk()
284 iop->cfg.tlb->tlb_flush_walk(iova, size, granule, iop->cookie); in io_pgtable_tlb_flush_walk()
292 if (iop->cfg.tlb && iop->cfg.tlb->tlb_add_page) in io_pgtable_tlb_add_page()
293 iop->cfg.tlb->tlb_add_page(gather, iova, granule, iop->cookie); in io_pgtable_tlb_add_page()
A Dhuge_mm.h28 bool madvise_free_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma,
30 int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma, pmd_t *pmd,
32 int zap_huge_pud(struct mmu_gather *tlb, struct vm_area_struct *vma, pud_t *pud,
36 int change_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma,
417 int change_huge_pud(struct mmu_gather *tlb, struct vm_area_struct *vma,
422 change_huge_pud(struct mmu_gather *tlb, struct vm_area_struct *vma, in change_huge_pud() argument
681 static inline int change_huge_pud(struct mmu_gather *tlb, in change_huge_pud() argument
A Dswiotlb.h40 int (*remap)(void *tlb, unsigned long nslabs));
42 int (*remap)(void *tlb, unsigned long nslabs));
A Dmm_types.h1440 extern void tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm);
1441 extern void tlb_gather_mmu_fullmm(struct mmu_gather *tlb, struct mm_struct *mm);
1442 extern void tlb_finish_mmu(struct mmu_gather *tlb);
A Dhugetlb.h134 void __unmap_hugepage_range(struct mmu_gather *tlb,
445 static inline void __unmap_hugepage_range(struct mmu_gather *tlb, in __unmap_hugepage_range() argument
A Dmm.h2364 void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
2370 void free_pgd_range(struct mmu_gather *tlb, unsigned long addr,
2566 extern long change_protection(struct mmu_gather *tlb,
2569 extern int mprotect_fixup(struct vma_iterator *vmi, struct mmu_gather *tlb,
/include/trace/events/
A Dtlb.h3 #define TRACE_SYSTEM tlb

Completed in 32 milliseconds