Lines Matching refs:vm_area_struct

110 	struct vm_area_struct *vma;
125 void hugetlb_dup_vma_private(struct vm_area_struct *vma);
126 void clear_vma_resv_huge_pages(struct vm_area_struct *vma);
135 int move_hugetlb_page_tables(struct vm_area_struct *vma,
136 struct vm_area_struct *new_vma,
140 struct vm_area_struct *, struct vm_area_struct *);
141 struct page *hugetlb_follow_page_mask(struct vm_area_struct *vma,
143 long follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *,
144 struct page **, struct vm_area_struct **,
147 void unmap_hugepage_range(struct vm_area_struct *,
151 struct vm_area_struct *vma,
158 vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
162 struct vm_area_struct *dst_vma,
170 struct vm_area_struct *vma,
185 pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
195 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
237 int huge_pmd_unshare(struct mm_struct *mm, struct vm_area_struct *vma,
239 void adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma,
242 void hugetlb_vma_lock_read(struct vm_area_struct *vma);
243 void hugetlb_vma_unlock_read(struct vm_area_struct *vma);
244 void hugetlb_vma_lock_write(struct vm_area_struct *vma);
245 void hugetlb_vma_unlock_write(struct vm_area_struct *vma);
246 int hugetlb_vma_trylock_write(struct vm_area_struct *vma);
247 void hugetlb_vma_assert_locked(struct vm_area_struct *vma);
252 long hugetlb_change_protection(struct vm_area_struct *vma,
257 void hugetlb_unshare_all_pmds(struct vm_area_struct *vma);
261 static inline void hugetlb_dup_vma_private(struct vm_area_struct *vma) in hugetlb_dup_vma_private()
265 static inline void clear_vma_resv_huge_pages(struct vm_area_struct *vma) in clear_vma_resv_huge_pages()
281 struct vm_area_struct *vma, in huge_pmd_unshare()
288 struct vm_area_struct *vma, in adjust_range_if_pmd_sharing_possible()
293 static inline struct page *hugetlb_follow_page_mask(struct vm_area_struct *vma, in hugetlb_follow_page_mask()
300 struct vm_area_struct *vma, struct page **pages, in follow_hugetlb_page()
301 struct vm_area_struct **vmas, unsigned long *position, in follow_hugetlb_page()
311 struct vm_area_struct *dst_vma, in copy_hugetlb_page_range()
312 struct vm_area_struct *src_vma) in copy_hugetlb_page_range()
318 static inline int move_hugetlb_page_tables(struct vm_area_struct *vma, in move_hugetlb_page_tables()
319 struct vm_area_struct *new_vma, in move_hugetlb_page_tables()
347 static inline void hugetlb_vma_lock_read(struct vm_area_struct *vma) in hugetlb_vma_lock_read()
351 static inline void hugetlb_vma_unlock_read(struct vm_area_struct *vma) in hugetlb_vma_unlock_read()
355 static inline void hugetlb_vma_lock_write(struct vm_area_struct *vma) in hugetlb_vma_lock_write()
359 static inline void hugetlb_vma_unlock_write(struct vm_area_struct *vma) in hugetlb_vma_unlock_write()
363 static inline int hugetlb_vma_trylock_write(struct vm_area_struct *vma) in hugetlb_vma_trylock_write()
368 static inline void hugetlb_vma_assert_locked(struct vm_area_struct *vma) in hugetlb_vma_assert_locked()
398 struct vm_area_struct *dst_vma, in hugetlb_mcopy_atomic_pte()
442 struct vm_area_struct *vma, unsigned long address, in hugetlb_change_protection()
450 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range_final()
458 struct vm_area_struct *vma, unsigned long address, in hugetlb_fault()
465 static inline void hugetlb_unshare_all_pmds(struct vm_area_struct *vma) { } in hugetlb_unshare_all_pmds()
720 struct folio *alloc_hugetlb_folio(struct vm_area_struct *vma,
724 struct folio *alloc_hugetlb_folio_vma(struct hstate *h, struct vm_area_struct *vma,
728 void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
790 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
800 extern unsigned long vma_kernel_pagesize(struct vm_area_struct *vma);
802 extern unsigned long vma_mmu_pagesize(struct vm_area_struct *vma);
998 static inline pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma, in huge_ptep_modify_prot_start()
1007 static inline void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, in huge_ptep_modify_prot_commit()
1039 static inline struct folio *alloc_hugetlb_folio(struct vm_area_struct *vma,
1054 struct vm_area_struct *vma,
1075 static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
1105 static inline unsigned long vma_kernel_pagesize(struct vm_area_struct *vma)
1110 static inline unsigned long vma_mmu_pagesize(struct vm_area_struct *vma)
1194 static inline pte_t huge_ptep_clear_flush(struct vm_area_struct *vma,
1244 bool want_pmd_share(struct vm_area_struct *vma, unsigned long addr);
1254 static inline bool __vma_shareable_lock(struct vm_area_struct *vma) in __vma_shareable_lock()
1264 hugetlb_walk(struct vm_area_struct *vma, unsigned long addr, unsigned long sz) in hugetlb_walk()