Lines Matching refs:vm_area_struct
234 struct vm_area_struct *vm_area_alloc(struct mm_struct *);
235 struct vm_area_struct *vm_area_dup(struct vm_area_struct *);
236 void vm_area_free(struct vm_area_struct *);
537 struct vm_area_struct *vma; /* Target VMA */
587 void (*open)(struct vm_area_struct * area);
592 void (*close)(struct vm_area_struct * area);
594 int (*may_split)(struct vm_area_struct *area, unsigned long addr);
595 int (*mremap)(struct vm_area_struct *area);
601 int (*mprotect)(struct vm_area_struct *vma, unsigned long start,
607 unsigned long (*pagesize)(struct vm_area_struct * area);
620 int (*access)(struct vm_area_struct *vma, unsigned long addr,
626 const char *(*name)(struct vm_area_struct *vma);
636 int (*set_policy)(struct vm_area_struct *vma, struct mempolicy *new);
648 struct mempolicy *(*get_policy)(struct vm_area_struct *vma,
656 struct page *(*find_special_page)(struct vm_area_struct *vma,
661 static inline void vma_numab_state_init(struct vm_area_struct *vma) in vma_numab_state_init()
665 static inline void vma_numab_state_free(struct vm_area_struct *vma) in vma_numab_state_free()
670 static inline void vma_numab_state_init(struct vm_area_struct *vma) {} in vma_numab_state_init()
671 static inline void vma_numab_state_free(struct vm_area_struct *vma) {} in vma_numab_state_free()
708 static inline void vma_init(struct vm_area_struct *vma, struct mm_struct *mm) in vma_init()
718 static inline void vm_flags_init(struct vm_area_struct *vma, in vm_flags_init()
729 static inline void vm_flags_reset(struct vm_area_struct *vma, in vm_flags_reset()
736 static inline void vm_flags_reset_once(struct vm_area_struct *vma, in vm_flags_reset_once()
743 static inline void vm_flags_set(struct vm_area_struct *vma, in vm_flags_set()
750 static inline void vm_flags_clear(struct vm_area_struct *vma, in vm_flags_clear()
761 static inline void __vm_flags_mod(struct vm_area_struct *vma, in __vm_flags_mod()
771 static inline void vm_flags_mod(struct vm_area_struct *vma, in vm_flags_mod()
778 static inline void vma_set_anonymous(struct vm_area_struct *vma) in vma_set_anonymous()
783 static inline bool vma_is_anonymous(struct vm_area_struct *vma) in vma_is_anonymous()
792 static inline bool vma_is_initial_heap(const struct vm_area_struct *vma) in vma_is_initial_heap()
802 static inline bool vma_is_initial_stack(const struct vm_area_struct *vma) in vma_is_initial_stack()
813 static inline bool vma_is_temporary_stack(struct vm_area_struct *vma) in vma_is_temporary_stack()
827 static inline bool vma_is_foreign(struct vm_area_struct *vma) in vma_is_foreign()
838 static inline bool vma_is_accessible(struct vm_area_struct *vma) in vma_is_accessible()
849 static inline bool vma_is_shared_maywrite(struct vm_area_struct *vma) in vma_is_shared_maywrite()
855 struct vm_area_struct *vma_find(struct vma_iterator *vmi, unsigned long max) in vma_find()
860 static inline struct vm_area_struct *vma_next(struct vma_iterator *vmi) in vma_next()
870 struct vm_area_struct *vma_iter_next_range(struct vma_iterator *vmi) in vma_iter_next_range()
876 static inline struct vm_area_struct *vma_prev(struct vma_iterator *vmi) in vma_prev()
899 struct vm_area_struct *vma) in vma_iter_bulk_store()
933 bool vma_is_shmem(struct vm_area_struct *vma);
934 bool vma_is_anon_shmem(struct vm_area_struct *vma);
936 static inline bool vma_is_shmem(struct vm_area_struct *vma) { return false; } in vma_is_shmem()
937 static inline bool vma_is_anon_shmem(struct vm_area_struct *vma) { return false; } in vma_is_anon_shmem()
940 int vma_is_stack_for_current(struct vm_area_struct *vma);
1233 static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) in maybe_mkwrite()
1611 static inline void vma_set_access_pid_bit(struct vm_area_struct *vma) in vma_set_access_pid_bit()
1672 static inline void vma_set_access_pid_bit(struct vm_area_struct *vma) in vma_set_access_pid_bit()
1898 static inline bool folio_needs_cow_for_dma(struct vm_area_struct *vma, in folio_needs_cow_for_dma()
2346 struct folio *vm_normal_folio(struct vm_area_struct *vma, unsigned long addr,
2348 struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr,
2350 struct folio *vm_normal_folio_pmd(struct vm_area_struct *vma,
2352 struct page *vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr,
2355 void zap_vma_ptes(struct vm_area_struct *vma, unsigned long address,
2357 void zap_page_range_single(struct vm_area_struct *vma, unsigned long address,
2359 static inline void zap_vma_pages(struct vm_area_struct *vma) in zap_vma_pages()
2365 struct vm_area_struct *start_vma, unsigned long start,
2373 copy_page_range(struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma);
2374 int generic_access_phys(struct vm_area_struct *vma, unsigned long addr,
2383 struct vm_area_struct *vma;
2417 struct vm_area_struct *lock_mm_and_find_vma(struct mm_struct *mm,
2421 extern vm_fault_t handle_mm_fault(struct vm_area_struct *vma,
2432 static inline vm_fault_t handle_mm_fault(struct vm_area_struct *vma, in handle_mm_fault()
2459 static inline struct vm_area_struct *vma_lookup(struct mm_struct *mm,
2487 struct vm_area_struct **vmap) in get_user_page_vma_remote()
2490 struct vm_area_struct *vma; in get_user_page_vma_remote()
2564 bool can_change_pte_writable(struct vm_area_struct *vma, unsigned long addr,
2567 struct vm_area_struct *vma, unsigned long start,
2570 struct vm_area_struct *vma, struct vm_area_struct **pprev,
3261 void vma_interval_tree_insert(struct vm_area_struct *node,
3263 void vma_interval_tree_insert_after(struct vm_area_struct *node,
3264 struct vm_area_struct *prev,
3266 void vma_interval_tree_remove(struct vm_area_struct *node,
3268 struct vm_area_struct *vma_interval_tree_iter_first(struct rb_root_cached *root,
3270 struct vm_area_struct *vma_interval_tree_iter_next(struct vm_area_struct *node,
3296 extern int insert_vm_struct(struct mm_struct *, struct vm_area_struct *);
3298 bool mmap_read_lock_maybe_expand(struct mm_struct *mm, struct vm_area_struct *vma,
3326 extern bool vma_is_special_mapping(const struct vm_area_struct *vma,
3328 struct vm_area_struct *_install_special_mapping(struct mm_struct *mm,
3354 int do_vmi_align_munmap(struct vma_iterator *vmi, struct vm_area_struct *vma,
3407 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address);
3408 struct vm_area_struct *expand_stack(struct mm_struct * mm, unsigned long addr);
3411 extern struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr);
3412 extern struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
3413 struct vm_area_struct **pprev);
3419 struct vm_area_struct *find_vma_intersection(struct mm_struct *mm,
3430 struct vm_area_struct *vma_lookup(struct mm_struct *mm, unsigned long addr) in vma_lookup()
3435 static inline unsigned long stack_guard_start_gap(struct vm_area_struct *vma) in stack_guard_start_gap()
3447 static inline unsigned long vm_start_gap(struct vm_area_struct *vma) in vm_start_gap()
3458 static inline unsigned long vm_end_gap(struct vm_area_struct *vma) in vm_end_gap()
3470 static inline unsigned long vma_pages(struct vm_area_struct *vma) in vma_pages()
3476 static inline struct vm_area_struct *find_exact_vma(struct mm_struct *mm, in find_exact_vma()
3479 struct vm_area_struct *vma = vma_lookup(mm, vm_start); in find_exact_vma()
3487 static inline bool range_in_vma(struct vm_area_struct *vma, in range_in_vma()
3495 void vma_set_page_prot(struct vm_area_struct *vma);
3501 static inline void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot()
3507 void vma_set_file(struct vm_area_struct *vma, struct file *file);
3510 unsigned long change_prot_numa(struct vm_area_struct *vma,
3514 struct vm_area_struct *find_extend_vma_locked(struct mm_struct *,
3516 int remap_pfn_range(struct vm_area_struct *, unsigned long addr,
3518 int remap_pfn_range_notrack(struct vm_area_struct *vma, unsigned long addr,
3520 int vm_insert_page(struct vm_area_struct *, unsigned long addr, struct page *);
3521 int vm_insert_pages(struct vm_area_struct *vma, unsigned long addr,
3523 int vm_map_pages(struct vm_area_struct *vma, struct page **pages,
3525 int vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages,
3529 vm_fault_t vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr,
3531 vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr,
3533 vm_fault_t vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr,
3535 vm_fault_t vmf_insert_mixed_mkwrite(struct vm_area_struct *vma,
3537 int vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len);
3539 static inline vm_fault_t vmf_insert_page(struct vm_area_struct *vma, in vmf_insert_page()
3553 static inline int io_remap_pfn_range(struct vm_area_struct *vma, in io_remap_pfn_range()
3603 static inline bool gup_can_follow_protnone(struct vm_area_struct *vma, in gup_can_follow_protnone()
3772 extern struct vm_area_struct *get_gate_vma(struct mm_struct *mm);
3776 static inline struct vm_area_struct *get_gate_vma(struct mm_struct *mm) in get_gate_vma()
3797 const char * arch_vma_name(struct vm_area_struct *vma);
4016 struct vm_area_struct *vma);
4031 static inline bool vma_is_special_huge(const struct vm_area_struct *vma) in vma_is_special_huge()
4100 void vma_pgtable_walk_begin(struct vm_area_struct *vma);
4101 void vma_pgtable_walk_end(struct vm_area_struct *vma);