Home
last modified time | relevance | path

Searched refs:vmf (Results 1 – 25 of 175) sorted by relevance

1234567

/linux/mm/
A Dmemory.c3389 vmf->pte = pte_offset_map_lock(mm, vmf->pmd, vmf->address, &vmf->ptl); in wp_page_copy()
3502 vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, vmf->address, in finish_mkwrite_fault()
3511 update_mmu_tlb(vmf->vma, vmf->address, vmf->pte); in finish_mkwrite_fault()
3942 vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, in pte_marker_clear()
3955 pte_clear(vmf->vma->vm_mm, vmf->address, vmf->pte); in pte_marker_clear()
4243 vmf->address, &vmf->ptl); in do_swap_page()
4335 vmf->address, &vmf->ptl); in do_swap_page()
4761 vmf->address, &vmf->ptl); in do_anonymous_page()
5411 vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, in do_fault()
5777 update_mmu_tlb(vmf->vma, vmf->address, vmf->pte); in handle_pte_fault()
[all …]
A Dhuge_memory.c1175 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
1201 update_mmu_cache_pmd(vma, vmf->address, vmf->pmd); in __do_huge_pmd_anonymous_page()
1303 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page()
1742 vmf->ptl = pud_lock(vmf->vma->vm_mm, vmf->pud); in huge_pud_set_accessed()
1746 touch_pud(vmf->vma, vmf->address, vmf->pud, write); in huge_pud_set_accessed()
1756 vmf->ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in huge_pmd_set_accessed()
1757 if (unlikely(!pmd_same(*vmf->pmd, vmf->orig_pmd))) in huge_pmd_set_accessed()
1760 touch_pmd(vmf->vma, vmf->address, vmf->pmd, write); in huge_pmd_set_accessed()
1775 vmf->ptl = pmd_lockptr(vma->vm_mm, vmf->pmd); in do_huge_pmd_wp_page()
1895 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page()
[all …]
A Dfilemap.c3261 ptep = pte_offset_map_nolock(vma->vm_mm, vmf->pmd, vmf->address, in filemap_fault_recheck_pte_none()
3262 &vmf->ptl); in filemap_fault_recheck_pte_none()
3269 spin_lock(vmf->ptl); in filemap_fault_recheck_pte_none()
3355 vmf->gfp_mask); in filemap_fault()
3484 if (pmd_none(*vmf->pmd) && vmf->prealloc_pte) in filemap_map_pmd()
3485 pmd_install(mm, vmf->pmd, &vmf->prealloc_pte); in filemap_map_pmd()
3580 vmf->pte += count; in filemap_map_folio_range()
3593 vmf->pte = old_ptep; in filemap_map_folio_range()
3655 vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl); in filemap_map_pages()
3656 if (!vmf->pte) { in filemap_map_pages()
[all …]
A Dhugetlb.c6107 if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) { in hugetlb_no_page()
6136 if (hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) in hugetlb_no_page()
6185 if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) { in hugetlb_no_page()
6209 vmf->ptl = huge_pte_lock(h, mm, vmf->pte); in hugetlb_no_page()
6212 if (!pte_same(huge_ptep_get(mm, vmf->address, vmf->pte), vmf->orig_pte)) in hugetlb_no_page()
6341 vmf.orig_pte = huge_ptep_get(mm, vmf.address, vmf.pte); in hugetlb_fault()
6412 vmf.ptl = huge_pte_lock(h, mm, vmf.pte); in hugetlb_fault()
6415 if (unlikely(!pte_same(vmf.orig_pte, huge_ptep_get(mm, vmf.address, vmf.pte)))) in hugetlb_fault()
6433 set_huge_pte_at(mm, vmf.address, vmf.pte, vmf.orig_pte, in hugetlb_fault()
6460 vmf.orig_pte = pte_mkyoung(vmf.orig_pte); in hugetlb_fault()
[all …]
A Dsecretmem.c50 static vm_fault_t secretmem_fault(struct vm_fault *vmf) in secretmem_fault() argument
52 struct address_space *mapping = vmf->vma->vm_file->f_mapping; in secretmem_fault()
53 struct inode *inode = file_inode(vmf->vma->vm_file); in secretmem_fault()
54 pgoff_t offset = vmf->pgoff; in secretmem_fault()
55 gfp_t gfp = vmf->gfp_mask; in secretmem_fault()
62 if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode)) in secretmem_fault()
105 vmf->page = page; in secretmem_fault()
A Dswap_state.c745 static int swap_vma_ra_win(struct vm_fault *vmf, unsigned long *start, in swap_vma_ra_win() argument
748 struct vm_area_struct *vma = vmf->vma; in swap_vma_ra_win()
757 faddr = vmf->address; in swap_vma_ra_win()
800 struct mempolicy *mpol, pgoff_t targ_ilx, struct vm_fault *vmf) in swap_vma_readahead() argument
812 win = swap_vma_ra_win(vmf, &start, &end); in swap_vma_readahead()
816 ilx = targ_ilx - PFN_DOWN(vmf->address - start); in swap_vma_readahead()
821 pte = pte_offset_map(vmf->pmd, addr); in swap_vma_readahead()
839 if (addr != vmf->address) { in swap_vma_readahead()
873 struct vm_fault *vmf) in swapin_readahead() argument
879 mpol = get_vma_policy(vmf->vma, vmf->address, 0, &ilx); in swapin_readahead()
[all …]
/linux/include/trace/events/
A Dfs_dax.h13 TP_ARGS(inode, vmf, max_pgoff, result),
30 __entry->vm_end = vmf->vma->vm_end;
32 __entry->address = vmf->address;
33 __entry->flags = vmf->flags;
34 __entry->pgoff = vmf->pgoff;
80 __entry->address = vmf->address;
123 __entry->address = vmf->address;
155 TP_ARGS(inode, vmf, result),
170 __entry->flags = vmf->flags;
171 __entry->pgoff = vmf->pgoff;
[all …]
/linux/drivers/dax/
A Ddevice.c88 pgoff = linear_page_index(vmf->vma, in dax_set_mapping()
104 struct vm_fault *vmf) in __dev_dax_pte_fault() argument
133 return vmf_insert_mixed(vmf->vma, vmf->address, pfn); in __dev_dax_pte_fault()
137 struct vm_fault *vmf) in __dev_dax_pmd_fault() argument
161 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault()
176 return vmf_insert_pfn_pmd(vmf, pfn, vmf->flags & FAULT_FLAG_WRITE); in __dev_dax_pmd_fault()
181 struct vm_fault *vmf) in __dev_dax_pud_fault() argument
206 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault()
221 return vmf_insert_pfn_pud(vmf, pfn, vmf->flags & FAULT_FLAG_WRITE); in __dev_dax_pud_fault()
225 struct vm_fault *vmf) in __dev_dax_pud_fault() argument
[all …]
/linux/fs/
A Ddax.c845 copy_user_page(vto, kaddr, vmf->address, vmf->cow_page); in copy_cow_page_dax()
1230 ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in dax_pmd_load_hole()
1242 set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry); in dax_pmd_load_hole()
1627 clear_user_highpage(vmf->cow_page, vmf->address); in dax_fault_cow_page()
1708 return vmf_insert_mixed_mkwrite(vmf->vma, vmf->address, pfn); in dax_fault_iter()
1709 return vmf_insert_mixed(vmf->vma, vmf->address, pfn); in dax_fault_iter()
1738 if ((vmf->flags & FAULT_FLAG_WRITE) && !vmf->cow_page) in dax_iomap_pte_fault()
1753 if (pmd_trans_huge(*vmf->pmd) || pmd_devmap(*vmf->pmd)) { in dax_iomap_pte_fault()
1874 if (!pmd_none(*vmf->pmd) && !pmd_trans_huge(*vmf->pmd) && in dax_iomap_pmd_fault()
1894 split_huge_pmd(vmf->vma, vmf->pmd, vmf->address); in dax_iomap_pmd_fault()
[all …]
A Duserfaultfd.c231 struct vm_fault *vmf, in userfaultfd_huge_must_wait() argument
234 struct vm_area_struct *vma = vmf->vma; in userfaultfd_huge_must_wait()
238 assert_fault_locked(vmf); in userfaultfd_huge_must_wait()
261 struct vm_fault *vmf, in userfaultfd_huge_must_wait() argument
276 struct vm_fault *vmf, in userfaultfd_must_wait() argument
280 unsigned long address = vmf->address; in userfaultfd_must_wait()
289 assert_fault_locked(vmf); in userfaultfd_must_wait()
380 assert_fault_locked(vmf); in handle_userfault()
445 vmf->flags); in handle_userfault()
465 uwq.msg = userfault_msg(vmf->address, vmf->real_address, vmf->flags, in handle_userfault()
[all …]
/linux/drivers/gpu/drm/ttm/
A Dttm_bo_vm.c42 struct vm_fault *vmf) in ttm_bo_vm_fault_idle() argument
58 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT) in ttm_bo_vm_fault_idle()
62 mmap_read_unlock(vmf->vma->vm_mm); in ttm_bo_vm_fault_idle()
117 struct vm_fault *vmf) in ttm_bo_vm_reserve() argument
134 mmap_read_unlock(vmf->vma->vm_mm); in ttm_bo_vm_reserve()
185 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault_reserved()
196 unsigned long address = vmf->address; in ttm_bo_vm_fault_reserved()
202 ret = ttm_bo_vm_fault_idle(bo, vmf); in ttm_bo_vm_fault_reserved()
293 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_dummy_page()
323 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault()
[all …]
/linux/fs/ocfs2/
A Dmmap.c31 static vm_fault_t ocfs2_fault(struct vm_fault *vmf) in ocfs2_fault() argument
33 struct vm_area_struct *vma = vmf->vma; in ocfs2_fault()
38 ret = filemap_fault(vmf); in ocfs2_fault()
42 vma, vmf->page, vmf->pgoff); in ocfs2_fault()
113 static vm_fault_t ocfs2_page_mkwrite(struct vm_fault *vmf) in ocfs2_page_mkwrite() argument
115 struct page *page = vmf->page; in ocfs2_page_mkwrite()
116 struct inode *inode = file_inode(vmf->vma->vm_file); in ocfs2_page_mkwrite()
144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page); in ocfs2_page_mkwrite()
/linux/drivers/video/fbdev/core/
A Dfb_defio.c133 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_fault()
135 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault()
143 if (vmf->vma->vm_file) in fb_deferred_io_fault()
144 page->mapping = vmf->vma->vm_file->f_mapping; in fb_deferred_io_fault()
149 page->index = vmf->pgoff; /* for folio_mkclean() */ in fb_deferred_io_fault()
151 vmf->page = page; in fb_deferred_io_fault()
232 unsigned long offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_page_mkwrite()
233 struct page *page = vmf->page; in fb_deferred_io_page_mkwrite()
235 file_update_time(vmf->vma->vm_file); in fb_deferred_io_page_mkwrite()
243 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_mkwrite()
[all …]
/linux/arch/x86/entry/vdso/
A Dvma.c66 struct vm_area_struct *vma, struct vm_fault *vmf) in vdso_fault() argument
70 if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size) in vdso_fault()
73 vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT)); in vdso_fault()
74 get_page(vmf->page); in vdso_fault()
132 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument
141 sym_offset = (long)(vmf->pgoff << PAGE_SHIFT) + in vvar_fault()
176 addr = vmf->address + (image->sym_timens_page - sym_offset); in vvar_fault()
184 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault()
189 return vmf_insert_pfn_prot(vma, vmf->address, in vvar_fault()
197 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault()
[all …]
/linux/drivers/gpu/drm/vmwgfx/
A Dvmwgfx_page_dirty.c375 vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf) in vmw_bo_vm_mkwrite() argument
377 struct vm_area_struct *vma = vmf->vma; in vmw_bo_vm_mkwrite()
389 save_flags = vmf->flags; in vmw_bo_vm_mkwrite()
390 vmf->flags &= ~FAULT_FLAG_ALLOW_RETRY; in vmw_bo_vm_mkwrite()
391 ret = ttm_bo_vm_reserve(bo, vmf); in vmw_bo_vm_mkwrite()
392 vmf->flags = save_flags; in vmw_bo_vm_mkwrite()
416 vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf) in vmw_bo_vm_fault() argument
418 struct vm_area_struct *vma = vmf->vma; in vmw_bo_vm_fault()
426 ret = ttm_bo_vm_reserve(bo, vmf); in vmw_bo_vm_fault()
437 page_offset = vmf->pgoff - in vmw_bo_vm_fault()
[all …]
/linux/fs/xfs/
A Dxfs_file.c1400 struct vm_fault *vmf, in xfs_dax_fault_locked() argument
1422 struct vm_fault *vmf, in xfs_dax_read_fault() argument
1437 struct vm_fault *vmf, in xfs_write_fault() argument
1482 struct vm_fault *vmf, in __xfs_filemap_fault() argument
1494 return filemap_fault(vmf); in __xfs_filemap_fault()
1499 struct vm_fault *vmf) in xfs_is_write_fault() argument
1507 struct vm_fault *vmf) in xfs_filemap_fault() argument
1512 xfs_is_write_fault(vmf)); in xfs_filemap_fault()
1517 struct vm_fault *vmf, in xfs_filemap_huge_fault() argument
1530 struct vm_fault *vmf) in xfs_filemap_page_mkwrite() argument
[all …]
/linux/drivers/xen/
A Dprivcmd-buf.c118 static vm_fault_t privcmd_buf_vma_fault(struct vm_fault *vmf) in privcmd_buf_vma_fault() argument
121 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, in privcmd_buf_vma_fault()
122 vmf->pgoff, (void *)vmf->address); in privcmd_buf_vma_fault()
/linux/fs/nilfs2/
A Dfile.c45 static vm_fault_t nilfs_page_mkwrite(struct vm_fault *vmf) in nilfs_page_mkwrite() argument
47 struct vm_area_struct *vma = vmf->vma; in nilfs_page_mkwrite()
48 struct folio *folio = page_folio(vmf->page); in nilfs_page_mkwrite()
101 ret = block_page_mkwrite(vma, vmf, nilfs_get_block); in nilfs_page_mkwrite()
/linux/drivers/misc/ocxl/
A Dsysfs.c109 static vm_fault_t global_mmio_fault(struct vm_fault *vmf) in global_mmio_fault() argument
111 struct vm_area_struct *vma = vmf->vma; in global_mmio_fault()
115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT)) in global_mmio_fault()
118 offset = vmf->pgoff; in global_mmio_fault()
120 return vmf_insert_pfn(vma, vmf->address, offset); in global_mmio_fault()
A Dcontext.c139 static vm_fault_t ocxl_mmap_fault(struct vm_fault *vmf) in ocxl_mmap_fault() argument
141 struct vm_area_struct *vma = vmf->vma; in ocxl_mmap_fault()
146 offset = vmf->pgoff << PAGE_SHIFT; in ocxl_mmap_fault()
148 ctx->pasid, vmf->address, offset); in ocxl_mmap_fault()
151 ret = map_pp_mmio(vma, vmf->address, offset, ctx); in ocxl_mmap_fault()
153 ret = map_afu_irq(vma, vmf->address, offset, ctx); in ocxl_mmap_fault()
/linux/drivers/misc/cxl/
A Dcontext.c126 static vm_fault_t cxl_mmap_fault(struct vm_fault *vmf) in cxl_mmap_fault() argument
128 struct vm_area_struct *vma = vmf->vma; in cxl_mmap_fault()
133 offset = vmf->pgoff << PAGE_SHIFT; in cxl_mmap_fault()
136 __func__, ctx->pe, vmf->address, offset); in cxl_mmap_fault()
161 vmf->page = ctx->ff_page; in cxl_mmap_fault()
168 ret = vmf_insert_pfn(vma, vmf->address, (area + offset) >> PAGE_SHIFT); in cxl_mmap_fault()
/linux/include/linux/
A Dhuge_mm.h11 vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf);
15 void huge_pmd_set_accessed(struct vm_fault *vmf);
21 void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud);
23 static inline void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) in huge_pud_set_accessed() argument
28 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf);
41 vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write);
42 vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write);
456 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf);
602 static inline vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf) in do_huge_pmd_numa_page() argument
/linux/fs/ext2/
A Dfile.c93 static vm_fault_t ext2_dax_fault(struct vm_fault *vmf) in ext2_dax_fault() argument
95 struct inode *inode = file_inode(vmf->vma->vm_file); in ext2_dax_fault()
97 bool write = (vmf->flags & FAULT_FLAG_WRITE) && in ext2_dax_fault()
98 (vmf->vma->vm_flags & VM_SHARED); in ext2_dax_fault()
102 file_update_time(vmf->vma->vm_file); in ext2_dax_fault()
106 ret = dax_iomap_fault(vmf, 0, NULL, NULL, &ext2_iomap_ops); in ext2_dax_fault()
/linux/arch/loongarch/kernel/
A Dvdso.c49 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument
54 switch (vmf->pgoff) { in vvar_fault()
76 pfn = sym_to_pfn(&loongarch_vdso_data) + vmf->pgoff - VVAR_LOONGARCH_PAGES_START; in vvar_fault()
82 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault()
/linux/drivers/char/agp/
A Dalpha-agp.c14 static vm_fault_t alpha_core_agp_vm_fault(struct vm_fault *vmf) in alpha_core_agp_vm_fault() argument
21 dma_addr = vmf->address - vmf->vma->vm_start + agp->aperture.bus_base; in alpha_core_agp_vm_fault()
32 vmf->page = page; in alpha_core_agp_vm_fault()

Completed in 101 milliseconds

1234567