Home
last modified time | relevance | path

Searched refs:pgoff (Results 1 – 25 of 62) sorted by relevance

123

/drivers/dax/
A Ddevice.c65 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys()
67 phys = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys()
81 pgoff_t pgoff; in dax_set_mapping() local
87 pgoff = linear_page_index(vmf->vma, in dax_set_mapping()
97 folio->index = pgoff + i; in dax_set_mapping()
141 pgoff_t pgoff; in __dev_dax_pmd_fault() local
164 pgoff = linear_page_index(vmf->vma, pmd_addr); in __dev_dax_pmd_fault()
186 pgoff_t pgoff; in __dev_dax_pud_fault() local
210 pgoff = linear_page_index(vmf->vma, pud_addr); in __dev_dax_pud_fault()
322 off = pgoff << PAGE_SHIFT; in dax_get_unmapped_area()
[all …]
A Dsuper.c149 long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, in dax_direct_access() argument
163 avail = dax_dev->ops->direct_access(dax_dev, pgoff, nr_pages, in dax_direct_access()
171 size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, in dax_copy_from_iter() argument
187 size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, in dax_copy_to_iter() argument
203 int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, in dax_zero_page_range() argument
218 ret = dax_dev->ops->zero_page_range(dax_dev, pgoff, nr_pages); in dax_zero_page_range()
223 size_t dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, in dax_recovery_write() argument
228 return dax_dev->ops->recovery_write(dax_dev, pgoff, addr, bytes, iter); in dax_recovery_write()
A Ddax-private.h62 unsigned long pgoff; member
113 phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, unsigned long size);
/drivers/md/
A Ddm-linear.c165 sector_t sector = linear_map_sector(ti, *pgoff << PAGE_SECTORS_SHIFT); in linear_dax_pgoff()
167 *pgoff = (get_start_sect(lc->dev->bdev) + sector) >> PAGE_SECTORS_SHIFT; in linear_dax_pgoff()
171 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() argument
175 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_direct_access()
177 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access()
180 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() argument
183 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_zero_page_range()
185 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in linear_dax_zero_page_range()
188 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write() argument
191 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_recovery_write()
[all …]
A Ddm-stripe.c309 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); in stripe_dax_pgoff()
313 *pgoff = (get_start_sect(bdev) + dev_sector) >> PAGE_SECTORS_SHIFT; in stripe_dax_pgoff()
317 static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_direct_access() argument
321 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_direct_access()
323 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in stripe_dax_direct_access()
326 static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_zero_page_range() argument
329 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_zero_page_range()
331 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in stripe_dax_zero_page_range()
334 static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_recovery_write() argument
337 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_recovery_write()
[all …]
A Ddm-log-writes.c886 pgoff_t *pgoff) in log_writes_dax_pgoff() argument
890 *pgoff += (get_start_sect(lc->dev->bdev) >> PAGE_SECTORS_SHIFT); in log_writes_dax_pgoff()
894 static long log_writes_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in log_writes_dax_direct_access() argument
898 struct dax_device *dax_dev = log_writes_dax_pgoff(ti, &pgoff); in log_writes_dax_direct_access()
900 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in log_writes_dax_direct_access()
903 static int log_writes_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in log_writes_dax_zero_page_range() argument
906 struct dax_device *dax_dev = log_writes_dax_pgoff(ti, &pgoff); in log_writes_dax_zero_page_range()
908 return dax_zero_page_range(dax_dev, pgoff, nr_pages << PAGE_SHIFT); in log_writes_dax_zero_page_range()
912 pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i) in log_writes_dax_recovery_write() argument
914 struct dax_device *dax_dev = log_writes_dax_pgoff(ti, &pgoff); in log_writes_dax_recovery_write()
[all …]
/drivers/w1/slaves/
A Dw1_ds250x.c49 int pgoff = pageno * W1_PAGE_SIZE; in w1_ds2502_read_page() local
63 buf[1] = pgoff & 0xff; in w1_ds2502_read_page()
64 buf[2] = pgoff >> 8; in w1_ds2502_read_page()
71 w1_read_block(sl->master, &data->eprom[pgoff], W1_PAGE_SIZE); in w1_ds2502_read_page()
74 if (w1_calc_crc8(&data->eprom[pgoff], W1_PAGE_SIZE) != crc8) in w1_ds2502_read_page()
88 int pgoff, epoff; in w1_ds2505_read_page() local
97 epoff = pgoff = pageno * W1_PAGE_SIZE; in w1_ds2505_read_page()
105 buf[1] = pgoff & 0xff; in w1_ds2505_read_page()
106 buf[2] = pgoff >> 8; in w1_ds2505_read_page()
121 pgoff = (redir ^ 0xff) * W1_PAGE_SIZE; in w1_ds2505_read_page()
/drivers/dma-buf/
A Dudmabuf.c51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault() local
55 if (pgoff >= ubuf->pagecount) in udmabuf_vm_fault()
58 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault()
59 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault()
66 pgoff = vma->vm_pgoff; in udmabuf_vm_fault()
73 if (WARN_ON(pgoff >= ubuf->pagecount)) in udmabuf_vm_fault()
76 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault()
77 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault()
331 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios() local
343 pgoff_t subpgoff = pgoff; in udmabuf_pin_folios()
[all …]
/drivers/video/fbdev/core/
A Dfb_defio.c51 unsigned long pgoff = offset >> PAGE_SHIFT; in fb_deferred_io_pageref_lookup() local
54 if (fb_WARN_ON_ONCE(info, pgoff >= info->npagerefs)) in fb_deferred_io_pageref_lookup()
58 pageref = &info->pagerefs[pgoff]; in fb_deferred_io_pageref_lookup()
64 pageref->offset = pgoff << PAGE_SHIFT; in fb_deferred_io_pageref_lookup()
128 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault()
222 unsigned long offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_page_mkwrite()
272 pgoff_t pgoff = pageref->offset >> PAGE_SHIFT; in fb_deferred_io_work() local
274 mapping_wrprotect_range(fbdefio->mapping, pgoff, in fb_deferred_io_work()
A Dfb_chrdev.c394 unsigned long pgoff, unsigned long flags) in get_fb_unmapped_area() argument
399 if (pgoff > fb_size || len > fb_size - pgoff) in get_fb_unmapped_area()
402 return (unsigned long)info->screen_base + pgoff; in get_fb_unmapped_area()
/drivers/infiniband/core/
A Dib_core_uverbs.c118 unsigned long pgoff) in rdma_user_mmap_entry_get_pgoff() argument
122 if (pgoff > U32_MAX) in rdma_user_mmap_entry_get_pgoff()
127 entry = xa_load(&ucontext->mmap_xa, pgoff); in rdma_user_mmap_entry_get_pgoff()
134 if (!entry || entry->start_pgoff != pgoff || entry->driver_removed || in rdma_user_mmap_entry_get_pgoff()
141 pgoff, entry->npages); in rdma_user_mmap_entry_get_pgoff()
A Dumem.c85 unsigned long va, pgoff; in ib_umem_find_best_pgsz() local
110 pgoff = umem->address & ~PAGE_MASK; in ib_umem_find_best_pgsz()
123 mask |= (curr_base + pgoff) ^ va; in ib_umem_find_best_pgsz()
135 va += sg_dma_len(sg) - pgoff; in ib_umem_find_best_pgsz()
137 pgoff = 0; in ib_umem_find_best_pgsz()
/drivers/android/
A Dbinder_alloc.c821 *pgoffp = pgoff; in binder_alloc_get_page()
842 pgoff_t pgoff; in binder_alloc_clear_buf() local
845 buffer_offset, &pgoff); in binder_alloc_clear_buf()
847 memset_page(page, pgoff, 0, size); in binder_alloc_clear_buf()
1343 pgoff_t pgoff; in binder_alloc_copy_user_to_buffer() local
1347 buffer_offset, &pgoff); in binder_alloc_copy_user_to_buffer()
1349 kptr = kmap_local_page(page) + pgoff; in binder_alloc_copy_user_to_buffer()
1375 pgoff_t pgoff; in binder_alloc_do_buffer_copy() local
1378 buffer_offset, &pgoff); in binder_alloc_do_buffer_copy()
1381 memcpy_to_page(page, pgoff, ptr, size); in binder_alloc_do_buffer_copy()
[all …]
/drivers/iommu/iommufd/
A Diova_bitmap.c45 unsigned long pgoff; member
210 mapped->pgoff = offset_in_page(addr); in iova_bitmap_get()
307 bytes = (bitmap->mapped.npages << PAGE_SHIFT) - bitmap->mapped.pgoff; in iova_bitmap_mapped_remaining()
421 mapped->pgshift) + mapped->pgoff * BITS_PER_BYTE; in iova_bitmap_set()
423 mapped->pgshift) + mapped->pgoff * BITS_PER_BYTE; in iova_bitmap_set()
A Ddriver.c196 max_pgoff = max(max_pgoff, cur->pgoff + 1); in iommufd_sw_msi_get_map()
211 cur->pgoff = max_pgoff; in iommufd_sw_msi_get_map()
225 iova = msi_map->sw_msi_start + msi_map->pgoff * PAGE_SIZE; in iommufd_sw_msi_install()
295 iova = msi_map->sw_msi_start + msi_map->pgoff * PAGE_SIZE; in iommufd_sw_msi()
/drivers/nvdimm/
A Dpmem.c242 __weak long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, in __pmem_direct_access() argument
246 resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; in __pmem_direct_access()
247 sector_t sector = PFN_PHYS(pgoff) >> SECTOR_SHIFT; in __pmem_direct_access()
293 static int pmem_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, in pmem_dax_zero_page_range() argument
299 PFN_PHYS(pgoff) >> SECTOR_SHIFT, in pmem_dax_zero_page_range()
304 pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, in pmem_dax_direct_access() argument
309 return __pmem_direct_access(pmem, pgoff, nr_pages, mode, kaddr, pfn); in pmem_dax_direct_access()
325 static size_t pmem_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, in pmem_recovery_write() argument
336 if (!is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) >> SECTOR_SHIFT, len)) in pmem_recovery_write()
349 pmem_off = PFN_PHYS(pgoff) + pmem->data_offset; in pmem_recovery_write()
/drivers/gpu/drm/omapdrm/
A Domap_gem.c357 pgoff_t pgoff; in omap_gem_fault_1d() local
360 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in omap_gem_fault_1d()
363 omap_gem_cpu_sync_page(obj, pgoff); in omap_gem_fault_1d()
364 pfn = page_to_pfn(omap_obj->pages[pgoff]); in omap_gem_fault_1d()
367 pfn = (omap_obj->dma_addr >> PAGE_SHIFT) + pgoff; in omap_gem_fault_1d()
386 pgoff_t pgoff, base_pgoff; in omap_gem_fault_2d() local
408 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in omap_gem_fault_2d()
414 base_pgoff = round_down(pgoff, m << n_shift); in omap_gem_fault_2d()
435 int off = pgoff % m; in omap_gem_fault_2d()
700 if (omap_obj->dma_addrs[pgoff]) { in omap_gem_cpu_sync_page()
[all …]
/drivers/xen/
A Dgntalloc.c83 uint16_t pgoff:12; /* Bits 0-11: Offset of the byte to clear */ member
180 tmp[gref->notify.pgoff] = 0; in __del_gref()
377 int pgoff; in gntalloc_ioctl_unmap_notify() local
384 pgoff = op.index & (PAGE_SIZE - 1); in gntalloc_ioctl_unmap_notify()
417 gref->notify.pgoff = pgoff; in gntalloc_ioctl_unmap_notify()
/drivers/hwtracing/intel_th/
A Dmsu.c70 unsigned long pgoff; member
993 unsigned long pgoff) in msc_buffer_contig_get_page() argument
995 if (pgoff >= msc->nr_pages) in msc_buffer_contig_get_page()
1122 win->pgoff = prev->pgoff + prev->nr_blocks; in msc_buffer_win_alloc()
1413 return msc_buffer_contig_get_page(msc, pgoff); in msc_buffer_get_page()
1416 if (pgoff >= win->pgoff && pgoff < win->pgoff + win->nr_blocks) in msc_buffer_get_page()
1422 pgoff -= win->pgoff; in msc_buffer_get_page()
1428 if (pgoff < pgsz) in msc_buffer_get_page()
1429 return page + pgoff; in msc_buffer_get_page()
1431 pgoff -= pgsz; in msc_buffer_get_page()
[all …]
/drivers/char/
A Dmem.c286 unsigned long pgoff, in get_unmapped_area_mem() argument
289 if (!valid_mmap_phys_addr_range(pgoff, len)) in get_unmapped_area_mem()
291 return pgoff << PAGE_SHIFT; in get_unmapped_area_mem()
517 unsigned long pgoff, unsigned long flags) in get_unmapped_area_zero() argument
527 return shmem_get_unmapped_area(NULL, addr, len, pgoff, flags); in get_unmapped_area_zero()
531 return mm_get_unmapped_area(current->mm, file, addr, len, pgoff, flags); in get_unmapped_area_zero()
/drivers/vfio/cdx/
A Dmain.c239 u64 pgoff, base; in vfio_cdx_mmap_mmio() local
241 pgoff = vma->vm_pgoff & in vfio_cdx_mmap_mmio()
243 base = pgoff << PAGE_SHIFT; in vfio_cdx_mmap_mmio()
248 vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff; in vfio_cdx_mmap_mmio()
/drivers/s390/block/
A Ddcssblk.c33 static long dcssblk_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff,
48 pgoff_t pgoff, size_t nr_pages) in dcssblk_dax_zero_page_range() argument
53 rc = dax_direct_access(dax_dev, pgoff, nr_pages, DAX_ACCESS, in dcssblk_dax_zero_page_range()
915 __dcssblk_direct_access(struct dcssblk_dev_info *dev_info, pgoff_t pgoff, in __dcssblk_direct_access() argument
918 resource_size_t offset = pgoff * PAGE_SIZE; in __dcssblk_direct_access()
931 dcssblk_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, in dcssblk_dax_direct_access() argument
937 return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn); in dcssblk_dax_direct_access()
/drivers/comedi/
A Dcomedi_buf.c159 unsigned int pgoff = offset_in_page(offset); in comedi_buf_map_access() local
164 int l = min_t(int, len - done, PAGE_SIZE - pgoff); in comedi_buf_map_access()
165 void *b = bm->page_list[pg].virt_addr + pgoff; in comedi_buf_map_access()
174 pgoff = 0; in comedi_buf_map_access()
/drivers/vfio/fsl-mc/
A Dvfio_fsl_mc.c363 u64 pgoff, base; in vfio_fsl_mc_mmap_mmio() local
366 pgoff = vma->vm_pgoff & in vfio_fsl_mc_mmap_mmio()
368 base = pgoff << PAGE_SHIFT; in vfio_fsl_mc_mmap_mmio()
378 vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff; in vfio_fsl_mc_mmap_mmio()
/drivers/misc/ocxl/
A Dsysfs.c115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT)) in global_mmio_fault()
118 offset = vmf->pgoff; in global_mmio_fault()

Completed in 105 milliseconds

123