| /drivers/dax/ |
| A D | device.c | 65 if (pgoff < dax_range->pgoff || pgoff > pgoff_end) in dax_pgoff_to_phys() 67 phys = PFN_PHYS(pgoff - dax_range->pgoff) + range->start; in dax_pgoff_to_phys() 81 pgoff_t pgoff; in dax_set_mapping() local 87 pgoff = linear_page_index(vmf->vma, in dax_set_mapping() 97 folio->index = pgoff + i; in dax_set_mapping() 141 pgoff_t pgoff; in __dev_dax_pmd_fault() local 164 pgoff = linear_page_index(vmf->vma, pmd_addr); in __dev_dax_pmd_fault() 186 pgoff_t pgoff; in __dev_dax_pud_fault() local 210 pgoff = linear_page_index(vmf->vma, pud_addr); in __dev_dax_pud_fault() 322 off = pgoff << PAGE_SHIFT; in dax_get_unmapped_area() [all …]
|
| A D | super.c | 149 long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, in dax_direct_access() argument 163 avail = dax_dev->ops->direct_access(dax_dev, pgoff, nr_pages, in dax_direct_access() 171 size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, in dax_copy_from_iter() argument 187 size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, in dax_copy_to_iter() argument 203 int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, in dax_zero_page_range() argument 218 ret = dax_dev->ops->zero_page_range(dax_dev, pgoff, nr_pages); in dax_zero_page_range() 223 size_t dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, in dax_recovery_write() argument 228 return dax_dev->ops->recovery_write(dax_dev, pgoff, addr, bytes, iter); in dax_recovery_write()
|
| A D | dax-private.h | 62 unsigned long pgoff; member 113 phys_addr_t dax_pgoff_to_phys(struct dev_dax *dev_dax, pgoff_t pgoff, unsigned long size);
|
| /drivers/md/ |
| A D | dm-linear.c | 165 sector_t sector = linear_map_sector(ti, *pgoff << PAGE_SECTORS_SHIFT); in linear_dax_pgoff() 167 *pgoff = (get_start_sect(lc->dev->bdev) + sector) >> PAGE_SECTORS_SHIFT; in linear_dax_pgoff() 171 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() argument 175 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_direct_access() 177 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access() 180 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() argument 183 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_zero_page_range() 185 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in linear_dax_zero_page_range() 188 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write() argument 191 struct dax_device *dax_dev = linear_dax_pgoff(ti, &pgoff); in linear_dax_recovery_write() [all …]
|
| A D | dm-stripe.c | 309 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); in stripe_dax_pgoff() 313 *pgoff = (get_start_sect(bdev) + dev_sector) >> PAGE_SECTORS_SHIFT; in stripe_dax_pgoff() 317 static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_direct_access() argument 321 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_direct_access() 323 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in stripe_dax_direct_access() 326 static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_zero_page_range() argument 329 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_zero_page_range() 331 return dax_zero_page_range(dax_dev, pgoff, nr_pages); in stripe_dax_zero_page_range() 334 static size_t stripe_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in stripe_dax_recovery_write() argument 337 struct dax_device *dax_dev = stripe_dax_pgoff(ti, &pgoff); in stripe_dax_recovery_write() [all …]
|
| A D | dm-log-writes.c | 886 pgoff_t *pgoff) in log_writes_dax_pgoff() argument 890 *pgoff += (get_start_sect(lc->dev->bdev) >> PAGE_SECTORS_SHIFT); in log_writes_dax_pgoff() 894 static long log_writes_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in log_writes_dax_direct_access() argument 898 struct dax_device *dax_dev = log_writes_dax_pgoff(ti, &pgoff); in log_writes_dax_direct_access() 900 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in log_writes_dax_direct_access() 903 static int log_writes_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in log_writes_dax_zero_page_range() argument 906 struct dax_device *dax_dev = log_writes_dax_pgoff(ti, &pgoff); in log_writes_dax_zero_page_range() 908 return dax_zero_page_range(dax_dev, pgoff, nr_pages << PAGE_SHIFT); in log_writes_dax_zero_page_range() 912 pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i) in log_writes_dax_recovery_write() argument 914 struct dax_device *dax_dev = log_writes_dax_pgoff(ti, &pgoff); in log_writes_dax_recovery_write() [all …]
|
| /drivers/w1/slaves/ |
| A D | w1_ds250x.c | 49 int pgoff = pageno * W1_PAGE_SIZE; in w1_ds2502_read_page() local 63 buf[1] = pgoff & 0xff; in w1_ds2502_read_page() 64 buf[2] = pgoff >> 8; in w1_ds2502_read_page() 71 w1_read_block(sl->master, &data->eprom[pgoff], W1_PAGE_SIZE); in w1_ds2502_read_page() 74 if (w1_calc_crc8(&data->eprom[pgoff], W1_PAGE_SIZE) != crc8) in w1_ds2502_read_page() 88 int pgoff, epoff; in w1_ds2505_read_page() local 97 epoff = pgoff = pageno * W1_PAGE_SIZE; in w1_ds2505_read_page() 105 buf[1] = pgoff & 0xff; in w1_ds2505_read_page() 106 buf[2] = pgoff >> 8; in w1_ds2505_read_page() 121 pgoff = (redir ^ 0xff) * W1_PAGE_SIZE; in w1_ds2505_read_page()
|
| /drivers/dma-buf/ |
| A D | udmabuf.c | 51 pgoff_t pgoff = vmf->pgoff; in udmabuf_vm_fault() local 55 if (pgoff >= ubuf->pagecount) in udmabuf_vm_fault() 58 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault() 59 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault() 66 pgoff = vma->vm_pgoff; in udmabuf_vm_fault() 73 if (WARN_ON(pgoff >= ubuf->pagecount)) in udmabuf_vm_fault() 76 pfn = folio_pfn(ubuf->folios[pgoff]); in udmabuf_vm_fault() 77 pfn += ubuf->offsets[pgoff] >> PAGE_SHIFT; in udmabuf_vm_fault() 331 pgoff_t pgoff, pgcnt; in udmabuf_pin_folios() local 343 pgoff_t subpgoff = pgoff; in udmabuf_pin_folios() [all …]
|
| /drivers/video/fbdev/core/ |
| A D | fb_defio.c | 51 unsigned long pgoff = offset >> PAGE_SHIFT; in fb_deferred_io_pageref_lookup() local 54 if (fb_WARN_ON_ONCE(info, pgoff >= info->npagerefs)) in fb_deferred_io_pageref_lookup() 58 pageref = &info->pagerefs[pgoff]; in fb_deferred_io_pageref_lookup() 64 pageref->offset = pgoff << PAGE_SHIFT; in fb_deferred_io_pageref_lookup() 128 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault() 222 unsigned long offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_page_mkwrite() 272 pgoff_t pgoff = pageref->offset >> PAGE_SHIFT; in fb_deferred_io_work() local 274 mapping_wrprotect_range(fbdefio->mapping, pgoff, in fb_deferred_io_work()
|
| A D | fb_chrdev.c | 394 unsigned long pgoff, unsigned long flags) in get_fb_unmapped_area() argument 399 if (pgoff > fb_size || len > fb_size - pgoff) in get_fb_unmapped_area() 402 return (unsigned long)info->screen_base + pgoff; in get_fb_unmapped_area()
|
| /drivers/infiniband/core/ |
| A D | ib_core_uverbs.c | 118 unsigned long pgoff) in rdma_user_mmap_entry_get_pgoff() argument 122 if (pgoff > U32_MAX) in rdma_user_mmap_entry_get_pgoff() 127 entry = xa_load(&ucontext->mmap_xa, pgoff); in rdma_user_mmap_entry_get_pgoff() 134 if (!entry || entry->start_pgoff != pgoff || entry->driver_removed || in rdma_user_mmap_entry_get_pgoff() 141 pgoff, entry->npages); in rdma_user_mmap_entry_get_pgoff()
|
| A D | umem.c | 85 unsigned long va, pgoff; in ib_umem_find_best_pgsz() local 110 pgoff = umem->address & ~PAGE_MASK; in ib_umem_find_best_pgsz() 123 mask |= (curr_base + pgoff) ^ va; in ib_umem_find_best_pgsz() 135 va += sg_dma_len(sg) - pgoff; in ib_umem_find_best_pgsz() 137 pgoff = 0; in ib_umem_find_best_pgsz()
|
| /drivers/android/ |
| A D | binder_alloc.c | 821 *pgoffp = pgoff; in binder_alloc_get_page() 842 pgoff_t pgoff; in binder_alloc_clear_buf() local 845 buffer_offset, &pgoff); in binder_alloc_clear_buf() 847 memset_page(page, pgoff, 0, size); in binder_alloc_clear_buf() 1343 pgoff_t pgoff; in binder_alloc_copy_user_to_buffer() local 1347 buffer_offset, &pgoff); in binder_alloc_copy_user_to_buffer() 1349 kptr = kmap_local_page(page) + pgoff; in binder_alloc_copy_user_to_buffer() 1375 pgoff_t pgoff; in binder_alloc_do_buffer_copy() local 1378 buffer_offset, &pgoff); in binder_alloc_do_buffer_copy() 1381 memcpy_to_page(page, pgoff, ptr, size); in binder_alloc_do_buffer_copy() [all …]
|
| /drivers/iommu/iommufd/ |
| A D | iova_bitmap.c | 45 unsigned long pgoff; member 210 mapped->pgoff = offset_in_page(addr); in iova_bitmap_get() 307 bytes = (bitmap->mapped.npages << PAGE_SHIFT) - bitmap->mapped.pgoff; in iova_bitmap_mapped_remaining() 421 mapped->pgshift) + mapped->pgoff * BITS_PER_BYTE; in iova_bitmap_set() 423 mapped->pgshift) + mapped->pgoff * BITS_PER_BYTE; in iova_bitmap_set()
|
| A D | driver.c | 196 max_pgoff = max(max_pgoff, cur->pgoff + 1); in iommufd_sw_msi_get_map() 211 cur->pgoff = max_pgoff; in iommufd_sw_msi_get_map() 225 iova = msi_map->sw_msi_start + msi_map->pgoff * PAGE_SIZE; in iommufd_sw_msi_install() 295 iova = msi_map->sw_msi_start + msi_map->pgoff * PAGE_SIZE; in iommufd_sw_msi()
|
| /drivers/nvdimm/ |
| A D | pmem.c | 242 __weak long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, in __pmem_direct_access() argument 246 resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; in __pmem_direct_access() 247 sector_t sector = PFN_PHYS(pgoff) >> SECTOR_SHIFT; in __pmem_direct_access() 293 static int pmem_dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, in pmem_dax_zero_page_range() argument 299 PFN_PHYS(pgoff) >> SECTOR_SHIFT, in pmem_dax_zero_page_range() 304 pgoff_t pgoff, long nr_pages, enum dax_access_mode mode, in pmem_dax_direct_access() argument 309 return __pmem_direct_access(pmem, pgoff, nr_pages, mode, kaddr, pfn); in pmem_dax_direct_access() 325 static size_t pmem_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, in pmem_recovery_write() argument 336 if (!is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) >> SECTOR_SHIFT, len)) in pmem_recovery_write() 349 pmem_off = PFN_PHYS(pgoff) + pmem->data_offset; in pmem_recovery_write()
|
| /drivers/gpu/drm/omapdrm/ |
| A D | omap_gem.c | 357 pgoff_t pgoff; in omap_gem_fault_1d() local 360 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in omap_gem_fault_1d() 363 omap_gem_cpu_sync_page(obj, pgoff); in omap_gem_fault_1d() 364 pfn = page_to_pfn(omap_obj->pages[pgoff]); in omap_gem_fault_1d() 367 pfn = (omap_obj->dma_addr >> PAGE_SHIFT) + pgoff; in omap_gem_fault_1d() 386 pgoff_t pgoff, base_pgoff; in omap_gem_fault_2d() local 408 pgoff = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in omap_gem_fault_2d() 414 base_pgoff = round_down(pgoff, m << n_shift); in omap_gem_fault_2d() 435 int off = pgoff % m; in omap_gem_fault_2d() 700 if (omap_obj->dma_addrs[pgoff]) { in omap_gem_cpu_sync_page() [all …]
|
| /drivers/xen/ |
| A D | gntalloc.c | 83 uint16_t pgoff:12; /* Bits 0-11: Offset of the byte to clear */ member 180 tmp[gref->notify.pgoff] = 0; in __del_gref() 377 int pgoff; in gntalloc_ioctl_unmap_notify() local 384 pgoff = op.index & (PAGE_SIZE - 1); in gntalloc_ioctl_unmap_notify() 417 gref->notify.pgoff = pgoff; in gntalloc_ioctl_unmap_notify()
|
| /drivers/hwtracing/intel_th/ |
| A D | msu.c | 70 unsigned long pgoff; member 993 unsigned long pgoff) in msc_buffer_contig_get_page() argument 995 if (pgoff >= msc->nr_pages) in msc_buffer_contig_get_page() 1122 win->pgoff = prev->pgoff + prev->nr_blocks; in msc_buffer_win_alloc() 1413 return msc_buffer_contig_get_page(msc, pgoff); in msc_buffer_get_page() 1416 if (pgoff >= win->pgoff && pgoff < win->pgoff + win->nr_blocks) in msc_buffer_get_page() 1422 pgoff -= win->pgoff; in msc_buffer_get_page() 1428 if (pgoff < pgsz) in msc_buffer_get_page() 1429 return page + pgoff; in msc_buffer_get_page() 1431 pgoff -= pgsz; in msc_buffer_get_page() [all …]
|
| /drivers/char/ |
| A D | mem.c | 286 unsigned long pgoff, in get_unmapped_area_mem() argument 289 if (!valid_mmap_phys_addr_range(pgoff, len)) in get_unmapped_area_mem() 291 return pgoff << PAGE_SHIFT; in get_unmapped_area_mem() 517 unsigned long pgoff, unsigned long flags) in get_unmapped_area_zero() argument 527 return shmem_get_unmapped_area(NULL, addr, len, pgoff, flags); in get_unmapped_area_zero() 531 return mm_get_unmapped_area(current->mm, file, addr, len, pgoff, flags); in get_unmapped_area_zero()
|
| /drivers/vfio/cdx/ |
| A D | main.c | 239 u64 pgoff, base; in vfio_cdx_mmap_mmio() local 241 pgoff = vma->vm_pgoff & in vfio_cdx_mmap_mmio() 243 base = pgoff << PAGE_SHIFT; in vfio_cdx_mmap_mmio() 248 vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff; in vfio_cdx_mmap_mmio()
|
| /drivers/s390/block/ |
| A D | dcssblk.c | 33 static long dcssblk_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, 48 pgoff_t pgoff, size_t nr_pages) in dcssblk_dax_zero_page_range() argument 53 rc = dax_direct_access(dax_dev, pgoff, nr_pages, DAX_ACCESS, in dcssblk_dax_zero_page_range() 915 __dcssblk_direct_access(struct dcssblk_dev_info *dev_info, pgoff_t pgoff, in __dcssblk_direct_access() argument 918 resource_size_t offset = pgoff * PAGE_SIZE; in __dcssblk_direct_access() 931 dcssblk_dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, in dcssblk_dax_direct_access() argument 937 return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn); in dcssblk_dax_direct_access()
|
| /drivers/comedi/ |
| A D | comedi_buf.c | 159 unsigned int pgoff = offset_in_page(offset); in comedi_buf_map_access() local 164 int l = min_t(int, len - done, PAGE_SIZE - pgoff); in comedi_buf_map_access() 165 void *b = bm->page_list[pg].virt_addr + pgoff; in comedi_buf_map_access() 174 pgoff = 0; in comedi_buf_map_access()
|
| /drivers/vfio/fsl-mc/ |
| A D | vfio_fsl_mc.c | 363 u64 pgoff, base; in vfio_fsl_mc_mmap_mmio() local 366 pgoff = vma->vm_pgoff & in vfio_fsl_mc_mmap_mmio() 368 base = pgoff << PAGE_SHIFT; in vfio_fsl_mc_mmap_mmio() 378 vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff; in vfio_fsl_mc_mmap_mmio()
|
| /drivers/misc/ocxl/ |
| A D | sysfs.c | 115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT)) in global_mmio_fault() 118 offset = vmf->pgoff; in global_mmio_fault()
|