| /drivers/android/ |
| A D | binder_alloc.c | 204 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_lru_freelist_add() 341 alloc->pid, addr - alloc->vm_start); in binder_install_single_page() 375 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_install_buffer_pages() 757 BUG_ON(buffer->user_data < alloc->vm_start); in binder_free_buf_locked() 817 (buffer->user_data - alloc->vm_start); in binder_alloc_get_page() 917 alloc->vm_start = vma->vm_start; in binder_alloc_mmap_handler() 935 buffer->user_data = alloc->vm_start; in binder_alloc_mmap_handler() 950 alloc->vm_start = 0; in binder_alloc_mmap_handler() 958 alloc->pid, vma->vm_start, vma->vm_end, in binder_alloc_mmap_handler() 1052 buffer->user_data - alloc->vm_start, in binder_alloc_print_allocated() [all …]
|
| /drivers/accel/habanalabs/common/ |
| A D | memory_mgr.c | 205 new_mmap_size = buf->real_mapped_size - (vma->vm_end - vma->vm_start); in hl_mmap_mem_buf_vm_close() 253 user_mem_size = vma->vm_end - vma->vm_start; 263 if (!access_ok(VERIFY_WRITE, (void __user *)(uintptr_t)vma->vm_start, 266 if (!access_ok((void __user *)(uintptr_t)vma->vm_start, 270 buf->behavior->topic, vma->vm_start);
|
| /drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_mman.c | 34 return vma->vm_start == addr && in __vma_matches() 35 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches() 281 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu() 304 unsigned long vm_start, vm_end, vma_size; /* user's memory parameters */ in set_address_limits() local 311 vm_start = area->vm_start >> PAGE_SHIFT; in set_address_limits() 320 start = vm_start; in set_address_limits() 325 start = max_t(long, start, vm_start); in set_address_limits() 333 *pfn += (*start_vaddr - area->vm_start) >> PAGE_SHIFT; in set_address_limits() 359 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; in vm_fault_gtt() 519 addr -= area->vm_start; in vm_access()
|
| /drivers/soc/qcom/ |
| A D | rmtfs_mem.c | 136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap() 139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap() 140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap() 146 vma->vm_start, in qcom_rmtfs_mem_mmap() 148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
|
| /drivers/xen/xenfs/ |
| A D | xenstored.c | 36 size_t size = vma->vm_end - vma->vm_start; in xsd_kva_mmap() 41 if (remap_pfn_range(vma, vma->vm_start, in xsd_kva_mmap()
|
| /drivers/sbus/char/ |
| A D | flash.c | 66 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) in flash_mmap() 67 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); in flash_mmap() 71 if (io_remap_pfn_range(vma, vma->vm_start, addr, size, vma->vm_page_prot)) in flash_mmap()
|
| /drivers/media/common/videobuf2/ |
| A D | videobuf2-memops.c | 95 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_open() 113 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_close()
|
| /drivers/pci/ |
| A D | mmap.c | 49 return io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_resource_range() 50 vma->vm_end - vma->vm_start, in pci_mmap_resource_range()
|
| /drivers/xen/xenbus/ |
| A D | xenbus_dev_backend.c | 94 size_t size = vma->vm_end - vma->vm_start; in xenbus_backend_mmap() 102 if (remap_pfn_range(vma, vma->vm_start, in xenbus_backend_mmap()
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_doorbell.c | 116 if (vma->vm_end - vma->vm_start != kfd_doorbell_process_slice(dev->kfd)) in kfd_doorbell_mmap() 137 (unsigned long long) vma->vm_start, address, vma->vm_flags, in kfd_doorbell_mmap() 142 vma->vm_start, in kfd_doorbell_mmap()
|
| A D | kfd_events.c | 1074 get_order(vma->vm_end - vma->vm_start)) { in kfd_event_mmap() 1093 pr_debug(" start user address == 0x%08lx\n", vma->vm_start); in kfd_event_mmap() 1098 vma->vm_end - vma->vm_start); in kfd_event_mmap() 1100 page->user_address = (uint64_t __user *)vma->vm_start; in kfd_event_mmap() 1103 ret = remap_pfn_range(vma, vma->vm_start, pfn, in kfd_event_mmap() 1104 vma->vm_end - vma->vm_start, vma->vm_page_prot); in kfd_event_mmap() 1106 p->signal_mapped_size = vma->vm_end - vma->vm_start; in kfd_event_mmap()
|
| /drivers/infiniband/hw/hfi1/ |
| A D | file_ops.c | 315 vma->vm_end - vma->vm_start, vma->vm_flags); in mmap_cdbg() 417 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap() 433 vm_start_save = vma->vm_start; in hfi1_file_mmap() 435 vma->vm_end = vma->vm_start; in hfi1_file_mmap() 446 vma->vm_start = vm_start_save; in hfi1_file_mmap() 450 vma->vm_start += memlen; in hfi1_file_mmap() 452 vma->vm_start = vm_start_save; in hfi1_file_mmap() 556 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap() 572 ret = io_remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap() 577 ret = remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap() [all …]
|
| /drivers/char/ |
| A D | uv_mmtimer.c | 151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) in uv_mmtimer_mmap() 166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, in uv_mmtimer_mmap()
|
| A D | bsr.c | 116 unsigned long size = vma->vm_end - vma->vm_start; in bsr_mmap() 124 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, in bsr_mmap() 127 ret = io_remap_pfn_range(vma, vma->vm_start, in bsr_mmap()
|
| /drivers/infiniband/core/ |
| A D | ib_core_uverbs.c | 77 if (vma->vm_end - vma->vm_start != size) in rdma_user_mmap_io() 91 if (io_remap_pfn_range(vma, vma->vm_start, pfn, size, prot)) { in rdma_user_mmap_io() 171 if (entry->npages * PAGE_SIZE != vma->vm_end - vma->vm_start) { in rdma_user_mmap_entry_get()
|
| /drivers/accel/amdxdna/ |
| A D | amdxdna_gem.c | 112 mapp->vma->vm_start, mapp->vma->vm_end, abo->type); in amdxdna_hmm_invalidate() 197 unsigned long len = vma->vm_end - vma->vm_start; in amdxdna_hmm_register() 198 unsigned long addr = vma->vm_start; in amdxdna_hmm_register() 229 mapp->range.start = vma->vm_start; in amdxdna_hmm_register() 286 ret = vm_insert_pages(vma, vma->vm_start, abo->base.pages, in amdxdna_insert_pages() 308 fault_ret = handle_mm_fault(vma, vma->vm_start + offset, in amdxdna_insert_pages() 344 vma->vm_start, gobj->size); in amdxdna_gem_obj_mmap() 369 ret = vm_insert_pages(vma, vma->vm_start, abo->base.pages, in amdxdna_gem_dmabuf_mmap()
|
| /drivers/infiniband/hw/cxgb4/ |
| A D | provider.c | 130 int len = vma->vm_end - vma->vm_start; in c4iw_mmap() 146 if (vma->vm_start & (PAGE_SIZE-1)) in c4iw_mmap() 164 ret = io_remap_pfn_range(vma, vma->vm_start, addr >> PAGE_SHIFT, in c4iw_mmap() 169 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap() 174 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
|
| /drivers/dax/ |
| A D | device.c | 34 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma() 37 current->comm, func, vma->vm_start, vma->vm_end, in check_vma() 160 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault() 206 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault()
|
| /drivers/gpu/drm/ttm/ |
| A D | ttm_bo_vm.c | 212 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + in ttm_bo_vm_fault_reserved() 314 for (address = vma->vm_start; address < vma->vm_end; in ttm_bo_vm_dummy_page() 461 unsigned long offset = (addr) - vma->vm_start + in ttm_bo_vm_access()
|
| /drivers/vfio/fsl-mc/ |
| A D | vfio_fsl_mc.c | 362 u64 size = vma->vm_end - vma->vm_start; in vfio_fsl_mc_mmap_mmio() 380 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in vfio_fsl_mc_mmap_mmio() 394 if (vma->vm_end < vma->vm_start) in vfio_fsl_mc_mmap() 396 if (vma->vm_start & ~PAGE_MASK) in vfio_fsl_mc_mmap()
|
| /drivers/uio/ |
| A D | uio.c | 754 if (vma->vm_end - vma->vm_start > mem->size) in uio_mmap_physical() 771 vma->vm_start, in uio_mmap_physical() 773 vma->vm_end - vma->vm_start, in uio_mmap_physical() 797 if (vma->vm_end - vma->vm_start > mem->size) in uio_mmap_dma_coherent() 814 vma->vm_end - vma->vm_start); in uio_mmap_dma_coherent() 828 if (vma->vm_end < vma->vm_start) in uio_mmap()
|
| /drivers/vfio/platform/ |
| A D | vfio_platform_common.c | 557 req_len = vma->vm_end - vma->vm_start; in vfio_platform_mmap_mmio() 568 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in vfio_platform_mmap_mmio() 580 if (vma->vm_end < vma->vm_start) in vfio_platform_mmap() 586 if (vma->vm_start & ~PAGE_MASK) in vfio_platform_mmap()
|
| /drivers/gpu/drm/virtio/ |
| A D | virtgpu_vram.c | 39 unsigned long vm_size = vma->vm_end - vma->vm_start; in virtio_gpu_vram_mmap() 66 ret = io_remap_pfn_range(vma, vma->vm_start, in virtio_gpu_vram_mmap()
|
| /drivers/misc/genwqe/ |
| A D | card_dev.c | 392 unsigned long vsize = vma->vm_end - vma->vm_start; in genwqe_vma_close() 401 dma_map = __genwqe_search_mapping(cfile, vma->vm_start, vsize, in genwqe_vma_close() 406 __func__, vma->vm_start, vma->vm_pgoff << PAGE_SHIFT, in genwqe_vma_close() 438 unsigned long pfn, vsize = vma->vm_end - vma->vm_start; in genwqe_mmap() 454 dma_map->u_vaddr = (void *)vma->vm_start; in genwqe_mmap() 469 vma->vm_start, in genwqe_mmap()
|
| /drivers/soc/aspeed/ |
| A D | aspeed-lpc-ctrl.c | 52 unsigned long vsize = vma->vm_end - vma->vm_start; in aspeed_lpc_ctrl_mmap() 61 if (remap_pfn_range(vma, vma->vm_start, in aspeed_lpc_ctrl_mmap()
|