Home
last modified time | relevance | path

Searched refs:vm_start (Results 1 – 25 of 124) sorted by relevance

12345

/drivers/android/
A Dbinder_alloc.c204 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_lru_freelist_add()
341 alloc->pid, addr - alloc->vm_start); in binder_install_single_page()
375 index = (page_addr - alloc->vm_start) / PAGE_SIZE; in binder_install_buffer_pages()
757 BUG_ON(buffer->user_data < alloc->vm_start); in binder_free_buf_locked()
817 (buffer->user_data - alloc->vm_start); in binder_alloc_get_page()
917 alloc->vm_start = vma->vm_start; in binder_alloc_mmap_handler()
935 buffer->user_data = alloc->vm_start; in binder_alloc_mmap_handler()
950 alloc->vm_start = 0; in binder_alloc_mmap_handler()
958 alloc->pid, vma->vm_start, vma->vm_end, in binder_alloc_mmap_handler()
1052 buffer->user_data - alloc->vm_start, in binder_alloc_print_allocated()
[all …]
/drivers/accel/habanalabs/common/
A Dmemory_mgr.c205 new_mmap_size = buf->real_mapped_size - (vma->vm_end - vma->vm_start); in hl_mmap_mem_buf_vm_close()
253 user_mem_size = vma->vm_end - vma->vm_start;
263 if (!access_ok(VERIFY_WRITE, (void __user *)(uintptr_t)vma->vm_start,
266 if (!access_ok((void __user *)(uintptr_t)vma->vm_start,
270 buf->behavior->topic, vma->vm_start);
/drivers/gpu/drm/i915/gem/
A Di915_gem_mman.c34 return vma->vm_start == addr && in __vma_matches()
35 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches()
281 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu()
304 unsigned long vm_start, vm_end, vma_size; /* user's memory parameters */ in set_address_limits() local
311 vm_start = area->vm_start >> PAGE_SHIFT; in set_address_limits()
320 start = vm_start; in set_address_limits()
325 start = max_t(long, start, vm_start); in set_address_limits()
333 *pfn += (*start_vaddr - area->vm_start) >> PAGE_SHIFT; in set_address_limits()
359 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; in vm_fault_gtt()
519 addr -= area->vm_start; in vm_access()
/drivers/soc/qcom/
A Drmtfs_mem.c136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap()
139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap()
140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap()
146 vma->vm_start, in qcom_rmtfs_mem_mmap()
148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
/drivers/xen/xenfs/
A Dxenstored.c36 size_t size = vma->vm_end - vma->vm_start; in xsd_kva_mmap()
41 if (remap_pfn_range(vma, vma->vm_start, in xsd_kva_mmap()
/drivers/sbus/char/
A Dflash.c66 if (vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)) > size) in flash_mmap()
67 size = vma->vm_end - (vma->vm_start + (vma->vm_pgoff << PAGE_SHIFT)); in flash_mmap()
71 if (io_remap_pfn_range(vma, vma->vm_start, addr, size, vma->vm_page_prot)) in flash_mmap()
/drivers/media/common/videobuf2/
A Dvideobuf2-memops.c95 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_open()
113 __func__, h, refcount_read(h->refcount), vma->vm_start, in vb2_common_vm_close()
/drivers/pci/
A Dmmap.c49 return io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in pci_mmap_resource_range()
50 vma->vm_end - vma->vm_start, in pci_mmap_resource_range()
/drivers/xen/xenbus/
A Dxenbus_dev_backend.c94 size_t size = vma->vm_end - vma->vm_start; in xenbus_backend_mmap()
102 if (remap_pfn_range(vma, vma->vm_start, in xenbus_backend_mmap()
/drivers/gpu/drm/amd/amdkfd/
A Dkfd_doorbell.c116 if (vma->vm_end - vma->vm_start != kfd_doorbell_process_slice(dev->kfd)) in kfd_doorbell_mmap()
137 (unsigned long long) vma->vm_start, address, vma->vm_flags, in kfd_doorbell_mmap()
142 vma->vm_start, in kfd_doorbell_mmap()
A Dkfd_events.c1074 get_order(vma->vm_end - vma->vm_start)) { in kfd_event_mmap()
1093 pr_debug(" start user address == 0x%08lx\n", vma->vm_start); in kfd_event_mmap()
1098 vma->vm_end - vma->vm_start); in kfd_event_mmap()
1100 page->user_address = (uint64_t __user *)vma->vm_start; in kfd_event_mmap()
1103 ret = remap_pfn_range(vma, vma->vm_start, pfn, in kfd_event_mmap()
1104 vma->vm_end - vma->vm_start, vma->vm_page_prot); in kfd_event_mmap()
1106 p->signal_mapped_size = vma->vm_end - vma->vm_start; in kfd_event_mmap()
/drivers/infiniband/hw/hfi1/
A Dfile_ops.c315 vma->vm_end - vma->vm_start, vma->vm_flags); in mmap_cdbg()
417 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap()
433 vm_start_save = vma->vm_start; in hfi1_file_mmap()
435 vma->vm_end = vma->vm_start; in hfi1_file_mmap()
446 vma->vm_start = vm_start_save; in hfi1_file_mmap()
450 vma->vm_start += memlen; in hfi1_file_mmap()
452 vma->vm_start = vm_start_save; in hfi1_file_mmap()
556 (vma->vm_end - vma->vm_start), memlen); in hfi1_file_mmap()
572 ret = io_remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap()
577 ret = remap_pfn_range(vma, vma->vm_start, in hfi1_file_mmap()
[all …]
/drivers/char/
A Duv_mmtimer.c151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) in uv_mmtimer_mmap()
166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, in uv_mmtimer_mmap()
A Dbsr.c116 unsigned long size = vma->vm_end - vma->vm_start; in bsr_mmap()
124 ret = remap_4k_pfn(vma, vma->vm_start, dev->bsr_addr >> 12, in bsr_mmap()
127 ret = io_remap_pfn_range(vma, vma->vm_start, in bsr_mmap()
/drivers/infiniband/core/
A Dib_core_uverbs.c77 if (vma->vm_end - vma->vm_start != size) in rdma_user_mmap_io()
91 if (io_remap_pfn_range(vma, vma->vm_start, pfn, size, prot)) { in rdma_user_mmap_io()
171 if (entry->npages * PAGE_SIZE != vma->vm_end - vma->vm_start) { in rdma_user_mmap_entry_get()
/drivers/accel/amdxdna/
A Damdxdna_gem.c112 mapp->vma->vm_start, mapp->vma->vm_end, abo->type); in amdxdna_hmm_invalidate()
197 unsigned long len = vma->vm_end - vma->vm_start; in amdxdna_hmm_register()
198 unsigned long addr = vma->vm_start; in amdxdna_hmm_register()
229 mapp->range.start = vma->vm_start; in amdxdna_hmm_register()
286 ret = vm_insert_pages(vma, vma->vm_start, abo->base.pages, in amdxdna_insert_pages()
308 fault_ret = handle_mm_fault(vma, vma->vm_start + offset, in amdxdna_insert_pages()
344 vma->vm_start, gobj->size); in amdxdna_gem_obj_mmap()
369 ret = vm_insert_pages(vma, vma->vm_start, abo->base.pages, in amdxdna_gem_dmabuf_mmap()
/drivers/infiniband/hw/cxgb4/
A Dprovider.c130 int len = vma->vm_end - vma->vm_start; in c4iw_mmap()
146 if (vma->vm_start & (PAGE_SIZE-1)) in c4iw_mmap()
164 ret = io_remap_pfn_range(vma, vma->vm_start, addr >> PAGE_SHIFT, in c4iw_mmap()
169 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
174 ret = io_remap_pfn_range(vma, vma->vm_start, in c4iw_mmap()
/drivers/dax/
A Ddevice.c34 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma()
37 current->comm, func, vma->vm_start, vma->vm_end, in check_vma()
160 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault()
206 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault()
/drivers/gpu/drm/ttm/
A Dttm_bo_vm.c212 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + in ttm_bo_vm_fault_reserved()
314 for (address = vma->vm_start; address < vma->vm_end; in ttm_bo_vm_dummy_page()
461 unsigned long offset = (addr) - vma->vm_start + in ttm_bo_vm_access()
/drivers/vfio/fsl-mc/
A Dvfio_fsl_mc.c362 u64 size = vma->vm_end - vma->vm_start; in vfio_fsl_mc_mmap_mmio()
380 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in vfio_fsl_mc_mmap_mmio()
394 if (vma->vm_end < vma->vm_start) in vfio_fsl_mc_mmap()
396 if (vma->vm_start & ~PAGE_MASK) in vfio_fsl_mc_mmap()
/drivers/uio/
A Duio.c754 if (vma->vm_end - vma->vm_start > mem->size) in uio_mmap_physical()
771 vma->vm_start, in uio_mmap_physical()
773 vma->vm_end - vma->vm_start, in uio_mmap_physical()
797 if (vma->vm_end - vma->vm_start > mem->size) in uio_mmap_dma_coherent()
814 vma->vm_end - vma->vm_start); in uio_mmap_dma_coherent()
828 if (vma->vm_end < vma->vm_start) in uio_mmap()
/drivers/vfio/platform/
A Dvfio_platform_common.c557 req_len = vma->vm_end - vma->vm_start; in vfio_platform_mmap_mmio()
568 return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in vfio_platform_mmap_mmio()
580 if (vma->vm_end < vma->vm_start) in vfio_platform_mmap()
586 if (vma->vm_start & ~PAGE_MASK) in vfio_platform_mmap()
/drivers/gpu/drm/virtio/
A Dvirtgpu_vram.c39 unsigned long vm_size = vma->vm_end - vma->vm_start; in virtio_gpu_vram_mmap()
66 ret = io_remap_pfn_range(vma, vma->vm_start, in virtio_gpu_vram_mmap()
/drivers/misc/genwqe/
A Dcard_dev.c392 unsigned long vsize = vma->vm_end - vma->vm_start; in genwqe_vma_close()
401 dma_map = __genwqe_search_mapping(cfile, vma->vm_start, vsize, in genwqe_vma_close()
406 __func__, vma->vm_start, vma->vm_pgoff << PAGE_SHIFT, in genwqe_vma_close()
438 unsigned long pfn, vsize = vma->vm_end - vma->vm_start; in genwqe_mmap()
454 dma_map->u_vaddr = (void *)vma->vm_start; in genwqe_mmap()
469 vma->vm_start, in genwqe_mmap()
/drivers/soc/aspeed/
A Daspeed-lpc-ctrl.c52 unsigned long vsize = vma->vm_end - vma->vm_start; in aspeed_lpc_ctrl_mmap()
61 if (remap_pfn_range(vma, vma->vm_start, in aspeed_lpc_ctrl_mmap()

Completed in 55 milliseconds

12345