| /drivers/misc/lkdtm/ |
| A D | usercopy.c | 57 unsigned long user_addr; in do_usercopy_stack() local 81 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in do_usercopy_stack() 84 if (user_addr >= TASK_SIZE) { in do_usercopy_stack() 127 vm_munmap(user_addr, PAGE_SIZE); in do_usercopy_stack() 136 unsigned long user_addr; in do_usercopy_slab_size() local 152 if (user_addr >= TASK_SIZE) { in do_usercopy_slab_size() 192 vm_munmap(user_addr, PAGE_SIZE); in do_usercopy_slab_size() 206 unsigned char __user *user_addr; in do_usercopy_slab_whitelist() local 313 unsigned long user_addr; in lkdtm_USERCOPY_KERNEL() local 318 if (user_addr >= TASK_SIZE) { in lkdtm_USERCOPY_KERNEL() [all …]
|
| A D | perms.c | 208 unsigned long user_addr; in lkdtm_EXEC_USERSPACE() local 210 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in lkdtm_EXEC_USERSPACE() 213 if (user_addr >= TASK_SIZE) { in lkdtm_EXEC_USERSPACE() 217 execute_user_location((void *)user_addr); in lkdtm_EXEC_USERSPACE() 218 vm_munmap(user_addr, PAGE_SIZE); in lkdtm_EXEC_USERSPACE() 228 unsigned long user_addr, tmp = 0; in lkdtm_ACCESS_USERSPACE() local 231 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in lkdtm_ACCESS_USERSPACE() 234 if (user_addr >= TASK_SIZE) { in lkdtm_ACCESS_USERSPACE() 241 vm_munmap(user_addr, PAGE_SIZE); in lkdtm_ACCESS_USERSPACE() 245 ptr = (unsigned long *)user_addr; in lkdtm_ACCESS_USERSPACE() [all …]
|
| /drivers/fpga/ |
| A D | dfl-afu-dma-region.c | 51 pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE, in afu_dma_pin_pages() 302 u64 user_addr, u64 length, u64 *iova) in afu_dma_map_region() argument 312 if (!PAGE_ALIGNED(user_addr) || !PAGE_ALIGNED(length) || !length) in afu_dma_map_region() 316 if (user_addr + length < user_addr) in afu_dma_map_region() 323 region->user_addr = user_addr; in afu_dma_map_region()
|
| A D | dfl-afu.h | 54 u64 user_addr; member 96 u64 user_addr, u64 length, u64 *iova);
|
| A D | dfl-afu-main.c | 726 ret = afu_dma_map_region(fdata, map.user_addr, map.length, &map.iova); in afu_ioctl_dma_map() 736 (unsigned long long)map.user_addr, in afu_ioctl_dma_map()
|
| /drivers/misc/genwqe/ |
| A D | card_utils.c | 292 void __user *user_addr, size_t user_size, int write) in genwqe_alloc_sync_sgl() argument 297 sgl->fpage_offs = offset_in_page((unsigned long)user_addr); in genwqe_alloc_sync_sgl() 303 __func__, user_addr, user_size, sgl->nr_pages, in genwqe_alloc_sync_sgl() 306 sgl->user_addr = user_addr; in genwqe_alloc_sync_sgl() 334 user_addr, sgl->fpage_size)) { in genwqe_alloc_sync_sgl() 346 if (copy_from_user(sgl->lpage, user_addr + user_size - in genwqe_alloc_sync_sgl() 482 res = copy_to_user(sgl->user_addr, in genwqe_free_sync_sgl() 499 res = copy_to_user(sgl->user_addr + offset, sgl->lpage, in genwqe_free_sync_sgl()
|
| A D | card_base.h | 351 void __user *user_addr; /* user-space base-address */ member 369 void __user *user_addr, size_t user_size, int write);
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_queue.c | 200 u64 user_addr; in kfd_queue_buffer_get() local 203 user_addr = (u64)addr >> AMDGPU_GPU_PAGE_SHIFT; in kfd_queue_buffer_get() 206 mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr); in kfd_queue_buffer_get() 210 if (user_addr != mapping->start || in kfd_queue_buffer_get() 211 (size != 0 && user_addr + size - 1 != mapping->last)) { in kfd_queue_buffer_get()
|
| A D | kfd_chardev.c | 1772 uint8_t __user *user_addr, in criu_checkpoint_devices() argument 1804 ret = copy_to_user(user_addr, device_buckets, num_devices * sizeof(*device_buckets)); in criu_checkpoint_devices() 1953 &bo_priv->user_addr); in criu_checkpoint_bos() 2325 offset = bo_priv->user_addr; in criu_restore_memory_of_gpu()
|
| A D | kfd_priv.h | 1223 uint64_t user_addr; member
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_amdkfd_gpuvm.c | 1037 static int init_user_pages(struct kgd_mem *mem, uint64_t user_addr, in init_user_pages() argument 1048 ret = amdgpu_ttm_tt_set_userptr(&bo->tbo, user_addr, 0); in init_user_pages() 1054 ret = amdgpu_hmm_register(bo, user_addr); in init_user_pages() 1668 uint64_t user_addr = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local 1709 user_addr = untagged_addr(*offset); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1784 if (user_addr) in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1792 add_kgd_mem_to_kfd_bo_list(*mem, avm->process_info, user_addr); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1794 if (user_addr) { in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1795 pr_debug("creating userptr BO for user_addr = %llx\n", user_addr); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() 1796 ret = init_user_pages(*mem, user_addr, criu_resume); in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
|
| A D | amdgpu_ttm.h | 218 uint64_t *user_addr);
|
| A D | amdgpu_ttm.c | 1222 uint64_t *user_addr) in amdgpu_ttm_tt_get_userptr() argument 1230 *user_addr = gtt->userptr; in amdgpu_ttm_tt_get_userptr()
|
| /drivers/infiniband/hw/hns/ |
| A D | hns_roce_mr.c | 585 struct ib_udata *udata, unsigned long user_addr) in mtr_alloc_bufs() argument 594 mtr->umem = ib_umem_get(ibdev, user_addr, total_size, in mtr_alloc_bufs() 1038 unsigned long user_addr) in hns_roce_mtr_create() argument 1051 ret = mtr_alloc_bufs(hr_dev, mtr, buf_attr, udata, user_addr); in hns_roce_mtr_create()
|
| A D | hns_roce_device.h | 1190 unsigned long user_addr);
|
| /drivers/hv/ |
| A D | mshv_root_main.c | 641 unsigned long user_addr = (user_pfn + completed) * PAGE_SIZE; in mshv_vp_ioctl_get_set_state_pfn() local 644 ret = pin_user_pages_fast(user_addr, remaining, FOLL_WRITE, in mshv_vp_ioctl_get_set_state_pfn()
|