| /drivers/acpi/ |
| A D | nvs.c | 78 void *kaddr; member 138 if (entry->kaddr) { in suspend_nvs_free() 140 iounmap(entry->kaddr); in suspend_nvs_free() 143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free() 146 entry->kaddr = NULL; in suspend_nvs_free() 182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save() 183 if (!entry->kaddr) { in suspend_nvs_save() 184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save() 185 entry->unmap = !!entry->kaddr; in suspend_nvs_save() 187 if (!entry->kaddr) { in suspend_nvs_save() [all …]
|
| /drivers/misc/ |
| A D | open-dice.c | 45 void *kaddr; in open_dice_wipe() local 48 kaddr = devm_memremap(drvdata->misc.this_device, drvdata->rmem->base, in open_dice_wipe() 50 if (IS_ERR(kaddr)) { in open_dice_wipe() 52 return PTR_ERR(kaddr); in open_dice_wipe() 55 memset(kaddr, 0, drvdata->rmem->size); in open_dice_wipe() 56 devm_memunmap(drvdata->misc.this_device, kaddr); in open_dice_wipe()
|
| /drivers/tee/amdtee/ |
| A D | shm_pool.c | 26 shm->kaddr = (void *)va; in pool_op_alloc() 34 shm->kaddr = NULL; in pool_op_alloc() 45 free_pages((unsigned long)shm->kaddr, get_order(shm->size)); in pool_op_free() 46 shm->kaddr = NULL; in pool_op_free()
|
| A D | core.c | 163 if (shmdata->kaddr == shm->kaddr) { in get_buffer_id() 359 shmem.kaddr = shm->kaddr; in amdtee_map_shmem() 373 shmnode->kaddr = shm->kaddr; in amdtee_map_shmem() 380 pr_debug("buf_id :[%x] kaddr[%p]\n", shmnode->buf_id, shmnode->kaddr); in amdtee_map_shmem()
|
| A D | amdtee_private.h | 82 void *kaddr; member 93 void *kaddr; member
|
| A D | call.c | 302 if (!start[i].kaddr || (start[i].size & (PAGE_SIZE - 1))) { in handle_map_shmem() 307 if ((u64)start[i].kaddr & (PAGE_SIZE - 1)) { in handle_map_shmem() 309 (u64)start[i].kaddr); in handle_map_shmem() 319 paddr = __psp_pa(start[i].kaddr); in handle_map_shmem()
|
| /drivers/tee/ |
| A D | tee_shm.c | 222 shm->kaddr = alloc_pages_exact(nr_pages * PAGE_SIZE, in tee_dyn_shm_alloc_helper() 224 if (!shm->kaddr) in tee_dyn_shm_alloc_helper() 227 shm->paddr = virt_to_phys(shm->kaddr); in tee_dyn_shm_alloc_helper() 237 pages[i] = virt_to_page((u8 *)shm->kaddr + i * PAGE_SIZE); in tee_dyn_shm_alloc_helper() 244 (unsigned long)shm->kaddr); in tee_dyn_shm_alloc_helper() 251 free_pages_exact(shm->kaddr, shm->size); in tee_dyn_shm_alloc_helper() 252 shm->kaddr = NULL; in tee_dyn_shm_alloc_helper() 263 free_pages_exact(shm->kaddr, shm->size); in tee_dyn_shm_free_helper() 264 shm->kaddr = NULL; in tee_dyn_shm_free_helper() 499 if (!shm->kaddr) in tee_shm_get_va() [all …]
|
| A D | tee_shm_pool.c | 26 shm->kaddr = (void *)va; in pool_op_gen_alloc() 39 gen_pool_free(pool->private_data, (unsigned long)shm->kaddr, in pool_op_gen_free() 41 shm->kaddr = NULL; in pool_op_gen_free()
|
| /drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
| A D | smu8_smumgr.c | 333 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_scratch_task() 370 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_ucode_load_task() 409 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_initialize_toc_empty_job_list() 420 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_enter() 590 entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr + in smu8_smu_populate_single_scratch_entry() 781 &smu8_smu->toc_buffer.kaddr); in smu8_smu_init() 791 &smu8_smu->smu_buffer.kaddr); in smu8_smu_init() 839 &smu8_smu->smu_buffer.kaddr); in smu8_smu_init() 843 &smu8_smu->toc_buffer.kaddr); in smu8_smu_init() 860 &smu8_smu->toc_buffer.kaddr); in smu8_smu_fini() [all …]
|
| A D | smu7_smumgr.c | 375 memcpy_toio(smu_data->header_buffer.kaddr, smu_data->toc, in smu7_request_smu_load_fw() 517 &smu_data->header_buffer.kaddr); in smu7_init() 532 &smu_data->smu_buffer.kaddr); in smu7_init() 537 &smu_data->header_buffer.kaddr); in smu7_init() 555 &smu_data->header_buffer.kaddr); in smu7_smu_fini() 560 &smu_data->smu_buffer.kaddr); in smu7_smu_fini()
|
| A D | smu8_smumgr.h | 64 void *kaddr; member
|
| A D | smu7_smumgr.h | 35 void *kaddr; member
|
| /drivers/misc/lkdtm/ |
| A D | usercopy.c | 351 static void do_usercopy_page_span(const char *name, void *kaddr) in do_usercopy_page_span() argument 363 memset(kaddr, 0xAA, PAGE_SIZE); in do_usercopy_page_span() 366 kaddr += PAGE_SIZE / 2; in do_usercopy_page_span() 369 name, kaddr); in do_usercopy_page_span() 370 if (copy_to_user((void __user *)uaddr, kaddr, in do_usercopy_page_span() 377 name, kaddr); in do_usercopy_page_span() 378 if (copy_to_user((void __user *)uaddr, kaddr, unconst + PAGE_SIZE)) { in do_usercopy_page_span()
|
| /drivers/block/ |
| A D | brd.c | 116 void *kaddr; in brd_rw_bvec() local 128 kaddr = bvec_kmap_local(&bv); in brd_rw_bvec() 130 memcpy_to_page(page, offset, kaddr, bv.bv_len); in brd_rw_bvec() 133 memcpy_from_page(kaddr, page, offset, bv.bv_len); in brd_rw_bvec() 135 memset(kaddr, 0, bv.bv_len); in brd_rw_bvec() 137 kunmap_local(kaddr); in brd_rw_bvec()
|
| /drivers/iommu/iommufd/ |
| A D | iova_bitmap.c | 430 void *kaddr; in iova_bitmap_set() local 441 kaddr = kmap_local_page(mapped->pages[page_idx]); in iova_bitmap_set() 442 bitmap_set(kaddr, offset, nbits); in iova_bitmap_set() 443 kunmap_local(kaddr); in iova_bitmap_set()
|
| /drivers/s390/block/ |
| A D | dcssblk.c | 34 long nr_pages, enum dax_access_mode mode, void **kaddr, 51 void *kaddr; in dcssblk_dax_zero_page_range() local 54 &kaddr, NULL); in dcssblk_dax_zero_page_range() 58 memset(kaddr, 0, nr_pages << PAGE_SHIFT); in dcssblk_dax_zero_page_range() 59 dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT); in dcssblk_dax_zero_page_range() 916 long nr_pages, void **kaddr, unsigned long *pfn) in __dcssblk_direct_access() argument 922 if (kaddr) in __dcssblk_direct_access() 923 *kaddr = __va(dev_info->start + offset); in __dcssblk_direct_access() 932 long nr_pages, enum dax_access_mode mode, void **kaddr, in dcssblk_dax_direct_access() argument 937 return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn); in dcssblk_dax_direct_access()
|
| /drivers/pci/ |
| A D | p2pdma.c | 100 void *kaddr; in p2pmem_alloc_mmap() local 125 kaddr = (void *)gen_pool_alloc_owner(p2pdma->pool, len, (void **)&ref); in p2pmem_alloc_mmap() 126 if (!kaddr) { in p2pmem_alloc_mmap() 143 struct page *page = virt_to_page(kaddr); in p2pmem_alloc_mmap() 154 gen_pool_free(p2pdma->pool, (uintptr_t)kaddr, len); in p2pmem_alloc_mmap() 159 kaddr += PAGE_SIZE; in p2pmem_alloc_mmap() 167 gen_pool_free(p2pdma->pool, (uintptr_t)kaddr, len); in p2pmem_alloc_mmap()
|
| /drivers/fpga/ |
| A D | stratix10-soc.c | 120 static void s10_unlock_bufs(struct s10_priv *priv, void *kaddr) in s10_unlock_bufs() argument 124 if (!kaddr) in s10_unlock_bufs() 128 if (priv->svc_bufs[i].buf == kaddr) { in s10_unlock_bufs() 134 WARN(1, "Unknown buffer returned from service layer %p\n", kaddr); in s10_unlock_bufs()
|
| /drivers/scsi/isci/ |
| A D | request.c | 1409 void *kaddr; in sci_stp_request_pio_data_in_copy_data_buffer() local 1423 kaddr = kmap_atomic(page); in sci_stp_request_pio_data_in_copy_data_buffer() 1425 kunmap_atomic(kaddr); in sci_stp_request_pio_data_in_copy_data_buffer() 1765 void *frame_header, *kaddr; in sci_io_request_frame_handler() local 1772 rsp = kaddr + sg->offset; in sci_io_request_frame_handler() 1808 kunmap_atomic(kaddr); in sci_io_request_frame_handler() 2912 void *kaddr; in isci_request_io_request_complete() local 2920 kunmap_atomic(kaddr); in isci_request_io_request_complete() 3182 void *kaddr; in sci_io_request_construct_smp() local 3187 smp_req = kaddr + sg->offset; in sci_io_request_construct_smp() [all …]
|
| /drivers/nvdimm/ |
| A D | pmem.c | 243 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument 253 if (kaddr) in __pmem_direct_access() 254 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access() 305 void **kaddr, unsigned long *pfn) in pmem_dax_direct_access() argument 309 return __pmem_direct_access(pmem, pgoff, nr_pages, mode, kaddr, pfn); in pmem_dax_direct_access()
|
| A D | pmem.h | 31 long nr_pages, enum dax_access_mode mode, void **kaddr,
|
| /drivers/md/ |
| A D | dm-linear.c | 172 long nr_pages, enum dax_access_mode mode, void **kaddr, in linear_dax_direct_access() argument 177 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access()
|
| A D | dm-stripe.c | 318 long nr_pages, enum dax_access_mode mode, void **kaddr, in stripe_dax_direct_access() argument 323 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in stripe_dax_direct_access()
|
| /drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| A D | chtls_hw.c | 358 int kaddr; in chtls_setkey() local 386 kaddr = keyid_to_addr(cdev->kmap.start, keyid); in chtls_setkey() 405 kwr->req.lock_addr = cpu_to_be32(ULP_MEMIO_ADDR_V(kaddr)); in chtls_setkey()
|
| /drivers/infiniband/sw/siw/ |
| A D | siw_qp_tx.c | 500 void *kaddr; in siw_tx_hdt() local 515 void *kaddr = kmap_local_page(p); in siw_tx_hdt() local 519 iov[seg].iov_base = kaddr + fp_off; in siw_tx_hdt() 528 kaddr = kmap_local_page(p); in siw_tx_hdt() 530 kaddr + fp_off, plen); in siw_tx_hdt() 531 kunmap_local(kaddr); in siw_tx_hdt()
|