| /drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| A D | chtls_hw.c | 157 cdev->kmap.size = num_key_ctx; in chtls_init_kmap() 158 cdev->kmap.available = bsize; in chtls_init_kmap() 161 if (!cdev->kmap.addr) in chtls_init_kmap() 165 spin_lock_init(&cdev->kmap.lock); in chtls_init_kmap() 180 spin_lock_bh(&cdev->kmap.lock); in get_new_keyid() 181 keyid = find_first_zero_bit(cdev->kmap.addr, cdev->kmap.size); in get_new_keyid() 182 if (keyid < cdev->kmap.size) { in get_new_keyid() 192 spin_unlock_bh(&cdev->kmap.lock); in get_new_keyid() 204 if (!cdev->kmap.addr) in free_tls_keyid() 210 spin_lock_bh(&cdev->kmap.lock); in free_tls_keyid() [all …]
|
| A D | chtls.h | 240 struct key_map kmap; member
|
| /drivers/gpu/drm/panthor/ |
| A D | panthor_gem.h | 142 void *kmap; member 180 if (bo->kmap) in panthor_kernel_bo_vmap() 187 bo->kmap = map.vaddr; in panthor_kernel_bo_vmap() 194 if (bo->kmap) { in panthor_kernel_bo_vunmap() 195 struct iosys_map map = IOSYS_MAP_INIT_VADDR(bo->kmap); in panthor_kernel_bo_vunmap() 198 bo->kmap = NULL; in panthor_kernel_bo_vunmap()
|
| A D | panthor_fw.c | 401 bool was_mapped = !!section->mem->kmap; in panthor_fw_init_section_mem() 412 memcpy(section->mem->kmap, section->data.buf, section->data.size); in panthor_fw_init_section_mem() 414 memset(section->mem->kmap + section->data.size, 0, in panthor_fw_init_section_mem() 463 memset(mem->kmap, 0, panthor_kernel_bo_size(mem)); in panthor_fw_alloc_queue_iface_mem() 464 *input = mem->kmap; in panthor_fw_alloc_queue_iface_mem() 465 *output = mem->kmap + SZ_4K; in panthor_fw_alloc_queue_iface_mem() 818 return ptdev->fw->shared_section->mem->kmap + (mcu_va - shared_mem_start); in iface_fw_to_cpu_addr() 839 cs_iface->control = ptdev->fw->shared_section->mem->kmap + iface_offset; in panthor_init_cs_iface() 890 csg_iface->control = ptdev->fw->shared_section->mem->kmap + iface_offset; in panthor_init_csg_iface() 938 if (!ptdev->fw->shared_section->mem->kmap) in panthor_fw_init_ifaces() [all …]
|
| A D | panthor_sched.c | 442 void *kmap; member 840 if (queue->syncwait.kmap) { in panthor_queue_put_syncwait_obj() 844 queue->syncwait.kmap = NULL; in panthor_queue_put_syncwait_obj() 859 if (queue->syncwait.kmap) in panthor_queue_get_syncwait_obj() 860 return queue->syncwait.kmap + queue->syncwait.offset; in panthor_queue_get_syncwait_obj() 873 queue->syncwait.kmap = map.vaddr; in panthor_queue_get_syncwait_obj() 874 if (drm_WARN_ON(&ptdev->base, !queue->syncwait.kmap)) in panthor_queue_get_syncwait_obj() 877 return queue->syncwait.kmap + queue->syncwait.offset; in panthor_queue_get_syncwait_obj() 2086 syncobj = group->syncobjs->kmap + (i * sizeof(*syncobj)); in group_term_post_processing() 2970 memcpy(queue->ringbuf->kmap, in copy_instrs_to_ringbuf() [all …]
|
| A D | panthor_heap.c | 120 return pool->gpu_contexts->kmap + in panthor_get_heap_ctx() 165 hdr = chunk->bo->kmap; in panthor_alloc_heap_chunk()
|
| /drivers/gpu/drm/nouveau/ |
| A D | nouveau_chan.c | 174 chan->chan.push.mem.object.map.ptr = chan->push.buffer->kmap.virtual; in nouveau_channel_prep() 440 (u8*)chan->push.buffer->kmap.virtual + 0x10000, 0x2000, in nouveau_channel_init() 441 chan->push.buffer->kmap.virtual, chan->push.addr, 0x10000); in nouveau_channel_init() 447 (u8*)chan->push.buffer->kmap.virtual + 0x10000, 0x2000, in nouveau_channel_init() 448 chan->push.buffer->kmap.virtual, chan->push.addr, 0x10000, in nouveau_channel_init() 449 chan->sema.bo->kmap.virtual, chan->sema.vma->addr); in nouveau_channel_init() 454 (u8*)chan->push.buffer->kmap.virtual + 0x10000, 0x2000, in nouveau_channel_init() 455 chan->push.buffer->kmap.virtual, chan->push.addr, 0x10000, in nouveau_channel_init() 456 chan->sema.bo->kmap.virtual, chan->sema.vma->addr, in nouveau_channel_init()
|
| A D | nouveau_bo.h | 20 struct ttm_bo_kmap_obj kmap; member 105 &nvbo->kmap, &is_iomem); in nvbo_kmap_obj_iovirtual()
|
| A D | nouveau_gem.c | 435 ttm_bo_kunmap(&nvbo->kmap); in validate_fini_no_ticket() 696 if (!nvbo->kmap.virtual) { in nouveau_gem_pushbuf_reloc_apply() 698 &nvbo->kmap); in nouveau_gem_pushbuf_reloc_apply() 901 if (!nvbo->kmap.virtual) { in nouveau_gem_ioctl_pushbuf() 904 &nvbo->kmap); in nouveau_gem_ioctl_pushbuf()
|
| A D | nouveau_bo.c | 676 ret = ttm_bo_kmap(&nvbo->bo, 0, PFN_UP(nvbo->bo.base.size), &nvbo->kmap); in nouveau_bo_map() 688 ttm_bo_kunmap(&nvbo->kmap); in nouveau_bo_unmap() 803 u16 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); in nouveau_bo_wr16() 817 u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); in nouveau_bo_rd32() 831 u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); in nouveau_bo_wr32()
|
| /drivers/gpu/drm/nouveau/include/nvkm/core/ |
| A D | memory.h | 41 int (*kmap)(struct nvkm_memory *, struct nvkm_memory **); member 68 #define nvkm_memory_kmap(p,i) ((p)->func->kmap ? (p)->func->kmap((p), (i)) : -ENOSYS)
|
| /drivers/gpu/drm/xe/ |
| A D | xe_bo_types.h | 50 struct ttm_bo_kmap_obj kmap; member
|
| /drivers/gpu/drm/loongson/ |
| A D | lsdc_ttm.c | 370 err = ttm_bo_kmap(tbo, 0, PFN_UP(lsdc_bo_size(lbo)), &lbo->kmap); in lsdc_bo_kmap() 376 lbo->kptr = ttm_kmap_obj_virtual(&lbo->kmap, &lbo->is_iomem); in lsdc_bo_kmap() 387 ttm_bo_kunmap(&lbo->kmap); in lsdc_bo_kunmap()
|
| A D | lsdc_ttm.h | 35 struct ttm_bo_kmap_obj kmap; member
|
| /drivers/gpu/drm/radeon/ |
| A D | radeon_object.c | 234 r = ttm_bo_kmap(&bo->tbo, 0, PFN_UP(bo->tbo.base.size), &bo->kmap); in radeon_bo_kmap() 238 bo->kptr = ttm_kmap_obj_virtual(&bo->kmap, &is_iomem); in radeon_bo_kmap() 252 ttm_bo_kunmap(&bo->kmap); in radeon_bo_kunmap()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_object.c | 838 r = ttm_bo_kmap(&bo->tbo, 0, PFN_UP(bo->tbo.base.size), &bo->kmap); in amdgpu_bo_kmap() 861 return ttm_kmap_obj_virtual(&bo->kmap, &is_iomem); in amdgpu_bo_kptr() 872 if (bo->kmap.bo) in amdgpu_bo_kunmap() 873 ttm_bo_kunmap(&bo->kmap); in amdgpu_bo_kunmap()
|
| A D | amdgpu_display.c | 1861 if (!abo->kmap.virtual && in amdgpu_display_get_scanout_buffer() 1862 ttm_bo_kmap(&abo->tbo, 0, PFN_UP(abo->tbo.base.size), &abo->kmap)) { in amdgpu_display_get_scanout_buffer() 1866 if (abo->kmap.bo_kmap_type & TTM_BO_MAP_IOMEM_MASK) in amdgpu_display_get_scanout_buffer() 1867 iosys_map_set_vaddr_iomem(&sb->map[0], abo->kmap.virtual); in amdgpu_display_get_scanout_buffer() 1869 iosys_map_set_vaddr(&sb->map[0], abo->kmap.virtual); in amdgpu_display_get_scanout_buffer()
|
| A D | amdgpu_object.h | 108 struct ttm_bo_kmap_obj kmap; member
|
| /drivers/mtd/ |
| A D | mtd_blkdevs.c | 67 buf = kmap(bio_page(req->bio)) + bio_offset(req->bio); in do_blktrans_request() 86 buf = kmap(bio_page(req->bio)) + bio_offset(req->bio); in do_blktrans_request()
|
| /drivers/firmware/efi/ |
| A D | capsule-loader.c | 200 kbuff = kmap(page); in efi_capsule_write()
|
| /drivers/gpu/drm/nouveau/dispnv50/ |
| A D | wndw.c | 676 if (nvbo->kmap.bo_kmap_type & TTM_BO_MAP_IOMEM_MASK) in nv50_wndw_get_scanout_buffer() 677 iosys_map_set_vaddr_iomem(&sb->map[0], (void __iomem *)nvbo->kmap.virtual); in nv50_wndw_get_scanout_buffer() 679 iosys_map_set_vaddr(&sb->map[0], nvbo->kmap.virtual); in nv50_wndw_get_scanout_buffer()
|
| /drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
| A D | ram.c | 116 .kmap = nvkm_vram_kmap,
|
| /drivers/mmc/host/ |
| A D | sdricoh_cs.c | 307 buf = kmap(page) + data->sg->offset + (len * i); in sdricoh_request()
|
| A D | usdhi6rol0.c | 327 host->pg.mapped = kmap(host->pg.page); in usdhi6_blk_bounce() 369 host->pg.mapped = kmap(host->pg.page); in usdhi6_sg_map() 507 host->pg.mapped = kmap(host->pg.page); in usdhi6_sg_advance()
|
| /drivers/gpu/drm/i915/ |
| A D | i915_gem.c | 212 vaddr = kmap(page); in shmem_pread() 649 vaddr = kmap(page); in shmem_pwrite()
|