| /linux/drivers/gpu/drm/nouveau/ |
| A D | nouveau_bo85b5.c | 45 u64 src_offset = mem->vma[0].addr; in nva3_bo_move_copy() local 58 PUSH_NVSQ(push, NV85B5, 0x030c, upper_32_bits(src_offset), in nva3_bo_move_copy() 59 0x0310, lower_32_bits(src_offset), in nva3_bo_move_copy() 69 src_offset += (PAGE_SIZE * line_count); in nva3_bo_move_copy()
|
| A D | nouveau_bo90b5.c | 38 u64 src_offset = mem->vma[0].addr; in nvc0_bo_move_copy() local 51 PUSH_NVSQ(push, NV90B5, 0x030c, upper_32_bits(src_offset), in nvc0_bo_move_copy() 52 0x0310, lower_32_bits(src_offset), in nvc0_bo_move_copy() 62 src_offset += (PAGE_SIZE * line_count); in nvc0_bo_move_copy()
|
| A D | nouveau_bo9039.c | 43 u64 src_offset = mem->vma[0].addr; in nvc0_bo_move_m2mf() local 62 NVVAL(NV9039, OFFSET_IN_UPPER, VALUE, upper_32_bits(src_offset)), in nvc0_bo_move_m2mf() 64 OFFSET_IN, lower_32_bits(src_offset), in nvc0_bo_move_m2mf() 79 src_offset += (PAGE_SIZE * line_count); in nvc0_bo_move_m2mf()
|
| A D | nouveau_bo5039.c | 45 u64 src_offset = mem->vma[0].addr; in nv50_bo_move_m2mf() local 107 NVVAL(NV5039, OFFSET_IN_UPPER, VALUE, upper_32_bits(src_offset)), in nv50_bo_move_m2mf() 112 PUSH_MTHD(push, NV5039, OFFSET_IN, lower_32_bits(src_offset), in nv50_bo_move_m2mf() 129 src_offset += amount; in nv50_bo_move_m2mf()
|
| A D | nouveau_bo0039.c | 52 u32 src_offset = old_reg->start << PAGE_SHIFT; in nv04_bo_move_m2mf() local 73 PUSH_MTHD(push, NV039, OFFSET_IN, src_offset, in nv04_bo_move_m2mf() 89 src_offset += (PAGE_SIZE * line_count); in nv04_bo_move_m2mf()
|
| /linux/drivers/staging/media/atomisp/pci/isp/kernels/sdis/sdis_2/ |
| A D | ia_css_sdis2.host.c | 210 unsigned int src_offset = 0, dst_offset = 0; in ia_css_translate_dvs2_statistics() local 247 &htemp_ptr[0 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 249 &htemp_ptr[1 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 251 &htemp_ptr[2 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 253 &htemp_ptr[3 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 257 &vtemp_ptr[0 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 259 &vtemp_ptr[1 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 261 &vtemp_ptr[2 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 263 &vtemp_ptr[3 * table_size + src_offset], size_bytes); in ia_css_translate_dvs2_statistics() 265 src_offset += table_width; /* aligned table width */ in ia_css_translate_dvs2_statistics()
|
| /linux/drivers/gpu/drm/radeon/ |
| A D | rv770_dma.c | 43 uint64_t src_offset, uint64_t dst_offset, in rv770_copy_dma() argument 76 radeon_ring_write(ring, src_offset & 0xfffffffc); in rv770_copy_dma() 78 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); in rv770_copy_dma() 79 src_offset += cur_size_in_dw * 4; in rv770_copy_dma()
|
| A D | evergreen_dma.c | 107 uint64_t src_offset, in evergreen_copy_dma() argument 141 radeon_ring_write(ring, src_offset & 0xfffffffc); in evergreen_copy_dma() 143 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); in evergreen_copy_dma() 144 src_offset += cur_size_in_dw * 4; in evergreen_copy_dma()
|
| A D | si_dma.c | 231 uint64_t src_offset, uint64_t dst_offset, in si_copy_dma() argument 264 radeon_ring_write(ring, lower_32_bits(src_offset)); in si_copy_dma() 266 radeon_ring_write(ring, upper_32_bits(src_offset) & 0xff); in si_copy_dma() 267 src_offset += cur_size_in_bytes; in si_copy_dma()
|
| A D | evergreen_cs.c | 2805 u64 src_offset, dst_offset, dst2_offset; in evergreen_dma_cs_parse() local 2870 src_offset = radeon_get_ib_value(p, idx+2); in evergreen_dma_cs_parse() 2895 src_offset = radeon_get_ib_value(p, idx+1); in evergreen_dma_cs_parse() 2896 src_offset <<= 8; in evergreen_dma_cs_parse() 2905 src_offset = radeon_get_ib_value(p, idx+7); in evergreen_dma_cs_parse() 2929 src_offset = radeon_get_ib_value(p, idx+2); in evergreen_dma_cs_parse() 2975 src_offset = radeon_get_ib_value(p, idx+3); in evergreen_dma_cs_parse() 3015 src_offset = radeon_get_ib_value(p, idx+8); in evergreen_dma_cs_parse() 3077 src_offset = radeon_get_ib_value(p, idx+8); in evergreen_dma_cs_parse() 3107 src_offset <<= 8; in evergreen_dma_cs_parse() [all …]
|
| A D | r600_dma.c | 444 uint64_t src_offset, uint64_t dst_offset, in r600_copy_dma() argument 477 radeon_ring_write(ring, src_offset & 0xfffffffc); in r600_copy_dma() 479 (upper_32_bits(src_offset) & 0xff))); in r600_copy_dma() 480 src_offset += cur_size_in_dw * 4; in r600_copy_dma()
|
| /linux/crypto/async_tx/ |
| A D | async_memcpy.c | 33 unsigned int src_offset, size_t len, in async_memcpy() argument 45 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy() 54 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy() 79 src_buf = kmap_atomic(src) + src_offset; in async_memcpy()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| A D | vmwgfx_blit.c | 361 u32 src_offset, in vmw_bo_cpu_blit_line() argument 369 u32 src_page = src_offset >> PAGE_SHIFT; in vmw_bo_cpu_blit_line() 371 u32 src_page_offset = src_offset & ~PAGE_MASK; in vmw_bo_cpu_blit_line() 419 src_offset += copy_size; in vmw_bo_cpu_blit_line() 458 u32 src_offset, u32 src_stride, in vmw_external_bo_copy() argument 487 vsrc += src_offset; in vmw_external_bo_copy() 491 src_size -= src_offset; in vmw_external_bo_copy() 544 u32 src_offset, u32 src_stride, in vmw_bo_cpu_blit() argument 585 vmw_src, src_offset, src_stride, in vmw_bo_cpu_blit() 626 ret = vmw_bo_cpu_blit_line(&d, dst_offset, src_offset, w); in vmw_bo_cpu_blit() [all …]
|
| /linux/fs/bcachefs/ |
| A D | io_misc.c | 315 prt_printf(out, " src_offset=%llu", le64_to_cpu(op.v->src_offset)); in bch2_logged_op_finsert_to_text() 363 u64 src_offset = le64_to_cpu(op->v.src_offset); in __bch2_resume_logged_op_finsert() local 364 s64 shift = dst_offset - src_offset; in __bch2_resume_logged_op_finsert() 394 adjust_i_size(trans, inum, src_offset, len, warn_errors) ?: in __bch2_resume_logged_op_finsert() 399 bch2_btree_iter_set_pos(&iter, POS(inum.inum, src_offset)); in __bch2_resume_logged_op_finsert() 401 ret = bch2_fpunch_at(trans, &iter, inum, src_offset + len, i_sectors_delta); in __bch2_resume_logged_op_finsert() 416 struct bpos src_pos = POS(inum.inum, src_offset); in __bch2_resume_logged_op_finsert() 436 bkey_le(k.k->p, POS(inum.inum, src_offset))) in __bch2_resume_logged_op_finsert() 484 adjust_i_size(trans, inum, src_offset, shift, warn_errors) ?: in __bch2_resume_logged_op_finsert() 520 op.v.src_offset = cpu_to_le64(offset); in bch2_fcollapse_finsert()
|
| A D | logged_ops_format.h | 26 __le64 src_offset; member
|
| /linux/drivers/gpu/drm/vc4/ |
| A D | vc4_validate.c | 496 uint32_t src_offset = 0; in vc4_validate_bin_cl() local 501 while (src_offset < len) { in vc4_validate_bin_cl() 503 void *src_pkt = unvalidated + src_offset; in vc4_validate_bin_cl() 509 src_offset, cmd); in vc4_validate_bin_cl() 516 src_offset, cmd); in vc4_validate_bin_cl() 520 if (src_offset + info->len > len) { in vc4_validate_bin_cl() 523 src_offset, cmd, info->name, info->len, in vc4_validate_bin_cl() 524 src_offset + len); in vc4_validate_bin_cl() 535 src_offset, cmd, info->name); in vc4_validate_bin_cl() 539 src_offset += info->len; in vc4_validate_bin_cl() [all …]
|
| /linux/drivers/gpu/drm/i915/gt/ |
| A D | intel_migrate.c | 566 *cs++ = src_offset; in emit_copy_ccs() 601 *cs++ = src_offset; in emit_copy() 612 *cs++ = src_offset; in emit_copy() 621 *cs++ = src_offset; in emit_copy() 692 u32 src_offset, dst_offset; in intel_context_migrate_copy() local 739 src_offset = 0; in intel_context_migrate_copy() 742 src_offset = 0; in intel_context_migrate_copy() 745 src_offset = CHUNK_SZ; in intel_context_migrate_copy() 782 src_offset, src_sz); in intel_context_migrate_copy() 834 src_offset, src_access, len); in intel_context_migrate_copy() [all …]
|
| /linux/include/linux/ |
| A D | async_tx.h | 167 struct page **src_list, unsigned int *src_offset, 177 struct page **src_list, unsigned int *src_offset, 183 unsigned int src_offset, size_t len,
|
| A D | iosys-map.h | 307 size_t src_offset, size_t len) in iosys_map_memcpy_from() argument 310 memcpy_fromio(dst, src->vaddr_iomem + src_offset, len); in iosys_map_memcpy_from() 312 memcpy(dst, src->vaddr + src_offset, len); in iosys_map_memcpy_from()
|
| /linux/drivers/gpu/drm/qxl/ |
| A D | qxl_ioctl.c | 76 int src_offset; member 92 info->src_offset); in apply_reloc() 239 reloc_info[i].src_offset = reloc.src_offset; in qxl_process_single_command() 242 reloc_info[i].src_offset = 0; in qxl_process_single_command()
|
| /linux/drivers/gpu/drm/xe/ |
| A D | xe_map.h | 33 size_t src_offset, size_t len) in xe_map_memcpy_from() argument 36 iosys_map_memcpy_from(dst, src, src_offset, len); in xe_map_memcpy_from()
|
| /linux/tools/testing/selftests/sgx/ |
| A D | load.c | 182 off_t src_offset; in encl_load() local 270 src_offset = phdr->p_offset & PAGE_MASK; in encl_load() 271 encl->src = encl->bin + src_offset; in encl_load() 282 seg->offset = (phdr->p_offset & PAGE_MASK) - src_offset; in encl_load()
|
| /linux/fs/smb/client/ |
| A D | trace.h | 220 __u64 src_offset, 231 __field(__u64, src_offset) 242 __entry->src_offset = src_offset; 259 __u64 src_offset, \ 274 __u64 src_offset, 277 TP_ARGS(xid, src_fid, target_fid, tid, sesid, src_offset, target_offset, len), 284 __field(__u64, src_offset) 294 __entry->src_offset = src_offset; 300 __entry->src_offset, __entry->target_fid, __entry->target_offset, __entry->len) 310 __u64 src_offset, \ [all …]
|
| /linux/fs/btrfs/ |
| A D | extent_io.h | 336 unsigned long dst_offset, unsigned long src_offset, 339 unsigned long dst_offset, unsigned long src_offset, 342 unsigned long dst_offset, unsigned long src_offset,
|
| /linux/drivers/crypto/ccp/ |
| A D | ccp-dmaengine.c | 366 unsigned int src_offset, src_len; in ccp_create_desc() local 385 src_offset = 0; in ccp_create_desc() 401 src_offset = 0; in ccp_create_desc() 433 ccp_pt->src_dma = sg_dma_address(src_sg) + src_offset; in ccp_create_desc() 450 src_offset += len; in ccp_create_desc()
|