| /drivers/gpu/drm/radeon/ |
| A D | si_dma.c | 80 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pages() 81 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pages() 120 ib->ptr[ib->length_dw++] = pe; in si_dma_vm_write_pages() 132 ib->ptr[ib->length_dw++] = value; in si_dma_vm_write_pages() 133 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_write_pages() 172 ib->ptr[ib->length_dw++] = pe; /* dst addr */ in si_dma_vm_set_pages() 174 ib->ptr[ib->length_dw++] = flags; /* mask */ in si_dma_vm_set_pages() 175 ib->ptr[ib->length_dw++] = 0; in si_dma_vm_set_pages() 176 ib->ptr[ib->length_dw++] = value; /* value */ in si_dma_vm_set_pages() 177 ib->ptr[ib->length_dw++] = upper_32_bits(value); in si_dma_vm_set_pages() [all …]
|
| A D | radeon_vce.c | 362 ib.length_dw = 0; in radeon_vce_get_create_msg() 365 ib.ptr[ib.length_dw++] = cpu_to_le32(handle); in radeon_vce_get_create_msg() 383 ib.ptr[ib.length_dw++] = cpu_to_le32(dummy); in radeon_vce_get_create_msg() 386 for (i = ib.length_dw; i < ib_size_dw; ++i) in radeon_vce_get_create_msg() 429 ib.length_dw = 0; in radeon_vce_get_destroy_msg() 432 ib.ptr[ib.length_dw++] = cpu_to_le32(handle); in radeon_vce_get_destroy_msg() 437 ib.ptr[ib.length_dw++] = cpu_to_le32(dummy); in radeon_vce_get_destroy_msg() 443 for (i = ib.length_dw; i < ib_size_dw; ++i) in radeon_vce_get_destroy_msg() 481 if (idx >= relocs_chunk->length_dw) { in radeon_vce_cs_reloc() 483 idx, relocs_chunk->length_dw); in radeon_vce_cs_reloc() [all …]
|
| A D | ni_dma.c | 328 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cayman_dma_vm_copy_pages() 329 ib->ptr[ib->length_dw++] = lower_32_bits(src); in cayman_dma_vm_copy_pages() 369 ib->ptr[ib->length_dw++] = pe; in cayman_dma_vm_write_pages() 381 ib->ptr[ib->length_dw++] = value; in cayman_dma_vm_write_pages() 421 ib->ptr[ib->length_dw++] = pe; /* dst addr */ in cayman_dma_vm_set_pages() 423 ib->ptr[ib->length_dw++] = flags; /* mask */ in cayman_dma_vm_set_pages() 424 ib->ptr[ib->length_dw++] = 0; in cayman_dma_vm_set_pages() 425 ib->ptr[ib->length_dw++] = value; /* value */ in cayman_dma_vm_set_pages() 426 ib->ptr[ib->length_dw++] = upper_32_bits(value); in cayman_dma_vm_set_pages() 428 ib->ptr[ib->length_dw++] = 0; in cayman_dma_vm_set_pages() [all …]
|
| A D | radeon_cs.c | 95 p->nrelocs = chunk->length_dw / 4; in radeon_cs_parser_relocs() 314 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init() 321 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 327 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 337 size = p->chunks[i].length_dw; in radeon_cs_parser_init() 358 if (p->chunks[i].length_dw > 1) in radeon_cs_parser_init() 360 if (p->chunks[i].length_dw > 2) in radeon_cs_parser_init() 563 if (parser->const_ib.length_dw) { in radeon_cs_ib_vm_chunk() 640 parser->const_ib.length_dw = ib_chunk->length_dw; in radeon_cs_ib_fill() 661 parser->ib.length_dw = ib_chunk->length_dw; in radeon_cs_ib_fill() [all …]
|
| A D | cik_sdma.c | 156 radeon_ring_write(ring, ib->length_dw); in cik_sdma_ring_ib_execute() 731 ib.length_dw = 5; in cik_sdma_ib_test() 814 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pages() 818 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in cik_sdma_vm_copy_pages() 857 ib->ptr[ib->length_dw++] = pe; in cik_sdma_vm_write_pages() 859 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pages() 870 ib->ptr[ib->length_dw++] = value; in cik_sdma_vm_write_pages() 912 ib->ptr[ib->length_dw++] = flags; /* mask */ in cik_sdma_vm_set_pages() 913 ib->ptr[ib->length_dw++] = 0; in cik_sdma_vm_set_pages() 917 ib->ptr[ib->length_dw++] = 0; in cik_sdma_vm_set_pages() [all …]
|
| A D | radeon_vm.c | 410 ib.length_dw = 0; in radeon_vm_clear_bo() 414 WARN_ON(ib.length_dw > 64); in radeon_vm_clear_bo() 662 ib.length_dw = 0; in radeon_vm_update_page_directory() 699 if (ib.length_dw != 0) { in radeon_vm_update_page_directory() 703 WARN_ON(ib.length_dw > ndw); in radeon_vm_update_page_directory() 1000 ib.length_dw = 0; in radeon_vm_bo_update() 1018 WARN_ON(ib.length_dw > ndw); in radeon_vm_bo_update()
|
| A D | radeon_uvd.c | 574 if (idx >= relocs_chunk->length_dw) { in radeon_uvd_cs_reloc() 576 idx, relocs_chunk->length_dw); in radeon_uvd_cs_reloc() 689 if (p->chunk_ib->length_dw % 16) { in radeon_uvd_cs_parse() 691 p->chunk_ib->length_dw); in radeon_uvd_cs_parse() 719 } while (p->idx < p->chunk_ib->length_dw); in radeon_uvd_cs_parse() 750 ib.length_dw = 16; in radeon_uvd_send_msg()
|
| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_vce.c | 469 ib->length_dw = 0; in amdgpu_vce_get_create_msg() 472 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_create_msg() 479 ib->ptr[ib->length_dw++] = 0x00000000; in amdgpu_vce_get_create_msg() 480 ib->ptr[ib->length_dw++] = 0x00000042; in amdgpu_vce_get_create_msg() 481 ib->ptr[ib->length_dw++] = 0x0000000a; in amdgpu_vce_get_create_msg() 499 ib->ptr[ib->length_dw++] = addr; in amdgpu_vce_get_create_msg() 550 ib->length_dw = 0; in amdgpu_vce_get_destroy_msg() 553 ib->ptr[ib->length_dw++] = handle; in amdgpu_vce_get_destroy_msg() 753 for (idx = 0; idx < ib->length_dw;) { in amdgpu_vce_ring_parse_cs() 819 for (idx = 0; idx < ib->length_dw;) { in amdgpu_vce_ring_parse_cs() [all …]
|
| A D | si_dma.c | 286 ib.length_dw = 4; in si_dma_ring_test_ib() 330 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_copy_pte() 331 ib->ptr[ib->length_dw++] = lower_32_bits(src); in si_dma_vm_copy_pte() 354 ib->ptr[ib->length_dw++] = lower_32_bits(pe); in si_dma_vm_write_pte() 355 ib->ptr[ib->length_dw++] = upper_32_bits(pe); in si_dma_vm_write_pte() 357 ib->ptr[ib->length_dw++] = lower_32_bits(value); in si_dma_vm_write_pte() 395 ib->ptr[ib->length_dw++] = pe; /* dst addr */ in si_dma_vm_set_pte_pde() 399 ib->ptr[ib->length_dw++] = value; /* value */ in si_dma_vm_set_pte_pde() 402 ib->ptr[ib->length_dw++] = 0; in si_dma_vm_set_pte_pde() 418 while (ib->length_dw & 0x7) in si_dma_ring_pad_ib() [all …]
|
| A D | sdma_v2_4.c | 262 amdgpu_ring_write(ring, ib->length_dw); in sdma_v2_4_ring_emit_ib() 616 ib.length_dw = 8; in sdma_v2_4_ring_test_ib() 661 ib->ptr[ib->length_dw++] = bytes; in sdma_v2_4_vm_copy_pte() 688 ib->ptr[ib->length_dw++] = pe; in sdma_v2_4_vm_write_pte() 690 ib->ptr[ib->length_dw++] = ndw; in sdma_v2_4_vm_write_pte() 723 ib->ptr[ib->length_dw++] = 0; in sdma_v2_4_vm_set_pte_pde() 740 pad_count = (-ib->length_dw) & 7; in sdma_v2_4_ring_pad_ib() 743 ib->ptr[ib->length_dw++] = in sdma_v2_4_ring_pad_ib() 747 ib->ptr[ib->length_dw++] = in sdma_v2_4_ring_pad_ib() 1189 ib->ptr[ib->length_dw++] = byte_count; in sdma_v2_4_emit_copy_buffer() [all …]
|
| A D | cik_sdma.c | 236 amdgpu_ring_write(ring, ib->length_dw); in cik_sdma_ring_emit_ib() 681 ib.length_dw = 5; in cik_sdma_ring_test_ib() 725 ib->ptr[ib->length_dw++] = bytes; in cik_sdma_vm_copy_pte() 754 ib->ptr[ib->length_dw++] = ndw; in cik_sdma_vm_write_pte() 787 ib->ptr[ib->length_dw++] = 0; in cik_sdma_vm_set_pte_pde() 804 pad_count = (-ib->length_dw) & 7; in cik_sdma_ring_pad_ib() 807 ib->ptr[ib->length_dw++] = in cik_sdma_ring_pad_ib() 811 ib->ptr[ib->length_dw++] = in cik_sdma_ring_pad_ib() 1294 ib->ptr[ib->length_dw++] = byte_count; in cik_sdma_emit_copy_buffer() 1320 ib->ptr[ib->length_dw++] = src_data; in cik_sdma_emit_fill_buffer() [all …]
|
| A D | sdma_v3_0.c | 438 amdgpu_ring_write(ring, ib->length_dw); in sdma_v3_0_ring_emit_ib() 890 ib.length_dw = 8; in sdma_v3_0_ring_test_ib() 934 ib->ptr[ib->length_dw++] = bytes; in sdma_v3_0_vm_copy_pte() 963 ib->ptr[ib->length_dw++] = ndw; in sdma_v3_0_vm_write_pte() 996 ib->ptr[ib->length_dw++] = 0; in sdma_v3_0_vm_set_pte_pde() 1013 pad_count = (-ib->length_dw) & 7; in sdma_v3_0_ring_pad_ib() 1016 ib->ptr[ib->length_dw++] = in sdma_v3_0_ring_pad_ib() 1020 ib->ptr[ib->length_dw++] = in sdma_v3_0_ring_pad_ib() 1631 ib->ptr[ib->length_dw++] = byte_count; in sdma_v3_0_emit_copy_buffer() 1657 ib->ptr[ib->length_dw++] = src_data; in sdma_v3_0_emit_fill_buffer() [all …]
|
| A D | amdgpu_vcn.c | 619 ib->length_dw = 16; in amdgpu_vcn_dec_send_msg() 740 ib->ptr[ib->length_dw++] = 0x30000002; in amdgpu_vcn_unified_ring_ib_header() 745 ib->ptr[ib->length_dw++] = 0x30000001; in amdgpu_vcn_unified_ring_ib_header() 789 ib->length_dw = 0; in amdgpu_vcn_dec_sw_send_msg() 921 ib->length_dw = 0; in amdgpu_vcn_enc_get_create_msg() 926 ib->ptr[ib->length_dw++] = 0x00000018; in amdgpu_vcn_enc_get_create_msg() 928 ib->ptr[ib->length_dw++] = handle; in amdgpu_vcn_enc_get_create_msg() 930 ib->ptr[ib->length_dw++] = addr; in amdgpu_vcn_enc_get_create_msg() 988 ib->length_dw = 0; in amdgpu_vcn_enc_get_destroy_msg() 995 ib->ptr[ib->length_dw++] = handle; in amdgpu_vcn_enc_get_destroy_msg() [all …]
|
| A D | sdma_v7_0.c | 291 amdgpu_ring_write(ring, ib->length_dw); in sdma_v7_0_ring_emit_ib() 1034 ib.length_dw = 8; in sdma_v7_0_ring_test_ib() 1086 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v7_0_vm_copy_pte() 1092 ib->ptr[ib->length_dw++] = 0; in sdma_v7_0_vm_copy_pte() 1117 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v7_0_vm_write_pte() 1151 ib->ptr[ib->length_dw++] = 0; in sdma_v7_0_vm_set_pte_pde() 1169 pad_count = (-ib->length_dw) & 0x7; in sdma_v7_0_ring_pad_ib() 1172 ib->ptr[ib->length_dw++] = in sdma_v7_0_ring_pad_ib() 1176 ib->ptr[ib->length_dw++] = in sdma_v7_0_ring_pad_ib() 1792 ib->ptr[ib->length_dw++] = 0; in sdma_v7_0_emit_copy_buffer() [all …]
|
| A D | uvd_v6_0.c | 227 ib->length_dw = 0; in uvd_v6_0_enc_get_create_msg() 228 ib->ptr[ib->length_dw++] = 0x00000018; in uvd_v6_0_enc_get_create_msg() 230 ib->ptr[ib->length_dw++] = handle; in uvd_v6_0_enc_get_create_msg() 231 ib->ptr[ib->length_dw++] = 0x00010000; in uvd_v6_0_enc_get_create_msg() 233 ib->ptr[ib->length_dw++] = addr; in uvd_v6_0_enc_get_create_msg() 235 ib->ptr[ib->length_dw++] = 0x00000014; in uvd_v6_0_enc_get_create_msg() 237 ib->ptr[ib->length_dw++] = 0x0000001c; in uvd_v6_0_enc_get_create_msg() 238 ib->ptr[ib->length_dw++] = 0x00000001; in uvd_v6_0_enc_get_create_msg() 291 ib->length_dw = 0; in uvd_v6_0_enc_get_destroy_msg() 294 ib->ptr[ib->length_dw++] = handle; in uvd_v6_0_enc_get_destroy_msg() [all …]
|
| A D | sdma_v6_0.c | 287 amdgpu_ring_write(ring, ib->length_dw); in sdma_v6_0_ring_emit_ib() 1017 ib.length_dw = 8; in sdma_v6_0_ring_test_ib() 1067 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v6_0_vm_copy_pte() 1097 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v6_0_vm_write_pte() 1131 ib->ptr[ib->length_dw++] = 0; in sdma_v6_0_vm_set_pte_pde() 1148 pad_count = (-ib->length_dw) & 0x7; in sdma_v6_0_ring_pad_ib() 1151 ib->ptr[ib->length_dw++] = in sdma_v6_0_ring_pad_ib() 1155 ib->ptr[ib->length_dw++] = in sdma_v6_0_ring_pad_ib() 1845 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v6_0_emit_copy_buffer() 1871 ib->ptr[ib->length_dw++] = src_data; in sdma_v6_0_emit_fill_buffer() [all …]
|
| A D | sdma_v5_2.c | 301 amdgpu_ring_write(ring, ib->length_dw); in sdma_v5_2_ring_emit_ib() 1011 ib.length_dw = 8; in sdma_v5_2_ring_test_ib() 1061 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v5_2_vm_copy_pte() 1091 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v5_2_vm_write_pte() 1125 ib->ptr[ib->length_dw++] = 0; in sdma_v5_2_vm_set_pte_pde() 1143 pad_count = (-ib->length_dw) & 0x7; in sdma_v5_2_ring_pad_ib() 1146 ib->ptr[ib->length_dw++] = in sdma_v5_2_ring_pad_ib() 1150 ib->ptr[ib->length_dw++] = in sdma_v5_2_ring_pad_ib() 2037 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v5_2_emit_copy_buffer() 2063 ib->ptr[ib->length_dw++] = src_data; in sdma_v5_2_emit_fill_buffer() [all …]
|
| A D | sdma_v5_0.c | 453 amdgpu_ring_write(ring, ib->length_dw); in sdma_v5_0_ring_emit_ib() 1112 ib.length_dw = 8; in sdma_v5_0_ring_test_ib() 1162 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v5_0_vm_copy_pte() 1192 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v5_0_vm_write_pte() 1226 ib->ptr[ib->length_dw++] = 0; in sdma_v5_0_vm_set_pte_pde() 1243 pad_count = (-ib->length_dw) & 0x7; in sdma_v5_0_ring_pad_ib() 1246 ib->ptr[ib->length_dw++] = in sdma_v5_0_ring_pad_ib() 1250 ib->ptr[ib->length_dw++] = in sdma_v5_0_ring_pad_ib() 2027 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v5_0_emit_copy_buffer() 2053 ib->ptr[ib->length_dw++] = src_data; in sdma_v5_0_emit_fill_buffer() [all …]
|
| A D | uvd_v7_0.c | 235 ib->length_dw = 0; in uvd_v7_0_enc_get_create_msg() 236 ib->ptr[ib->length_dw++] = 0x00000018; in uvd_v7_0_enc_get_create_msg() 238 ib->ptr[ib->length_dw++] = handle; in uvd_v7_0_enc_get_create_msg() 239 ib->ptr[ib->length_dw++] = 0x00000000; in uvd_v7_0_enc_get_create_msg() 241 ib->ptr[ib->length_dw++] = addr; in uvd_v7_0_enc_get_create_msg() 243 ib->ptr[ib->length_dw++] = 0x00000014; in uvd_v7_0_enc_get_create_msg() 245 ib->ptr[ib->length_dw++] = 0x0000001c; in uvd_v7_0_enc_get_create_msg() 246 ib->ptr[ib->length_dw++] = 0x00000000; in uvd_v7_0_enc_get_create_msg() 298 ib->length_dw = 0; in uvd_v7_0_enc_get_destroy_msg() 301 ib->ptr[ib->length_dw++] = handle; in uvd_v7_0_enc_get_destroy_msg() [all …]
|
| A D | sdma_v4_0.c | 822 amdgpu_ring_write(ring, ib->length_dw); in sdma_v4_0_ring_emit_ib() 1548 ib.length_dw = 8; in sdma_v4_0_ring_test_ib() 1594 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v4_0_vm_copy_pte() 1624 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v4_0_vm_write_pte() 1658 ib->ptr[ib->length_dw++] = 0; in sdma_v4_0_vm_set_pte_pde() 1674 pad_count = (-ib->length_dw) & 7; in sdma_v4_0_ring_pad_ib() 1677 ib->ptr[ib->length_dw++] = in sdma_v4_0_ring_pad_ib() 1681 ib->ptr[ib->length_dw++] = in sdma_v4_0_ring_pad_ib() 2569 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v4_0_emit_copy_buffer() 2595 ib->ptr[ib->length_dw++] = src_data; in sdma_v4_0_emit_fill_buffer() [all …]
|
| A D | sdma_v4_4_2.c | 389 amdgpu_ring_write(ring, ib->length_dw); in sdma_v4_4_2_ring_emit_ib() 1144 ib.length_dw = 8; in sdma_v4_4_2_ring_test_ib() 1190 ib->ptr[ib->length_dw++] = bytes - 1; in sdma_v4_4_2_vm_copy_pte() 1220 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v4_4_2_vm_write_pte() 1254 ib->ptr[ib->length_dw++] = 0; in sdma_v4_4_2_vm_set_pte_pde() 1270 pad_count = (-ib->length_dw) & 7; in sdma_v4_4_2_ring_pad_ib() 1273 ib->ptr[ib->length_dw++] = in sdma_v4_4_2_ring_pad_ib() 1277 ib->ptr[ib->length_dw++] = in sdma_v4_4_2_ring_pad_ib() 2270 ib->ptr[ib->length_dw++] = byte_count - 1; in sdma_v4_4_2_emit_copy_buffer() 2296 ib->ptr[ib->length_dw++] = src_data; in sdma_v4_4_2_emit_fill_buffer() [all …]
|
| A D | gfx_v9_4_2.c | 377 ib->length_dw = 0; in gfx_v9_4_2_run_shader() 384 ib->ptr[ib->length_dw++] = init_regs[i].reg_value; in gfx_v9_4_2_run_shader() 392 ib->ptr[ib->length_dw++] = lower_32_bits(gpu_addr); in gfx_v9_4_2_run_shader() 393 ib->ptr[ib->length_dw++] = upper_32_bits(gpu_addr); in gfx_v9_4_2_run_shader() 399 ib->ptr[ib->length_dw++] = lower_32_bits(wb_gpu_addr); in gfx_v9_4_2_run_shader() 400 ib->ptr[ib->length_dw++] = upper_32_bits(wb_gpu_addr); in gfx_v9_4_2_run_shader() 401 ib->ptr[ib->length_dw++] = pattern; in gfx_v9_4_2_run_shader() 405 ib->ptr[ib->length_dw++] = compute_dim_x; /* x */ in gfx_v9_4_2_run_shader() 406 ib->ptr[ib->length_dw++] = 1; /* y */ in gfx_v9_4_2_run_shader() 407 ib->ptr[ib->length_dw++] = 1; /* z */ in gfx_v9_4_2_run_shader() [all …]
|
| A D | amdgpu_vm_sdma.c | 115 WARN_ON(ib->length_dw == 0); in amdgpu_vm_sdma_commit() 121 WARN_ON(ib->length_dw > p->num_dw_left); in amdgpu_vm_sdma_commit() 245 ndw -= p->job->ibs->length_dw; in amdgpu_vm_sdma_update()
|
| A D | gfx_v8_0.c | 896 ib.length_dw = 5; in gfx_v8_0_ring_test_ib() 1525 ib.length_dw = 0; in gfx_v8_0_do_edc_gpr_workarounds() 1543 ib.ptr[ib.length_dw++] = 8; /* x */ in gfx_v8_0_do_edc_gpr_workarounds() 1544 ib.ptr[ib.length_dw++] = 1; /* y */ in gfx_v8_0_do_edc_gpr_workarounds() 1545 ib.ptr[ib.length_dw++] = 1; /* z */ in gfx_v8_0_do_edc_gpr_workarounds() 1546 ib.ptr[ib.length_dw++] = in gfx_v8_0_do_edc_gpr_workarounds() 1569 ib.ptr[ib.length_dw++] = 8; /* x */ in gfx_v8_0_do_edc_gpr_workarounds() 1570 ib.ptr[ib.length_dw++] = 1; /* y */ in gfx_v8_0_do_edc_gpr_workarounds() 1571 ib.ptr[ib.length_dw++] = 1; /* z */ in gfx_v8_0_do_edc_gpr_workarounds() 1572 ib.ptr[ib.length_dw++] = in gfx_v8_0_do_edc_gpr_workarounds() [all …]
|
| /drivers/net/ethernet/qlogic/qed/ |
| A D | qed_hw.c | 469 le16_to_cpu(p_command->length_dw), in qed_dmae_post_command() 484 le16_to_cpu(p_command->length_dw), in qed_dmae_post_command() 616 u32 length_dw) in qed_dmae_execute_sub_operation() argument 634 length_dw * sizeof(u32)); in qed_dmae_execute_sub_operation() 655 cmd->length_dw = cpu_to_le16((u16)length_dw); in qed_dmae_execute_sub_operation() 664 src_addr, dst_addr, length_dw); in qed_dmae_execute_sub_operation() 671 length_dw * sizeof(u32)); in qed_dmae_execute_sub_operation()
|