| /linux/drivers/accel/habanalabs/common/ |
| A D | security.c | 327 for (j = 0 ; j < num_instances ; j++) { in hl_init_pb_with_mask() 328 int seq = i * num_instances + j; in hl_init_pb_with_mask() 360 u32 num_instances, u32 instance_offset, in hl_init_pb() argument 411 for (j = 0 ; j < num_instances ; j++) { in hl_init_pb_ranges_with_mask() 412 int seq = i * num_instances + j; in hl_init_pb_ranges_with_mask() 472 u32 num_instances, u32 instance_offset, in hl_init_pb_single_dcore() argument 492 for (i = 0 ; i < num_instances ; i++) in hl_init_pb_single_dcore() 520 u32 num_instances, u32 instance_offset, in hl_init_pb_ranges_single_dcore() argument 539 for (i = 0 ; i < num_instances ; i++) in hl_init_pb_ranges_single_dcore() 572 int seq = i * num_instances + j; in hl_ack_pb_with_mask() [all …]
|
| A D | habanalabs.h | 4203 u32 dcore_offset, u32 num_instances, u32 instance_offset, 4207 u32 num_instances, u32 instance_offset, 4211 u32 dcore_offset, u32 num_instances, u32 instance_offset, 4216 u32 dcore_offset, u32 num_instances, u32 instance_offset, 4221 u32 num_instances, u32 instance_offset, 4225 u32 num_instances, u32 instance_offset, 4230 u32 num_instances, u32 instance_offset, 4233 u32 dcore_offset, u32 num_instances, u32 instance_offset, 4236 u32 num_instances, u32 instance_offset,
|
| /linux/drivers/iommu/arm/arm-smmu/ |
| A D | arm-smmu-nvidia.c | 37 unsigned int num_instances; member 69 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_write_reg() 90 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_write_reg64() 112 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_tlb_sync() 137 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_reset() 182 for (inst = 0; inst < nvidia->num_instances; inst++) { in nvidia_smmu_global_fault() 230 for (inst = 0; inst < nvidia->num_instances; inst++) { in nvidia_smmu_context_fault() 323 nvidia_smmu->num_instances++; in nvidia_smmu_impl_init() 334 nvidia_smmu->num_instances++; in nvidia_smmu_impl_init() 337 if (nvidia_smmu->num_instances == 1) in nvidia_smmu_impl_init()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| A D | sdma_v4_0.c | 602 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_setup_ulv() 627 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_init_microcode() 927 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_gfx_enable() 961 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_page_stop() 1010 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_ctx_switch_enable() 1056 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_enable() 1349 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_load_microcode() 1405 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_0_start() 2045 if (j == adev->sdma.num_instances) in sdma_v4_0_wait_for_idle() 2529 switch (adev->sdma.num_instances) { in sdma_v4_0_set_irq_funcs() [all …]
|
| A D | amdgpu_sdma.c | 42 for (i = 0; i < adev->sdma.num_instances; i++) in amdgpu_sdma_get_instance_from_ring() 55 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_get_index_from_ring() 108 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_ras_late_init() 198 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_destroy_inst_ctx() 243 for (i = 1; i < adev->sdma.num_instances; i++) in amdgpu_sdma_init_microcode() 255 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_init_microcode()
|
| A D | sdma_v4_4_2.c | 161 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_inst_init_golden_registers() 190 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_init_microcode() 1389 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_sw_init() 1453 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_sw_fini() 1533 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_is_idle() 1555 if (j == adev->sdma.num_instances) in sdma_v4_4_2_wait_for_idle() 1596 for (i = instance; i < adev->sdma.num_instances; in sdma_v4_4_2_process_trap_irq() 1603 if (i >= adev->sdma.num_instances) { in sdma_v4_4_2_process_trap_irq() 1871 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_print_ip_state() 1892 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_2_dump_ip_state() [all …]
|
| A D | sdma_v3_0.c | 254 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v3_0_free_microcode() 305 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v3_0_init_microcode() 332 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v3_0_init_microcode() 517 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v3_0_gfx_stop() 576 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v3_0_ctx_switch_enable() 618 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v3_0_enable() 645 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v3_0_gfx_resume() 743 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v3_0_gfx_resume() 1090 adev->sdma.num_instances = 1; in sdma_v3_0_early_init() 1093 adev->sdma.num_instances = SDMA_MAX_INSTANCE; in sdma_v3_0_early_init() [all …]
|
| A D | cik_sdma.c | 77 for (i = 0; i < adev->sdma.num_instances; i++) in cik_sdma_free_microcode() 133 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_sdma_init_microcode() 147 for (i = 0; i < adev->sdma.num_instances; i++) in cik_sdma_init_microcode() 312 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_sdma_gfx_stop() 369 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_ctx_switch_enable() 407 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_sdma_enable() 432 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_sdma_gfx_resume() 495 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_sdma_gfx_resume() 537 for (i = 0; i < adev->sdma.num_instances; i++) { in cik_sdma_load_microcode() 926 adev->sdma.num_instances = SDMA_MAX_INSTANCE; in cik_sdma_early_init() [all …]
|
| A D | si_dma.c | 118 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_stop() 133 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_start() 464 adev->sdma.num_instances = 2; in si_dma_early_init() 492 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_sw_init() 514 for (i = 0; i < adev->sdma.num_instances; i++) in si_dma_sw_fini() 648 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_clockgating_state() 660 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_clockgating_state() 744 for (i = 0; i < adev->sdma.num_instances; i++) in si_dma_set_ring_funcs() 838 for (i = 0; i < adev->sdma.num_instances; i++) { in si_dma_set_vm_pte_funcs() 842 adev->vm_manager.vm_pte_num_scheds = adev->sdma.num_instances; in si_dma_set_vm_pte_funcs()
|
| A D | sdma_v2_4.c | 114 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v2_4_free_microcode() 145 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v2_4_init_microcode() 174 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v2_4_init_microcode() 341 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v2_4_gfx_stop() 381 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v2_4_enable() 406 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v2_4_gfx_resume() 469 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v2_4_gfx_resume() 815 adev->sdma.num_instances = SDMA_MAX_INSTANCE; in sdma_v2_4_early_init() 853 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v2_4_sw_init() 874 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v2_4_sw_fini() [all …]
|
| A D | sdma_v7_0.c | 430 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_gfx_stop() 485 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_enable() 511 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_gfx_resume() 668 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v12_0_free_ucode_buffer() 705 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_load_microcode() 764 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_soft_reset() 799 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_check_soft_reset() 1297 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_sw_init() 1334 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v7_0_sw_fini() 1386 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v7_0_is_idle() [all …]
|
| A D | sdma_v6_0.c | 398 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_gfx_stop() 434 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_ctxempty_int_enable() 464 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_enable() 490 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_gfx_resume() 744 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_soft_reset() 779 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_check_soft_reset() 1309 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_sw_init() 1351 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v6_0_sw_fini() 1400 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_is_idle() 1570 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_print_ip_state() [all …]
|
| A D | sdma_v5_2.c | 418 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_gfx_stop() 477 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_ctx_switch_enable() 516 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_enable() 544 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_gfx_resume() 713 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_load_microcode() 746 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_soft_reset() 1280 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_sw_init() 1288 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_sw_init() 1324 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v5_2_sw_fini() 1373 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_is_idle() [all …]
|
| A D | vpe_v6_1.c | 78 for (i = 0; i < vpe->num_instances; i++) { in vpe_v6_1_halt() 108 for (i = 0; i < vpe->num_instances; i++) { in vpe_v6_1_set_collaborate_mode() 133 for (j = 0; j < vpe->num_instances; j++) { in vpe_v6_1_load_microcode() 183 for (j = 0; j < vpe->num_instances; j++) { in vpe_v6_1_load_microcode() 215 for (i = 0; i < vpe->num_instances; i++) { in vpe_v6_1_ring_start() 282 for (i = 0; i < vpe->num_instances; i++) { in vpe_v_6_1_ring_stop()
|
| A D | sdma_v5_0.c | 292 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_init_microcode() 599 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_gfx_stop() 658 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_ctx_switch_enable() 700 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_enable() 727 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_gfx_resume() 899 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_load_microcode() 1410 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_sw_init() 1447 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v5_0_sw_fini() 1501 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_is_idle() 1676 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_update_medium_grain_clock_gating() [all …]
|
| A D | sdma_v4_4.c | 243 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_reset_ras_error_count() 256 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_query_ras_error_count()
|
| A D | aqua_vanjaram.c | 51 for (i = 0; i < adev->sdma.num_instances; i++) in aqua_vanjaram_doorbell_index_init() 400 num_sdma = adev->sdma.num_instances; in __aqua_vanjaram_get_xcp_ip_info() 688 adev->sdma.num_instances = NUM_SDMA(adev->sdma.sdma_mask); in aqua_vanjaram_init_soc_config() 821 pcie_reg_state->common_header.num_instances = 1; in aqua_vanjaram_read_pcie_state() 905 xgmi_reg_state->common_header.num_instances = max_xgmi_instances; in aqua_vanjaram_read_xgmi_state() 978 wafl_reg_state->common_header.num_instances = max_wafl_instances; in aqua_vanjaram_read_wafl_state() 1097 usr_reg_state->common_header.num_instances = max_usr_instances; in aqua_vanjaram_read_usr_state()
|
| A D | amdgpu_discovery.c | 1361 if (adev->sdma.num_instances < in amdgpu_discovery_reg_base_init() 1363 adev->sdma.num_instances++; in amdgpu_discovery_reg_base_init() 1368 adev->sdma.num_instances + 1, in amdgpu_discovery_reg_base_init() 1375 adev->vpe.num_instances++; in amdgpu_discovery_reg_base_init() 1378 adev->vpe.num_instances + 1, in amdgpu_discovery_reg_base_init() 2432 adev->sdma.num_instances = 2; in amdgpu_discovery_set_ip_blocks() 2454 adev->sdma.num_instances = 2; in amdgpu_discovery_set_ip_blocks() 2476 adev->sdma.num_instances = 1; in amdgpu_discovery_set_ip_blocks() 2515 adev->sdma.num_instances = 2; in amdgpu_discovery_set_ip_blocks() 2538 adev->sdma.num_instances = 8; in amdgpu_discovery_set_ip_blocks() [all …]
|
| A D | amdgpu_vpe.h | 80 uint32_t num_instances; member
|
| A D | amdgpu_sdma.h | 111 int num_instances; member
|
| /linux/drivers/gpu/drm/amd/include/ |
| A D | amdgpu_reg_state.h | 51 uint8_t num_instances; member
|
| /linux/drivers/media/platform/samsung/exynos4-is/ |
| A D | fimc-lite.h | 69 unsigned short num_instances; member
|
| /linux/include/sound/ |
| A D | timer.h | 79 int num_instances; /* current number of timer instances */ member
|
| /linux/drivers/hwmon/ |
| A D | ibmaem.c | 191 u8 num_instances; member 203 u8 num_instances; member 519 return ff_resp.num_instances; in aem_find_aem1_count() 657 fi_resp->num_instances <= instance_num) in aem_find_aem2()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_device.c | 130 kfd->adev->sdma.num_instances * in kfd_device_info_set_sdma_info() 1416 return node->adev->sdma.num_instances/(int)node->kfd->num_nodes; in kfd_get_num_sdma_engines() 1418 return min(node->adev->sdma.num_instances/(int)node->kfd->num_nodes, 2); in kfd_get_num_sdma_engines() 1424 return node->adev->sdma.num_instances/(int)node->kfd->num_nodes - in kfd_get_num_xgmi_sdma_engines()
|