Lines Matching refs:gpu
108 return top_dev->gpu; in kfd_device_by_id()
261 if (iolink->gpu && kfd_devcgroup_check_permission(iolink->gpu)) in iolink_show()
303 if (mem->gpu && kfd_devcgroup_check_permission(mem->gpu)) in mem_show()
335 if (cache->gpu && kfd_devcgroup_check_permission(cache->gpu)) in kfd_cache_show()
417 if (dev->gpu && kfd_devcgroup_check_permission(dev->gpu)) in node_show()
426 if (dev->gpu && kfd_devcgroup_check_permission(dev->gpu)) in node_show()
433 if (dev->gpu && kfd_devcgroup_check_permission(dev->gpu)) in node_show()
438 dev->gpu ? dev->node_props.simd_count : 0); in node_show()
462 dev->gpu ? (dev->node_props.array_count * in node_show()
463 NUM_XCC(dev->gpu->xcc_mask)) : 0); in node_show()
495 if (dev->gpu) { in node_show()
497 __ilog2_u32(dev->gpu->kfd->device_info.num_of_watch_points); in node_show()
509 if (dev->gpu->adev->asic_type == CHIP_TONGA) in node_show()
513 if (KFD_GC_VERSION(dev->gpu) < IP_VERSION(10, 0, 0) && in node_show()
514 (dev->gpu->adev->sdma.supported_reset & AMDGPU_RESET_TYPE_PER_QUEUE)) in node_show()
523 dev->gpu->kfd->mec_fw_version); in node_show()
531 dev->gpu->kfd->sdma_fw_version); in node_show()
533 dev->gpu->adev->unique_id); in node_show()
535 NUM_XCC(dev->gpu->xcc_mask)); in node_show()
986 if (!kdev->gpu) { in kfd_add_non_crat_information()
1082 static uint32_t kfd_generate_gpu_id(struct kfd_node *gpu) in kfd_generate_gpu_id() argument
1091 if (!gpu) in kfd_generate_gpu_id()
1095 local_mem_size = gpu->local_mem_info.local_mem_size_private + in kfd_generate_gpu_id()
1096 gpu->local_mem_info.local_mem_size_public; in kfd_generate_gpu_id()
1097 buf[0] = gpu->adev->pdev->devfn; in kfd_generate_gpu_id()
1098 buf[1] = gpu->adev->pdev->subsystem_vendor | in kfd_generate_gpu_id()
1099 (gpu->adev->pdev->subsystem_device << 16); in kfd_generate_gpu_id()
1100 buf[2] = pci_domain_nr(gpu->adev->pdev->bus); in kfd_generate_gpu_id()
1101 buf[3] = gpu->adev->pdev->device; in kfd_generate_gpu_id()
1102 buf[4] = gpu->adev->pdev->bus->number; in kfd_generate_gpu_id()
1105 buf[7] = (ffs(gpu->xcc_mask) - 1) | (NUM_XCC(gpu->xcc_mask) << 16); in kfd_generate_gpu_id()
1123 if (dev->gpu && dev->gpu_id == gpu_id) { in kfd_generate_gpu_id()
1141 static struct kfd_topology_device *kfd_assign_gpu(struct kfd_node *gpu) in kfd_assign_gpu() argument
1157 if (!dev->gpu && (dev->node_props.simd_count > 0)) { in kfd_assign_gpu()
1158 dev->gpu = gpu; in kfd_assign_gpu()
1162 mem->gpu = dev->gpu; in kfd_assign_gpu()
1164 cache->gpu = dev->gpu; in kfd_assign_gpu()
1166 iolink->gpu = dev->gpu; in kfd_assign_gpu()
1168 p2plink->gpu = dev->gpu; in kfd_assign_gpu()
1200 amdgpu_amdkfd_get_local_mem_info(dev->gpu->adev, &local_mem_info, in kfd_fill_mem_clk_max_info()
1201 dev->gpu->xcp); in kfd_fill_mem_clk_max_info()
1219 pcie_capability_read_dword(target_gpu_dev->gpu->adev->pdev, in kfd_set_iolink_no_atomics()
1228 if (!dev->gpu->kfd->pci_atomic_requested || in kfd_set_iolink_no_atomics()
1229 dev->gpu->adev->asic_type == CHIP_HAWAII) in kfd_set_iolink_no_atomics()
1240 if (!to_dev->gpu && in kfd_set_iolink_non_coherent()
1244 if (to_dev->gpu) { in kfd_set_iolink_non_coherent()
1250 KFD_GC_VERSION(to_dev->gpu) == IP_VERSION(9, 4, 0))) { in kfd_set_iolink_non_coherent()
1272 struct kfd_node *gpu = outbound_link->gpu; in kfd_set_recommended_sdma_engines() local
1273 struct amdgpu_device *adev = gpu->adev; in kfd_set_recommended_sdma_engines()
1275 unsigned int num_xgmi_sdma_engines = kfd_get_num_xgmi_sdma_engines(gpu); in kfd_set_recommended_sdma_engines()
1276 unsigned int num_sdma_engines = kfd_get_num_sdma_engines(gpu); in kfd_set_recommended_sdma_engines()
1281 bool support_rec_eng = !amdgpu_sriov_vf(adev) && to_dev->gpu && in kfd_set_recommended_sdma_engines()
1282 adev->aid_mask && num_xgmi_nodes && gpu->kfd->num_nodes == 1 && in kfd_set_recommended_sdma_engines()
1288 int dst_socket_id = to_dev->gpu->adev->gmc.xgmi.physical_node_id; in kfd_set_recommended_sdma_engines()
1304 num_xgmi_sdma_engines && to_dev->gpu) ? xgmi_sdma_eng_id_mask : in kfd_set_recommended_sdma_engines()
1317 if (!dev || !dev->gpu) in kfd_fill_iolink_non_crat_info()
1331 if (!peer_dev->gpu && in kfd_fill_iolink_non_crat_info()
1338 dev->node_props.hive_id = pci_dev_id(dev->gpu->adev->pdev); in kfd_fill_iolink_non_crat_info()
1412 if (cpu_dev->gpu) in kfd_create_indirect_link_prop()
1465 if (kfd_dev_is_large_bar(kdev->gpu)) { in kfd_create_indirect_link_prop()
1495 kdev->gpu->adev, in kfd_add_peer_prop()
1496 peer->gpu->adev)) in kfd_add_peer_prop()
1572 if (WARN_ON(!new_dev->gpu)) in kfd_dev_create_p2p_links()
1588 if (!dev->gpu || !dev->gpu->adev || in kfd_dev_create_p2p_links()
1589 (dev->gpu->kfd->hive_id && in kfd_dev_create_p2p_links()
1590 dev->gpu->kfd->hive_id == new_dev->gpu->kfd->hive_id)) in kfd_dev_create_p2p_links()
1864 static int kfd_topology_add_device_locked(struct kfd_node *gpu, in kfd_topology_add_device_locked() argument
1874 COMPUTE_UNIT_GPU, gpu, in kfd_topology_add_device_locked()
1877 dev_err(gpu->adev->dev, "Error creating VCRAT\n"); in kfd_topology_add_device_locked()
1888 dev_err(gpu->adev->dev, "Error parsing VCRAT\n"); in kfd_topology_add_device_locked()
1896 *dev = kfd_assign_gpu(gpu); in kfd_topology_add_device_locked()
1905 kfd_fill_cache_non_crat_info(*dev, gpu); in kfd_topology_add_device_locked()
1914 dev_err(gpu->adev->dev, "Failed to update GPU to sysfs topology. res=%d\n", in kfd_topology_add_device_locked()
1926 if (KFD_GC_VERSION(dev->gpu) >= IP_VERSION(11, 0, 0) && in kfd_topology_set_dbg_firmware_support()
1927 KFD_GC_VERSION(dev->gpu) < IP_VERSION(12, 0, 0)) { in kfd_topology_set_dbg_firmware_support()
1928 uint32_t mes_api_rev = (dev->gpu->adev->mes.sched_version & in kfd_topology_set_dbg_firmware_support()
1931 uint32_t mes_rev = dev->gpu->adev->mes.sched_version & in kfd_topology_set_dbg_firmware_support()
1942 switch (KFD_GC_VERSION(dev->gpu)) { in kfd_topology_set_dbg_firmware_support()
1944 firmware_supported = dev->gpu->kfd->mec_fw_version >= 459 + 32768; in kfd_topology_set_dbg_firmware_support()
1951 firmware_supported = dev->gpu->kfd->mec_fw_version >= 459; in kfd_topology_set_dbg_firmware_support()
1954 firmware_supported = dev->gpu->kfd->mec_fw_version >= 60; in kfd_topology_set_dbg_firmware_support()
1957 firmware_supported = dev->gpu->kfd->mec_fw_version >= 51; in kfd_topology_set_dbg_firmware_support()
1962 firmware_supported = dev->gpu->kfd->mec_fw_version >= 144; in kfd_topology_set_dbg_firmware_support()
1969 firmware_supported = dev->gpu->kfd->mec_fw_version >= 89; in kfd_topology_set_dbg_firmware_support()
1994 if (kfd_dbg_has_ttmps_always_setup(dev->gpu)) in kfd_topology_set_capabilities()
1997 if (KFD_GC_VERSION(dev->gpu) < IP_VERSION(10, 0, 0)) { in kfd_topology_set_capabilities()
1998 if (KFD_GC_VERSION(dev->gpu) == IP_VERSION(9, 4, 3) || in kfd_topology_set_capabilities()
1999 KFD_GC_VERSION(dev->gpu) == IP_VERSION(9, 4, 4)) in kfd_topology_set_capabilities()
2008 if (KFD_GC_VERSION(dev->gpu) >= IP_VERSION(9, 4, 2)) in kfd_topology_set_capabilities()
2012 if (!amdgpu_sriov_vf(dev->gpu->adev)) in kfd_topology_set_capabilities()
2019 if (KFD_GC_VERSION(dev->gpu) >= IP_VERSION(12, 0, 0)) in kfd_topology_set_capabilities()
2027 int kfd_topology_add_device(struct kfd_node *gpu) in kfd_topology_add_device() argument
2033 const char *asic_name = amdgpu_asic_name[gpu->adev->asic_type]; in kfd_topology_add_device()
2034 struct amdgpu_gfx_config *gfx_info = &gpu->adev->gfx.config; in kfd_topology_add_device()
2035 struct amdgpu_cu_info *cu_info = &gpu->adev->gfx.cu_info; in kfd_topology_add_device()
2037 if (gpu->xcp && !gpu->xcp->ddev) { in kfd_topology_add_device()
2038 dev_warn(gpu->adev->dev, in kfd_topology_add_device()
2042 dev_dbg(gpu->adev->dev, "Adding new GPU to topology\n"); in kfd_topology_add_device()
2052 dev = kfd_assign_gpu(gpu); in kfd_topology_add_device()
2054 res = kfd_topology_add_device_locked(gpu, &dev); in kfd_topology_add_device()
2059 gpu_id = kfd_generate_gpu_id(gpu); in kfd_topology_add_device()
2061 gpu->id = gpu_id; in kfd_topology_add_device()
2083 gpu->kfd->device_info.gfx_target_version; in kfd_topology_add_device()
2084 dev->node_props.vendor_id = gpu->adev->pdev->vendor; in kfd_topology_add_device()
2085 dev->node_props.device_id = gpu->adev->pdev->device; in kfd_topology_add_device()
2087 ((dev->gpu->adev->rev_id << HSA_CAP_ASIC_REVISION_SHIFT) & in kfd_topology_add_device()
2090 dev->node_props.location_id = pci_dev_id(gpu->adev->pdev); in kfd_topology_add_device()
2091 if (gpu->kfd->num_nodes > 1) in kfd_topology_add_device()
2092 dev->node_props.location_id |= dev->gpu->node_id; in kfd_topology_add_device()
2094 dev->node_props.domain = pci_domain_nr(gpu->adev->pdev->bus); in kfd_topology_add_device()
2096 amdgpu_amdkfd_get_max_engine_clock_in_mhz(dev->gpu->adev); in kfd_topology_add_device()
2100 if (gpu->xcp) in kfd_topology_add_device()
2101 dev->node_props.drm_render_minor = gpu->xcp->ddev->render->index; in kfd_topology_add_device()
2104 gpu->kfd->shared_resources.drm_render_minor; in kfd_topology_add_device()
2106 dev->node_props.hive_id = gpu->kfd->hive_id; in kfd_topology_add_device()
2107 dev->node_props.num_sdma_engines = kfd_get_num_sdma_engines(gpu); in kfd_topology_add_device()
2109 kfd_get_num_xgmi_sdma_engines(gpu); in kfd_topology_add_device()
2111 gpu->kfd->device_info.num_sdma_queues_per_engine - in kfd_topology_add_device()
2112 gpu->kfd->device_info.num_reserved_sdma_queues_per_engine; in kfd_topology_add_device()
2113 dev->node_props.num_gws = (dev->gpu->gws && in kfd_topology_add_device()
2114 dev->gpu->dqm->sched_policy != KFD_SCHED_POLICY_NO_HWS) ? in kfd_topology_add_device()
2115 dev->gpu->adev->gds.gws_size : 0; in kfd_topology_add_device()
2116 dev->node_props.num_cp_queues = get_cp_queues_num(dev->gpu->dqm); in kfd_topology_add_device()
2121 switch (dev->gpu->adev->asic_type) { in kfd_topology_add_device()
2141 if (KFD_GC_VERSION(dev->gpu) < IP_VERSION(9, 0, 1)) in kfd_topology_add_device()
2143 dev->gpu->adev->asic_type); in kfd_topology_add_device()
2159 if (dev->gpu->adev->asic_type == CHIP_CARRIZO) { in kfd_topology_add_device()
2167 ((dev->gpu->adev->ras_enabled & BIT(AMDGPU_RAS_BLOCK__GFX)) != 0) ? in kfd_topology_add_device()
2170 ((dev->gpu->adev->ras_enabled & BIT(AMDGPU_RAS_BLOCK__UMC)) != 0) ? in kfd_topology_add_device()
2173 if (KFD_GC_VERSION(dev->gpu) != IP_VERSION(9, 0, 1)) in kfd_topology_add_device()
2174 dev->node_props.capability |= (dev->gpu->adev->ras_enabled != 0) ? in kfd_topology_add_device()
2177 if (KFD_IS_SVM_API_SUPPORTED(dev->gpu->adev)) in kfd_topology_add_device()
2180 if (dev->gpu->adev->gmc.is_app_apu || in kfd_topology_add_device()
2181 dev->gpu->adev->gmc.xgmi.connected_to_cpu) in kfd_topology_add_device()
2255 int kfd_topology_remove_device(struct kfd_node *gpu) in kfd_topology_remove_device() argument
2265 if (dev->gpu == gpu) { in kfd_topology_remove_device()
2306 *kdev = top_dev->gpu; in kfd_topology_enum_kfd_devices()
2360 if (!dev->gpu) { in kfd_debugfs_hqds_by_device()
2365 seq_printf(m, "Node %u, gpu_id %x:\n", i++, dev->gpu->id); in kfd_debugfs_hqds_by_device()
2366 r = dqm_debugfs_hqds(m, dev->gpu->dqm); in kfd_debugfs_hqds_by_device()
2385 if (!dev->gpu) { in kfd_debugfs_rls_by_device()
2390 seq_printf(m, "Node %u, gpu_id %x:\n", i++, dev->gpu->id); in kfd_debugfs_rls_by_device()
2391 r = pm_debugfs_runlist(m, &dev->gpu->dqm->packet_mgr); in kfd_debugfs_rls_by_device()