Searched refs:xgmi (Results 1 – 15 of 15) sorted by relevance
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_xgmi.c | 220 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id() 330 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive() 342 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive() 367 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive() 433 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate() 434 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate() 462 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology() 463 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology() 483 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count() 497 if (!adev->gmc.xgmi.supported) in amdgpu_xgmi_add_device() [all …]
|
D | gfxhub_v1_1.c | 55 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info() 56 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info() 59 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info() 61 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info() 63 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info()
|
D | amdgpu_xgmi.h | 73 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive() 74 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
|
D | gfxhub_v2_1.c | 518 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v2_1_get_xgmi_info() 519 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v2_1_get_xgmi_info() 522 adev->gmc.xgmi.physical_node_id = in gfxhub_v2_1_get_xgmi_info() 524 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v2_1_get_xgmi_info() 527 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v2_1_get_xgmi_info()
|
D | amdgpu_device.c | 2254 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_init() 2466 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_late_init() 2482 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_ip_late_init() 2523 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini() 4256 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset() 4258 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset() 4274 gmc.xgmi.head) { in amdgpu_do_asic_reset() 4275 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset() 4286 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset() 4295 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) { in amdgpu_do_asic_reset() [all …]
|
D | gmc_v9_0.c | 729 if (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb() 855 bool vega20_xgmi_wa = (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb_pasid() 1241 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 1250 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 1464 if (adev->gmc.xgmi.supported) { in gmc_v9_0_sw_init()
|
D | gmc_v10_0.c | 697 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location() 707 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location() 857 if (adev->gmc.xgmi.supported) { in gmc_v10_0_sw_init()
|
D | amdgpu_amdkfd.c | 544 return adev->gmc.xgmi.hive_id; in amdgpu_amdkfd_get_hive_id() 562 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count() 563 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count()
|
D | amdgpu_gmc.h | 240 struct amdgpu_xgmi xgmi; member
|
D | amdgpu_ib.c | 384 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
|
D | amdgpu_ras.c | 871 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_error_inject() 1576 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_do_recovery() 1580 list_add_tail(&adev->gmc.xgmi.head, &device_list); in amdgpu_ras_do_recovery() 1585 device_list_handle, gmc.xgmi.head) { in amdgpu_ras_do_recovery()
|
D | amdgpu_gmc.c | 157 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location()
|
D | nv.c | 476 adev->gmc.xgmi.supported = true; in nv_set_ip_blocks()
|
D | amdgpu_psp.c | 2180 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend() 2263 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_resume()
|
D | soc15.c | 703 adev->gmc.xgmi.supported = true; in soc15_set_ip_blocks()
|