2 * Copyright 2018 Advanced Micro Devices, Inc.
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
11 * The above copyright notice and this permission notice shall be included in
12 * all copies or substantial portions of the Software.
14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 * OTHER DEALINGS IN THE SOFTWARE.
25 #include "amdgpu_discovery.h"
26 #include "soc15_hw_ip.h"
27 #include "discovery.h"
29 #define mmRCC_CONFIG_MEMSIZE 0xde3
30 #define mmMM_INDEX 0x0
31 #define mmMM_INDEX_HI 0x6
35 const char *hw_id_names[HW_ID_MAX] = {
39 [SMUIO_HWID] = "SMUIO",
45 [AUDIO_AZ_HWID] = "AUDIO_AZ",
52 [DCEAZ_HWID] = "DCEAZ",
54 [SDPMUX_HWID] = "SDPMUX",
57 [L2IMU_HWID] = "L2IMU",
59 [MMHUB_HWID] = "MMHUB",
60 [ATHUB_HWID] = "ATHUB",
61 [DBGU_NBIO_HWID] = "DBGU_NBIO",
63 [DBGU0_HWID] = "DBGU0",
64 [DBGU1_HWID] = "DBGU1",
65 [OSSSYS_HWID] = "OSSSYS",
67 [SDMA0_HWID] = "SDMA0",
68 [SDMA1_HWID] = "SDMA1",
70 [DBGU_IO_HWID] = "DBGU_IO",
74 [DFX_DAP_HWID] = "DFX_DAP",
75 [L1IMU_PCIE_HWID] = "L1IMU_PCIE",
76 [L1IMU_NBIF_HWID] = "L1IMU_NBIF",
77 [L1IMU_IOAGR_HWID] = "L1IMU_IOAGR",
78 [L1IMU3_HWID] = "L1IMU3",
79 [L1IMU4_HWID] = "L1IMU4",
80 [L1IMU5_HWID] = "L1IMU5",
81 [L1IMU6_HWID] = "L1IMU6",
82 [L1IMU7_HWID] = "L1IMU7",
83 [L1IMU8_HWID] = "L1IMU8",
84 [L1IMU9_HWID] = "L1IMU9",
85 [L1IMU10_HWID] = "L1IMU10",
86 [L1IMU11_HWID] = "L1IMU11",
87 [L1IMU12_HWID] = "L1IMU12",
88 [L1IMU13_HWID] = "L1IMU13",
89 [L1IMU14_HWID] = "L1IMU14",
90 [L1IMU15_HWID] = "L1IMU15",
91 [WAFLC_HWID] = "WAFLC",
92 [FCH_USB_PD_HWID] = "FCH_USB_PD",
97 [IOAGR_HWID] = "IOAGR",
99 [IOAPIC_HWID] = "IOAPIC",
100 [SYSTEMHUB_HWID] = "SYSTEMHUB",
101 [NTBCCP_HWID] = "NTBCCP",
103 [SATA_HWID] = "SATA",
105 [CCXSEC_HWID] = "CCXSEC",
106 [XGMI_HWID] = "XGMI",
107 [XGBE_HWID] = "XGBE",
111 static int hw_id_map[MAX_HWIP] = {
113 [HDP_HWIP] = HDP_HWID,
114 [SDMA0_HWIP] = SDMA0_HWID,
115 [SDMA1_HWIP] = SDMA1_HWID,
116 [MMHUB_HWIP] = MMHUB_HWID,
117 [ATHUB_HWIP] = ATHUB_HWID,
118 [NBIO_HWIP] = NBIF_HWID,
119 [MP0_HWIP] = MP0_HWID,
120 [MP1_HWIP] = MP1_HWID,
121 [UVD_HWIP] = UVD_HWID,
122 [VCE_HWIP] = VCE_HWID,
124 [DCE_HWIP] = DMU_HWID,
125 [OSSSYS_HWIP] = OSSSYS_HWID,
126 [SMUIO_HWIP] = SMUIO_HWID,
127 [PWR_HWIP] = PWR_HWID,
128 [NBIF_HWIP] = NBIF_HWID,
129 [THM_HWIP] = THM_HWID,
130 [CLK_HWIP] = CLKA_HWID,
133 static int amdgpu_discovery_read_binary(struct amdgpu_device *adev, uint8_t *binary)
135 uint64_t vram_size = (uint64_t)RREG32(mmRCC_CONFIG_MEMSIZE) << 20;
136 uint64_t pos = vram_size - adev->discovery_tmr_size;
138 amdgpu_device_vram_access(adev, pos, (uint32_t *)binary,
139 adev->discovery_tmr_size, false);
143 static uint16_t amdgpu_discovery_calculate_checksum(uint8_t *data, uint32_t size)
145 uint16_t checksum = 0;
148 for (i = 0; i < size; i++)
154 static inline bool amdgpu_discovery_verify_checksum(uint8_t *data, uint32_t size,
157 return !!(amdgpu_discovery_calculate_checksum(data, size) == expected);
160 static int amdgpu_discovery_init(struct amdgpu_device *adev)
162 struct table_info *info;
163 struct binary_header *bhdr;
164 struct ip_discovery_header *ihdr;
165 struct gpu_info_header *ghdr;
171 adev->discovery_tmr_size = DISCOVERY_TMR_SIZE;
172 adev->discovery_bin = kzalloc(adev->discovery_tmr_size, GFP_KERNEL);
173 if (!adev->discovery_bin)
176 r = amdgpu_discovery_read_binary(adev, adev->discovery_bin);
178 DRM_ERROR("failed to read ip discovery binary\n");
182 bhdr = (struct binary_header *)adev->discovery_bin;
184 if (le32_to_cpu(bhdr->binary_signature) != BINARY_SIGNATURE) {
185 DRM_ERROR("invalid ip discovery binary signature\n");
190 offset = offsetof(struct binary_header, binary_checksum) +
191 sizeof(bhdr->binary_checksum);
192 size = bhdr->binary_size - offset;
193 checksum = bhdr->binary_checksum;
195 if (!amdgpu_discovery_verify_checksum(adev->discovery_bin + offset,
197 DRM_ERROR("invalid ip discovery binary checksum\n");
202 info = &bhdr->table_list[IP_DISCOVERY];
203 offset = le16_to_cpu(info->offset);
204 checksum = le16_to_cpu(info->checksum);
205 ihdr = (struct ip_discovery_header *)(adev->discovery_bin + offset);
207 if (le32_to_cpu(ihdr->signature) != DISCOVERY_TABLE_SIGNATURE) {
208 DRM_ERROR("invalid ip discovery data table signature\n");
213 if (!amdgpu_discovery_verify_checksum(adev->discovery_bin + offset,
214 ihdr->size, checksum)) {
215 DRM_ERROR("invalid ip discovery data table checksum\n");
220 info = &bhdr->table_list[GC];
221 offset = le16_to_cpu(info->offset);
222 checksum = le16_to_cpu(info->checksum);
223 ghdr = (struct gpu_info_header *)(adev->discovery_bin + offset);
225 if (!amdgpu_discovery_verify_checksum(adev->discovery_bin + offset,
226 ghdr->size, checksum)) {
227 DRM_ERROR("invalid gc data table checksum\n");
235 kfree(adev->discovery_bin);
236 adev->discovery_bin = NULL;
241 void amdgpu_discovery_fini(struct amdgpu_device *adev)
243 kfree(adev->discovery_bin);
244 adev->discovery_bin = NULL;
247 int amdgpu_discovery_reg_base_init(struct amdgpu_device *adev)
249 struct binary_header *bhdr;
250 struct ip_discovery_header *ihdr;
251 struct die_header *dhdr;
257 uint8_t num_base_address;
262 r = amdgpu_discovery_init(adev);
264 DRM_ERROR("amdgpu_discovery_init failed\n");
268 bhdr = (struct binary_header *)adev->discovery_bin;
269 ihdr = (struct ip_discovery_header *)(adev->discovery_bin +
270 le16_to_cpu(bhdr->table_list[IP_DISCOVERY].offset));
271 num_dies = le16_to_cpu(ihdr->num_dies);
273 DRM_DEBUG("number of dies: %d\n", num_dies);
275 for (i = 0; i < num_dies; i++) {
276 die_offset = le16_to_cpu(ihdr->die_info[i].die_offset);
277 dhdr = (struct die_header *)(adev->discovery_bin + die_offset);
278 num_ips = le16_to_cpu(dhdr->num_ips);
279 ip_offset = die_offset + sizeof(*dhdr);
281 if (le16_to_cpu(dhdr->die_id) != i) {
282 DRM_ERROR("invalid die id %d, expected %d\n",
283 le16_to_cpu(dhdr->die_id), i);
287 DRM_DEBUG("number of hardware IPs on die%d: %d\n",
288 le16_to_cpu(dhdr->die_id), num_ips);
290 for (j = 0; j < num_ips; j++) {
291 ip = (struct ip *)(adev->discovery_bin + ip_offset);
292 num_base_address = ip->num_base_address;
294 DRM_DEBUG("%s(%d) #%d v%d.%d.%d:\n",
295 hw_id_names[le16_to_cpu(ip->hw_id)],
296 le16_to_cpu(ip->hw_id),
298 ip->major, ip->minor,
301 for (k = 0; k < num_base_address; k++) {
303 * convert the endianness of base addresses in place,
304 * so that we don't need to convert them when accessing adev->reg_offset.
306 ip->base_address[k] = le32_to_cpu(ip->base_address[k]);
307 DRM_DEBUG("\t0x%08x\n", ip->base_address[k]);
310 for (hw_ip = 0; hw_ip < MAX_HWIP; hw_ip++) {
311 if (hw_id_map[hw_ip] == le16_to_cpu(ip->hw_id)) {
312 DRM_DEBUG("set register base offset for %s\n",
313 hw_id_names[le16_to_cpu(ip->hw_id)]);
314 adev->reg_offset[hw_ip][ip->number_instance] =
320 ip_offset += sizeof(*ip) + 4 * (ip->num_base_address - 1);
327 int amdgpu_discovery_get_ip_version(struct amdgpu_device *adev, int hw_id,
328 int *major, int *minor, int *revision)
330 struct binary_header *bhdr;
331 struct ip_discovery_header *ihdr;
332 struct die_header *dhdr;
340 if (!adev->discovery_bin) {
341 DRM_ERROR("ip discovery uninitialized\n");
345 bhdr = (struct binary_header *)adev->discovery_bin;
346 ihdr = (struct ip_discovery_header *)(adev->discovery_bin +
347 le16_to_cpu(bhdr->table_list[IP_DISCOVERY].offset));
348 num_dies = le16_to_cpu(ihdr->num_dies);
350 for (i = 0; i < num_dies; i++) {
351 die_offset = le16_to_cpu(ihdr->die_info[i].die_offset);
352 dhdr = (struct die_header *)(adev->discovery_bin + die_offset);
353 num_ips = le16_to_cpu(dhdr->num_ips);
354 ip_offset = die_offset + sizeof(*dhdr);
356 for (j = 0; j < num_ips; j++) {
357 ip = (struct ip *)(adev->discovery_bin + ip_offset);
359 if (le16_to_cpu(ip->hw_id) == hw_id) {
365 *revision = ip->revision;
368 ip_offset += sizeof(*ip) + 4 * (ip->num_base_address - 1);
375 int amdgpu_discovery_get_gfx_info(struct amdgpu_device *adev)
377 struct binary_header *bhdr;
378 struct gc_info_v1_0 *gc_info;
380 if (!adev->discovery_bin) {
381 DRM_ERROR("ip discovery uninitialized\n");
385 bhdr = (struct binary_header *)adev->discovery_bin;
386 gc_info = (struct gc_info_v1_0 *)(adev->discovery_bin +
387 le16_to_cpu(bhdr->table_list[GC].offset));
389 adev->gfx.config.max_shader_engines = le32_to_cpu(gc_info->gc_num_se);
390 adev->gfx.config.max_cu_per_sh = 2 * (le32_to_cpu(gc_info->gc_num_wgp0_per_sa) +
391 le32_to_cpu(gc_info->gc_num_wgp1_per_sa));
392 adev->gfx.config.max_sh_per_se = le32_to_cpu(gc_info->gc_num_sa_per_se);
393 adev->gfx.config.max_backends_per_se = le32_to_cpu(gc_info->gc_num_rb_per_se);
394 adev->gfx.config.max_texture_channel_caches = le32_to_cpu(gc_info->gc_num_gl2c);
395 adev->gfx.config.max_gprs = le32_to_cpu(gc_info->gc_num_gprs);
396 adev->gfx.config.max_gs_threads = le32_to_cpu(gc_info->gc_num_max_gs_thds);
397 adev->gfx.config.gs_vgt_table_depth = le32_to_cpu(gc_info->gc_gs_table_depth);
398 adev->gfx.config.gs_prim_buffer_depth = le32_to_cpu(gc_info->gc_gsprim_buff_depth);
399 adev->gfx.config.double_offchip_lds_buf = le32_to_cpu(gc_info->gc_double_offchip_lds_buffer);
400 adev->gfx.cu_info.wave_front_size = le32_to_cpu(gc_info->gc_wave_size);
401 adev->gfx.cu_info.max_waves_per_simd = le32_to_cpu(gc_info->gc_max_waves_per_simd);
402 adev->gfx.cu_info.max_scratch_slots_per_cu = le32_to_cpu(gc_info->gc_max_scratch_slots_per_cu);
403 adev->gfx.cu_info.lds_size = le32_to_cpu(gc_info->gc_lds_size);
404 adev->gfx.config.num_sc_per_sh = le32_to_cpu(gc_info->gc_num_sc_per_se) /
405 le32_to_cpu(gc_info->gc_num_sa_per_se);
406 adev->gfx.config.num_packer_per_sc = le32_to_cpu(gc_info->gc_num_packer_per_sc);