Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

drm/amdgpu: add gfxhub 11.5.0 support

Add initial gfxhub 11.5 support.

Signed-off-by: benl <ben.li@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>

authored by

benl and committed by
Alex Deucher
96271dd4 b90975fa

+552 -1
+2 -1
drivers/gpu/drm/amd/amdgpu/Makefile
··· 113 113 gfxhub_v1_0.o mmhub_v1_0.o gmc_v9_0.o gfxhub_v1_1.o mmhub_v9_4.o \ 114 114 gfxhub_v2_0.o mmhub_v2_0.o gmc_v10_0.o gfxhub_v2_1.o mmhub_v2_3.o \ 115 115 mmhub_v1_7.o gfxhub_v3_0.o mmhub_v3_0.o mmhub_v3_0_2.o gmc_v11_0.o \ 116 - mmhub_v3_0_1.o gfxhub_v3_0_3.o gfxhub_v1_2.o mmhub_v1_8.o mmhub_v3_3.o 116 + mmhub_v3_0_1.o gfxhub_v3_0_3.o gfxhub_v1_2.o mmhub_v1_8.o mmhub_v3_3.o \ 117 + gfxhub_v11_5_0.o 117 118 118 119 # add UMC block 119 120 amdgpu-y += \
+517
drivers/gpu/drm/amd/amdgpu/gfxhub_v11_5_0.c
··· 1 + /* 2 + * Copyright 2023 Advanced Micro Devices, Inc. 3 + * 4 + * Permission is hereby granted, free of charge, to any person obtaining a 5 + * copy of this software and associated documentation files (the "Software"), 6 + * to deal in the Software without restriction, including without limitation 7 + * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 + * and/or sell copies of the Software, and to permit persons to whom the 9 + * Software is furnished to do so, subject to the following conditions: 10 + * 11 + * The above copyright notice and this permission notice shall be included in 12 + * all copies or substantial portions of the Software. 13 + * 14 + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 + * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 + * OTHER DEALINGS IN THE SOFTWARE. 21 + * 22 + */ 23 + 24 + #include "amdgpu.h" 25 + #include "gfxhub_v11_5_0.h" 26 + 27 + #include "gc/gc_11_5_0_offset.h" 28 + #include "gc/gc_11_5_0_sh_mask.h" 29 + 30 + #include "navi10_enum.h" 31 + #include "soc15_common.h" 32 + 33 + #define regGCVM_L2_CNTL3_DEFAULT 0x80100007 34 + #define regGCVM_L2_CNTL4_DEFAULT 0x000000c1 35 + #define regGCVM_L2_CNTL5_DEFAULT 0x00003fe0 36 + 37 + 38 + static const char *gfxhub_client_ids[] = { 39 + "CB/DB", 40 + "Reserved", 41 + "GE1", 42 + "GE2", 43 + "CPF", 44 + "CPC", 45 + "CPG", 46 + "RLC", 47 + "TCP", 48 + "SQC (inst)", 49 + "SQC (data)", 50 + "SQG", 51 + "Reserved", 52 + "SDMA0", 53 + "SDMA1", 54 + "GCR", 55 + "SDMA2", 56 + "SDMA3", 57 + }; 58 + 59 + static uint32_t gfxhub_v11_5_0_get_invalidate_req(unsigned int vmid, 60 + uint32_t flush_type) 61 + { 62 + u32 req = 0; 63 + 64 + /* invalidate using legacy mode on vmid*/ 65 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, 66 + PER_VMID_INVALIDATE_REQ, 1 << vmid); 67 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, FLUSH_TYPE, flush_type); 68 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PTES, 1); 69 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PDE0, 1); 70 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PDE1, 1); 71 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PDE2, 1); 72 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, INVALIDATE_L1_PTES, 1); 73 + req = REG_SET_FIELD(req, GCVM_INVALIDATE_ENG0_REQ, 74 + CLEAR_PROTECTION_FAULT_STATUS_ADDR, 0); 75 + 76 + return req; 77 + } 78 + 79 + static void 80 + gfxhub_v11_5_0_print_l2_protection_fault_status(struct amdgpu_device *adev, 81 + uint32_t status) 82 + { 83 + u32 cid = REG_GET_FIELD(status, 84 + GCVM_L2_PROTECTION_FAULT_STATUS, CID); 85 + 86 + dev_err(adev->dev, 87 + "GCVM_L2_PROTECTION_FAULT_STATUS:0x%08X\n", 88 + status); 89 + dev_err(adev->dev, "\t Faulty UTCL2 client ID: %s (0x%x)\n", 90 + cid >= ARRAY_SIZE(gfxhub_client_ids) ? "unknown" : gfxhub_client_ids[cid], 91 + cid); 92 + dev_err(adev->dev, "\t MORE_FAULTS: 0x%lx\n", 93 + REG_GET_FIELD(status, 94 + GCVM_L2_PROTECTION_FAULT_STATUS, MORE_FAULTS)); 95 + dev_err(adev->dev, "\t WALKER_ERROR: 0x%lx\n", 96 + REG_GET_FIELD(status, 97 + GCVM_L2_PROTECTION_FAULT_STATUS, WALKER_ERROR)); 98 + dev_err(adev->dev, "\t PERMISSION_FAULTS: 0x%lx\n", 99 + REG_GET_FIELD(status, 100 + GCVM_L2_PROTECTION_FAULT_STATUS, PERMISSION_FAULTS)); 101 + dev_err(adev->dev, "\t MAPPING_ERROR: 0x%lx\n", 102 + REG_GET_FIELD(status, 103 + GCVM_L2_PROTECTION_FAULT_STATUS, MAPPING_ERROR)); 104 + dev_err(adev->dev, "\t RW: 0x%lx\n", 105 + REG_GET_FIELD(status, 106 + GCVM_L2_PROTECTION_FAULT_STATUS, RW)); 107 + } 108 + 109 + static u64 gfxhub_v11_5_0_get_fb_location(struct amdgpu_device *adev) 110 + { 111 + u64 base = RREG32_SOC15(GC, 0, regGCMC_VM_FB_LOCATION_BASE); 112 + 113 + base &= GCMC_VM_FB_LOCATION_BASE__FB_BASE_MASK; 114 + base <<= 24; 115 + 116 + return base; 117 + } 118 + 119 + static u64 gfxhub_v11_5_0_get_mc_fb_offset(struct amdgpu_device *adev) 120 + { 121 + return (u64)RREG32_SOC15(GC, 0, regGCMC_VM_FB_OFFSET) << 24; 122 + } 123 + 124 + static void gfxhub_v11_5_0_setup_vm_pt_regs(struct amdgpu_device *adev, uint32_t vmid, 125 + uint64_t page_table_base) 126 + { 127 + struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_GFXHUB(0)]; 128 + 129 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_LO32, 130 + hub->ctx_addr_distance * vmid, 131 + lower_32_bits(page_table_base)); 132 + 133 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_HI32, 134 + hub->ctx_addr_distance * vmid, 135 + upper_32_bits(page_table_base)); 136 + } 137 + 138 + static void gfxhub_v11_5_0_init_gart_aperture_regs(struct amdgpu_device *adev) 139 + { 140 + uint64_t pt_base = amdgpu_gmc_pd_addr(adev->gart.bo); 141 + 142 + gfxhub_v11_5_0_setup_vm_pt_regs(adev, 0, pt_base); 143 + 144 + WREG32_SOC15(GC, 0, regGCVM_CONTEXT0_PAGE_TABLE_START_ADDR_LO32, 145 + (u32)(adev->gmc.gart_start >> 12)); 146 + WREG32_SOC15(GC, 0, regGCVM_CONTEXT0_PAGE_TABLE_START_ADDR_HI32, 147 + (u32)(adev->gmc.gart_start >> 44)); 148 + 149 + WREG32_SOC15(GC, 0, regGCVM_CONTEXT0_PAGE_TABLE_END_ADDR_LO32, 150 + (u32)(adev->gmc.gart_end >> 12)); 151 + WREG32_SOC15(GC, 0, regGCVM_CONTEXT0_PAGE_TABLE_END_ADDR_HI32, 152 + (u32)(adev->gmc.gart_end >> 44)); 153 + } 154 + 155 + static void gfxhub_v11_5_0_init_system_aperture_regs(struct amdgpu_device *adev) 156 + { 157 + uint64_t value; 158 + 159 + WREG32_SOC15(GC, 0, regGCMC_VM_AGP_BASE, 0); 160 + WREG32_SOC15(GC, 0, regGCMC_VM_AGP_BOT, adev->gmc.agp_start >> 24); 161 + WREG32_SOC15(GC, 0, regGCMC_VM_AGP_TOP, adev->gmc.agp_end >> 24); 162 + 163 + /* Program the system aperture low logical page number. */ 164 + WREG32_SOC15(GC, 0, regGCMC_VM_SYSTEM_APERTURE_LOW_ADDR, 165 + min(adev->gmc.fb_start, adev->gmc.agp_start) >> 18); 166 + 167 + WREG32_SOC15(GC, 0, regGCMC_VM_SYSTEM_APERTURE_HIGH_ADDR, 168 + max(adev->gmc.fb_end, adev->gmc.agp_end) >> 18); 169 + 170 + /* Set default page address. */ 171 + value = adev->mem_scratch.gpu_addr - adev->gmc.vram_start 172 + + adev->vm_manager.vram_base_offset; 173 + WREG32_SOC15(GC, 0, regGCMC_VM_SYSTEM_APERTURE_DEFAULT_ADDR_LSB, 174 + (u32)(value >> 12)); 175 + WREG32_SOC15(GC, 0, regGCMC_VM_SYSTEM_APERTURE_DEFAULT_ADDR_MSB, 176 + (u32)(value >> 44)); 177 + 178 + /* Program "protection fault". */ 179 + WREG32_SOC15(GC, 0, regGCVM_L2_PROTECTION_FAULT_DEFAULT_ADDR_LO32, 180 + (u32)(adev->dummy_page_addr >> 12)); 181 + WREG32_SOC15(GC, 0, regGCVM_L2_PROTECTION_FAULT_DEFAULT_ADDR_HI32, 182 + (u32)((u64)adev->dummy_page_addr >> 44)); 183 + 184 + WREG32_FIELD15_PREREG(GC, 0, GCVM_L2_PROTECTION_FAULT_CNTL2, 185 + ACTIVE_PAGE_MIGRATION_PTE_READ_RETRY, 1); 186 + } 187 + 188 + static void gfxhub_v11_5_0_init_tlb_regs(struct amdgpu_device *adev) 189 + { 190 + uint32_t tmp; 191 + 192 + /* Setup TLB control */ 193 + tmp = RREG32_SOC15(GC, 0, regGCMC_VM_MX_L1_TLB_CNTL); 194 + 195 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, ENABLE_L1_TLB, 1); 196 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, SYSTEM_ACCESS_MODE, 3); 197 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, 198 + ENABLE_ADVANCED_DRIVER_MODEL, 1); 199 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, 200 + SYSTEM_APERTURE_UNMAPPED_ACCESS, 0); 201 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, ECO_BITS, 0); 202 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, 203 + MTYPE, MTYPE_UC); /* UC, uncached */ 204 + 205 + WREG32_SOC15(GC, 0, regGCMC_VM_MX_L1_TLB_CNTL, tmp); 206 + } 207 + 208 + static void gfxhub_v11_5_0_init_cache_regs(struct amdgpu_device *adev) 209 + { 210 + uint32_t tmp; 211 + 212 + /* These registers are not accessible to VF-SRIOV. 213 + * The PF will program them instead. 214 + */ 215 + if (amdgpu_sriov_vf(adev)) 216 + return; 217 + 218 + /* Setup L2 cache */ 219 + tmp = RREG32_SOC15(GC, 0, regGCVM_L2_CNTL); 220 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, ENABLE_L2_CACHE, 1); 221 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, ENABLE_L2_FRAGMENT_PROCESSING, 0); 222 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, 223 + ENABLE_DEFAULT_PAGE_OUT_TO_SYSTEM_MEMORY, 1); 224 + /* XXX for emulation, Refer to closed source code.*/ 225 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, 226 + L2_PDE0_CACHE_TAG_GENERATION_MODE, 0); 227 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, PDE_FAULT_CLASSIFICATION, 0); 228 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, CONTEXT1_IDENTITY_ACCESS_MODE, 1); 229 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL, IDENTITY_MODE_FRAGMENT_SIZE, 0); 230 + WREG32_SOC15(GC, 0, regGCVM_L2_CNTL, tmp); 231 + 232 + tmp = RREG32_SOC15(GC, 0, regGCVM_L2_CNTL2); 233 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL2, INVALIDATE_ALL_L1_TLBS, 1); 234 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL2, INVALIDATE_L2_CACHE, 1); 235 + WREG32_SOC15(GC, 0, regGCVM_L2_CNTL2, tmp); 236 + 237 + tmp = regGCVM_L2_CNTL3_DEFAULT; 238 + if (adev->gmc.translate_further) { 239 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL3, BANK_SELECT, 12); 240 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL3, 241 + L2_CACHE_BIGK_FRAGMENT_SIZE, 9); 242 + } else { 243 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL3, BANK_SELECT, 9); 244 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL3, 245 + L2_CACHE_BIGK_FRAGMENT_SIZE, 6); 246 + } 247 + WREG32_SOC15(GC, 0, regGCVM_L2_CNTL3, tmp); 248 + 249 + tmp = regGCVM_L2_CNTL4_DEFAULT; 250 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL4, VMC_TAP_PDE_REQUEST_PHYSICAL, 0); 251 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL4, VMC_TAP_PTE_REQUEST_PHYSICAL, 0); 252 + WREG32_SOC15(GC, 0, regGCVM_L2_CNTL4, tmp); 253 + 254 + tmp = regGCVM_L2_CNTL5_DEFAULT; 255 + tmp = REG_SET_FIELD(tmp, GCVM_L2_CNTL5, L2_CACHE_SMALLK_FRAGMENT_SIZE, 0); 256 + WREG32_SOC15(GC, 0, regGCVM_L2_CNTL5, tmp); 257 + } 258 + 259 + static void gfxhub_v11_5_0_enable_system_domain(struct amdgpu_device *adev) 260 + { 261 + uint32_t tmp; 262 + 263 + tmp = RREG32_SOC15(GC, 0, regGCVM_CONTEXT0_CNTL); 264 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT0_CNTL, ENABLE_CONTEXT, 1); 265 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT0_CNTL, PAGE_TABLE_DEPTH, 0); 266 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT0_CNTL, 267 + RETRY_PERMISSION_OR_INVALID_PAGE_FAULT, 0); 268 + WREG32_SOC15(GC, 0, regGCVM_CONTEXT0_CNTL, tmp); 269 + } 270 + 271 + static void gfxhub_v11_5_0_disable_identity_aperture(struct amdgpu_device *adev) 272 + { 273 + /* These registers are not accessible to VF-SRIOV. 274 + * The PF will program them instead. 275 + */ 276 + if (amdgpu_sriov_vf(adev)) 277 + return; 278 + 279 + WREG32_SOC15(GC, 0, regGCVM_L2_CONTEXT1_IDENTITY_APERTURE_LOW_ADDR_LO32, 280 + 0xFFFFFFFF); 281 + WREG32_SOC15(GC, 0, regGCVM_L2_CONTEXT1_IDENTITY_APERTURE_LOW_ADDR_HI32, 282 + 0x0000000F); 283 + 284 + WREG32_SOC15(GC, 0, regGCVM_L2_CONTEXT1_IDENTITY_APERTURE_HIGH_ADDR_LO32, 285 + 0); 286 + WREG32_SOC15(GC, 0, regGCVM_L2_CONTEXT1_IDENTITY_APERTURE_HIGH_ADDR_HI32, 287 + 0); 288 + 289 + WREG32_SOC15(GC, 0, regGCVM_L2_CONTEXT_IDENTITY_PHYSICAL_OFFSET_LO32, 0); 290 + WREG32_SOC15(GC, 0, regGCVM_L2_CONTEXT_IDENTITY_PHYSICAL_OFFSET_HI32, 0); 291 + 292 + } 293 + 294 + static void gfxhub_v11_5_0_setup_vmid_config(struct amdgpu_device *adev) 295 + { 296 + struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_GFXHUB(0)]; 297 + int i; 298 + uint32_t tmp; 299 + 300 + for (i = 0; i <= 14; i++) { 301 + tmp = RREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT1_CNTL, i); 302 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, ENABLE_CONTEXT, 1); 303 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, PAGE_TABLE_DEPTH, 304 + adev->vm_manager.num_level); 305 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 306 + RANGE_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 307 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 308 + DUMMY_PAGE_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 309 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 310 + PDE0_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 311 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 312 + VALID_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 313 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 314 + READ_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 315 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 316 + WRITE_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 317 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 318 + EXECUTE_PROTECTION_FAULT_ENABLE_DEFAULT, 1); 319 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 320 + PAGE_TABLE_BLOCK_SIZE, 321 + adev->vm_manager.block_size - 9); 322 + /* Send no-retry XNACK on fault to suppress VM fault storm. */ 323 + tmp = REG_SET_FIELD(tmp, GCVM_CONTEXT1_CNTL, 324 + RETRY_PERMISSION_OR_INVALID_PAGE_FAULT, 325 + !amdgpu_noretry); 326 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT1_CNTL, 327 + i * hub->ctx_distance, tmp); 328 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT1_PAGE_TABLE_START_ADDR_LO32, 329 + i * hub->ctx_addr_distance, 0); 330 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT1_PAGE_TABLE_START_ADDR_HI32, 331 + i * hub->ctx_addr_distance, 0); 332 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT1_PAGE_TABLE_END_ADDR_LO32, 333 + i * hub->ctx_addr_distance, 334 + lower_32_bits(adev->vm_manager.max_pfn - 1)); 335 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT1_PAGE_TABLE_END_ADDR_HI32, 336 + i * hub->ctx_addr_distance, 337 + upper_32_bits(adev->vm_manager.max_pfn - 1)); 338 + } 339 + 340 + hub->vm_cntx_cntl = tmp; 341 + } 342 + 343 + static void gfxhub_v11_5_0_program_invalidation(struct amdgpu_device *adev) 344 + { 345 + struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_GFXHUB(0)]; 346 + unsigned i; 347 + 348 + for (i = 0 ; i < 18; ++i) { 349 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_INVALIDATE_ENG0_ADDR_RANGE_LO32, 350 + i * hub->eng_addr_distance, 0xffffffff); 351 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_INVALIDATE_ENG0_ADDR_RANGE_HI32, 352 + i * hub->eng_addr_distance, 0x1f); 353 + } 354 + } 355 + 356 + static int gfxhub_v11_5_0_gart_enable(struct amdgpu_device *adev) 357 + { 358 + if (amdgpu_sriov_vf(adev)) { 359 + /* 360 + * GCMC_VM_FB_LOCATION_BASE/TOP is NULL for VF, becuase they are 361 + * VF copy registers so vbios post doesn't program them, for 362 + * SRIOV driver need to program them 363 + */ 364 + WREG32_SOC15(GC, 0, regGCMC_VM_FB_LOCATION_BASE, 365 + adev->gmc.vram_start >> 24); 366 + WREG32_SOC15(GC, 0, regGCMC_VM_FB_LOCATION_TOP, 367 + adev->gmc.vram_end >> 24); 368 + } 369 + 370 + /* GART Enable. */ 371 + gfxhub_v11_5_0_init_gart_aperture_regs(adev); 372 + gfxhub_v11_5_0_init_system_aperture_regs(adev); 373 + gfxhub_v11_5_0_init_tlb_regs(adev); 374 + gfxhub_v11_5_0_init_cache_regs(adev); 375 + 376 + gfxhub_v11_5_0_enable_system_domain(adev); 377 + gfxhub_v11_5_0_disable_identity_aperture(adev); 378 + gfxhub_v11_5_0_setup_vmid_config(adev); 379 + gfxhub_v11_5_0_program_invalidation(adev); 380 + 381 + return 0; 382 + } 383 + 384 + static void gfxhub_v11_5_0_gart_disable(struct amdgpu_device *adev) 385 + { 386 + struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_GFXHUB(0)]; 387 + u32 tmp; 388 + u32 i; 389 + 390 + /* Disable all tables */ 391 + for (i = 0; i < 16; i++) 392 + WREG32_SOC15_OFFSET(GC, 0, regGCVM_CONTEXT0_CNTL, 393 + i * hub->ctx_distance, 0); 394 + 395 + /* Setup TLB control */ 396 + tmp = RREG32_SOC15(GC, 0, regGCMC_VM_MX_L1_TLB_CNTL); 397 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, ENABLE_L1_TLB, 0); 398 + tmp = REG_SET_FIELD(tmp, GCMC_VM_MX_L1_TLB_CNTL, 399 + ENABLE_ADVANCED_DRIVER_MODEL, 0); 400 + WREG32_SOC15(GC, 0, regGCMC_VM_MX_L1_TLB_CNTL, tmp); 401 + 402 + /* Setup L2 cache */ 403 + WREG32_FIELD15_PREREG(GC, 0, GCVM_L2_CNTL, ENABLE_L2_CACHE, 0); 404 + WREG32_SOC15(GC, 0, regGCVM_L2_CNTL3, 0); 405 + } 406 + 407 + /** 408 + * gfxhub_v11_5_0_set_fault_enable_default - update GART/VM fault handling 409 + * 410 + * @adev: amdgpu_device pointer 411 + * @value: true redirects VM faults to the default page 412 + */ 413 + static void gfxhub_v11_5_0_set_fault_enable_default(struct amdgpu_device *adev, 414 + bool value) 415 + { 416 + u32 tmp; 417 + 418 + /* NO halt CP when page fault */ 419 + tmp = RREG32_SOC15(GC, 0, regCP_DEBUG); 420 + tmp = REG_SET_FIELD(tmp, CP_DEBUG, CPG_UTCL1_ERROR_HALT_DISABLE, 1); 421 + WREG32_SOC15(GC, 0, regCP_DEBUG, tmp); 422 + 423 + /* These registers are not accessible to VF-SRIOV. 424 + * The PF will program them instead. 425 + */ 426 + if (amdgpu_sriov_vf(adev)) 427 + return; 428 + 429 + tmp = RREG32_SOC15(GC, 0, regGCVM_L2_PROTECTION_FAULT_CNTL); 430 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 431 + RANGE_PROTECTION_FAULT_ENABLE_DEFAULT, value); 432 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 433 + PDE0_PROTECTION_FAULT_ENABLE_DEFAULT, value); 434 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 435 + PDE1_PROTECTION_FAULT_ENABLE_DEFAULT, value); 436 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 437 + PDE2_PROTECTION_FAULT_ENABLE_DEFAULT, value); 438 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 439 + TRANSLATE_FURTHER_PROTECTION_FAULT_ENABLE_DEFAULT, 440 + value); 441 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 442 + NACK_PROTECTION_FAULT_ENABLE_DEFAULT, value); 443 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 444 + DUMMY_PAGE_PROTECTION_FAULT_ENABLE_DEFAULT, value); 445 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 446 + VALID_PROTECTION_FAULT_ENABLE_DEFAULT, value); 447 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 448 + READ_PROTECTION_FAULT_ENABLE_DEFAULT, value); 449 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 450 + WRITE_PROTECTION_FAULT_ENABLE_DEFAULT, value); 451 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 452 + EXECUTE_PROTECTION_FAULT_ENABLE_DEFAULT, value); 453 + if (!value) { 454 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 455 + CRASH_ON_NO_RETRY_FAULT, 1); 456 + tmp = REG_SET_FIELD(tmp, GCVM_L2_PROTECTION_FAULT_CNTL, 457 + CRASH_ON_RETRY_FAULT, 1); 458 + } 459 + WREG32_SOC15(GC, 0, regGCVM_L2_PROTECTION_FAULT_CNTL, tmp); 460 + } 461 + 462 + static const struct amdgpu_vmhub_funcs gfxhub_v11_5_0_vmhub_funcs = { 463 + .print_l2_protection_fault_status = gfxhub_v11_5_0_print_l2_protection_fault_status, 464 + .get_invalidate_req = gfxhub_v11_5_0_get_invalidate_req, 465 + }; 466 + 467 + static void gfxhub_v11_5_0_init(struct amdgpu_device *adev) 468 + { 469 + struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_GFXHUB(0)]; 470 + 471 + hub->ctx0_ptb_addr_lo32 = 472 + SOC15_REG_OFFSET(GC, 0, 473 + regGCVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_LO32); 474 + hub->ctx0_ptb_addr_hi32 = 475 + SOC15_REG_OFFSET(GC, 0, 476 + regGCVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_HI32); 477 + hub->vm_inv_eng0_sem = 478 + SOC15_REG_OFFSET(GC, 0, regGCVM_INVALIDATE_ENG0_SEM); 479 + hub->vm_inv_eng0_req = 480 + SOC15_REG_OFFSET(GC, 0, regGCVM_INVALIDATE_ENG0_REQ); 481 + hub->vm_inv_eng0_ack = 482 + SOC15_REG_OFFSET(GC, 0, regGCVM_INVALIDATE_ENG0_ACK); 483 + hub->vm_context0_cntl = 484 + SOC15_REG_OFFSET(GC, 0, regGCVM_CONTEXT0_CNTL); 485 + hub->vm_l2_pro_fault_status = 486 + SOC15_REG_OFFSET(GC, 0, regGCVM_L2_PROTECTION_FAULT_STATUS); 487 + hub->vm_l2_pro_fault_cntl = 488 + SOC15_REG_OFFSET(GC, 0, regGCVM_L2_PROTECTION_FAULT_CNTL); 489 + 490 + hub->ctx_distance = regGCVM_CONTEXT1_CNTL - regGCVM_CONTEXT0_CNTL; 491 + hub->ctx_addr_distance = regGCVM_CONTEXT1_PAGE_TABLE_BASE_ADDR_LO32 - 492 + regGCVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_LO32; 493 + hub->eng_distance = regGCVM_INVALIDATE_ENG1_REQ - 494 + regGCVM_INVALIDATE_ENG0_REQ; 495 + hub->eng_addr_distance = regGCVM_INVALIDATE_ENG1_ADDR_RANGE_LO32 - 496 + regGCVM_INVALIDATE_ENG0_ADDR_RANGE_LO32; 497 + 498 + hub->vm_cntx_cntl_vm_fault = GCVM_CONTEXT1_CNTL__RANGE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK | 499 + GCVM_CONTEXT1_CNTL__DUMMY_PAGE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK | 500 + GCVM_CONTEXT1_CNTL__PDE0_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK | 501 + GCVM_CONTEXT1_CNTL__VALID_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK | 502 + GCVM_CONTEXT1_CNTL__READ_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK | 503 + GCVM_CONTEXT1_CNTL__WRITE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK | 504 + GCVM_CONTEXT1_CNTL__EXECUTE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK; 505 + 506 + hub->vmhub_funcs = &gfxhub_v11_5_0_vmhub_funcs; 507 + } 508 + 509 + const struct amdgpu_gfxhub_funcs gfxhub_v11_5_0_funcs = { 510 + .get_fb_location = gfxhub_v11_5_0_get_fb_location, 511 + .get_mc_fb_offset = gfxhub_v11_5_0_get_mc_fb_offset, 512 + .setup_vm_pt_regs = gfxhub_v11_5_0_setup_vm_pt_regs, 513 + .gart_enable = gfxhub_v11_5_0_gart_enable, 514 + .gart_disable = gfxhub_v11_5_0_gart_disable, 515 + .set_fault_enable_default = gfxhub_v11_5_0_set_fault_enable_default, 516 + .init = gfxhub_v11_5_0_init, 517 + };
+29
drivers/gpu/drm/amd/amdgpu/gfxhub_v11_5_0.h
··· 1 + /* 2 + * Copyright 2023 Advanced Micro Devices, Inc. 3 + * 4 + * Permission is hereby granted, free of charge, to any person obtaining a 5 + * copy of this software and associated documentation files (the "Software"), 6 + * to deal in the Software without restriction, including without limitation 7 + * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 + * and/or sell copies of the Software, and to permit persons to whom the 9 + * Software is furnished to do so, subject to the following conditions: 10 + * 11 + * The above copyright notice and this permission notice shall be included in 12 + * all copies or substantial portions of the Software. 13 + * 14 + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 + * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 + * OTHER DEALINGS IN THE SOFTWARE. 21 + * 22 + */ 23 + 24 + #ifndef __GFXHUB_V11_5_0_H__ 25 + #define __GFXHUB_V11_5_0_H__ 26 + 27 + extern const struct amdgpu_gfxhub_funcs gfxhub_v11_5_0_funcs; 28 + 29 + #endif
+4
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
··· 42 42 #include "nbio_v4_3.h" 43 43 #include "gfxhub_v3_0.h" 44 44 #include "gfxhub_v3_0_3.h" 45 + #include "gfxhub_v11_5_0.h" 45 46 #include "mmhub_v3_0.h" 46 47 #include "mmhub_v3_0_1.h" 47 48 #include "mmhub_v3_0_2.h" ··· 633 632 switch (adev->ip_versions[GC_HWIP][0]) { 634 633 case IP_VERSION(11, 0, 3): 635 634 adev->gfxhub.funcs = &gfxhub_v3_0_3_funcs; 635 + break; 636 + case IP_VERSION(11, 5, 0): 637 + adev->gfxhub.funcs = &gfxhub_v11_5_0_funcs; 636 638 break; 637 639 default: 638 640 adev->gfxhub.funcs = &gfxhub_v3_0_funcs;