Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1// SPDX-License-Identifier: MIT
2/*
3 * Copyright © 2022 Intel Corporation
4 */
5
6#include "xe_debugfs.h"
7
8#include <linux/debugfs.h>
9#include <linux/fault-inject.h>
10#include <linux/string_helpers.h>
11
12#include <drm/drm_debugfs.h>
13
14#include "regs/xe_pmt.h"
15#include "xe_bo.h"
16#include "xe_device.h"
17#include "xe_force_wake.h"
18#include "xe_gt_debugfs.h"
19#include "xe_gt_printk.h"
20#include "xe_guc_ads.h"
21#include "xe_mmio.h"
22#include "xe_pm.h"
23#include "xe_psmi.h"
24#include "xe_pxp_debugfs.h"
25#include "xe_sriov.h"
26#include "xe_sriov_pf.h"
27#include "xe_sriov_vf.h"
28#include "xe_step.h"
29#include "xe_tile_debugfs.h"
30#include "xe_wa.h"
31#include "xe_vsec.h"
32
33#ifdef CONFIG_DRM_XE_DEBUG
34#include "xe_bo_evict.h"
35#include "xe_migrate.h"
36#include "xe_vm.h"
37#endif
38
39DECLARE_FAULT_ATTR(gt_reset_failure);
40DECLARE_FAULT_ATTR(inject_csc_hw_error);
41
42static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio,
43 u32 offset, const char *name, struct drm_printer *p)
44{
45 u64 residency = 0;
46 int ret;
47
48 ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev),
49 xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID),
50 &residency, offset, sizeof(residency));
51 if (ret != sizeof(residency)) {
52 drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret);
53 return;
54 }
55
56 drm_printf(p, "%s : %llu\n", name, residency);
57}
58
59static struct xe_device *node_to_xe(struct drm_info_node *node)
60{
61 return to_xe_device(node->minor->dev);
62}
63
64static int info(struct seq_file *m, void *data)
65{
66 struct xe_device *xe = node_to_xe(m->private);
67 struct drm_printer p = drm_seq_file_printer(m);
68 struct xe_gt *gt;
69 u8 id;
70
71 xe_pm_runtime_get(xe);
72
73 drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100);
74 drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100);
75 drm_printf(&p, "stepping G:%s M:%s B:%s\n",
76 xe_step_name(xe->info.step.graphics),
77 xe_step_name(xe->info.step.media),
78 xe_step_name(xe->info.step.basedie));
79 drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx));
80 drm_printf(&p, "platform %d\n", xe->info.platform);
81 drm_printf(&p, "subplatform %d\n",
82 xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0);
83 drm_printf(&p, "devid 0x%x\n", xe->info.devid);
84 drm_printf(&p, "revid %d\n", xe->info.revid);
85 drm_printf(&p, "tile_count %d\n", xe->info.tile_count);
86 drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level);
87 drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist));
88 drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs));
89 drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm));
90 drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc));
91 for_each_gt(gt, xe, id) {
92 drm_printf(&p, "gt%d force wake %d\n", id,
93 xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT));
94 drm_printf(&p, "gt%d engine_mask 0x%llx\n", id,
95 gt->info.engine_mask);
96 }
97
98 xe_pm_runtime_put(xe);
99 return 0;
100}
101
102static int sriov_info(struct seq_file *m, void *data)
103{
104 struct xe_device *xe = node_to_xe(m->private);
105 struct drm_printer p = drm_seq_file_printer(m);
106
107 xe_sriov_print_info(xe, &p);
108 return 0;
109}
110
111static int workarounds(struct xe_device *xe, struct drm_printer *p)
112{
113 xe_pm_runtime_get(xe);
114 xe_wa_device_dump(xe, p);
115 xe_pm_runtime_put(xe);
116
117 return 0;
118}
119
120static int workaround_info(struct seq_file *m, void *data)
121{
122 struct xe_device *xe = node_to_xe(m->private);
123 struct drm_printer p = drm_seq_file_printer(m);
124
125 workarounds(xe, &p);
126 return 0;
127}
128
129static int dgfx_pkg_residencies_show(struct seq_file *m, void *data)
130{
131 struct xe_device *xe;
132 struct xe_mmio *mmio;
133 struct drm_printer p;
134
135 xe = node_to_xe(m->private);
136 p = drm_seq_file_printer(m);
137 xe_pm_runtime_get(xe);
138 mmio = xe_root_tile_mmio(xe);
139 static const struct {
140 u32 offset;
141 const char *name;
142 } residencies[] = {
143 {BMG_G2_RESIDENCY_OFFSET, "Package G2"},
144 {BMG_G6_RESIDENCY_OFFSET, "Package G6"},
145 {BMG_G8_RESIDENCY_OFFSET, "Package G8"},
146 {BMG_G10_RESIDENCY_OFFSET, "Package G10"},
147 {BMG_MODS_RESIDENCY_OFFSET, "Package ModS"}
148 };
149
150 for (int i = 0; i < ARRAY_SIZE(residencies); i++)
151 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
152
153 xe_pm_runtime_put(xe);
154 return 0;
155}
156
157static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data)
158{
159 struct xe_device *xe;
160 struct xe_mmio *mmio;
161 struct drm_printer p;
162
163 xe = node_to_xe(m->private);
164 p = drm_seq_file_printer(m);
165 xe_pm_runtime_get(xe);
166 mmio = xe_root_tile_mmio(xe);
167
168 static const struct {
169 u32 offset;
170 const char *name;
171 } residencies[] = {
172 {BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"},
173 {BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"},
174 {BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"}
175 };
176
177 for (int i = 0; i < ARRAY_SIZE(residencies); i++)
178 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
179
180 xe_pm_runtime_put(xe);
181 return 0;
182}
183
184static const struct drm_info_list debugfs_list[] = {
185 {"info", info, 0},
186 { .name = "sriov_info", .show = sriov_info, },
187 { .name = "workarounds", .show = workaround_info, },
188};
189
190static const struct drm_info_list debugfs_residencies[] = {
191 { .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, },
192 { .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, },
193};
194
195static int forcewake_open(struct inode *inode, struct file *file)
196{
197 struct xe_device *xe = inode->i_private;
198 struct xe_gt *gt;
199 u8 id, last_gt;
200 unsigned int fw_ref;
201
202 xe_pm_runtime_get(xe);
203 for_each_gt(gt, xe, id) {
204 last_gt = id;
205
206 fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL);
207 if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL))
208 goto err_fw_get;
209 }
210
211 return 0;
212
213err_fw_get:
214 for_each_gt(gt, xe, id) {
215 if (id < last_gt)
216 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
217 else if (id == last_gt)
218 xe_force_wake_put(gt_to_fw(gt), fw_ref);
219 else
220 break;
221 }
222
223 xe_pm_runtime_put(xe);
224 return -ETIMEDOUT;
225}
226
227static int forcewake_release(struct inode *inode, struct file *file)
228{
229 struct xe_device *xe = inode->i_private;
230 struct xe_gt *gt;
231 u8 id;
232
233 for_each_gt(gt, xe, id)
234 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
235 xe_pm_runtime_put(xe);
236
237 return 0;
238}
239
240static const struct file_operations forcewake_all_fops = {
241 .owner = THIS_MODULE,
242 .open = forcewake_open,
243 .release = forcewake_release,
244};
245
246static ssize_t wedged_mode_show(struct file *f, char __user *ubuf,
247 size_t size, loff_t *pos)
248{
249 struct xe_device *xe = file_inode(f)->i_private;
250 char buf[32];
251 int len = 0;
252
253 len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode);
254
255 return simple_read_from_buffer(ubuf, size, pos, buf, len);
256}
257
258static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf,
259 size_t size, loff_t *pos)
260{
261 struct xe_device *xe = file_inode(f)->i_private;
262 struct xe_gt *gt;
263 u32 wedged_mode;
264 ssize_t ret;
265 u8 id;
266
267 ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode);
268 if (ret)
269 return ret;
270
271 if (wedged_mode > 2)
272 return -EINVAL;
273
274 if (xe->wedged.mode == wedged_mode)
275 return size;
276
277 xe->wedged.mode = wedged_mode;
278
279 xe_pm_runtime_get(xe);
280 for_each_gt(gt, xe, id) {
281 ret = xe_guc_ads_scheduler_policy_toggle_reset(>->uc.guc.ads);
282 if (ret) {
283 xe_gt_err(gt, "Failed to update GuC ADS scheduler policy. GuC may still cause engine reset even with wedged_mode=2\n");
284 xe_pm_runtime_put(xe);
285 return -EIO;
286 }
287 }
288 xe_pm_runtime_put(xe);
289
290 return size;
291}
292
293static const struct file_operations wedged_mode_fops = {
294 .owner = THIS_MODULE,
295 .read = wedged_mode_show,
296 .write = wedged_mode_set,
297};
298
299static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf,
300 size_t size, loff_t *pos)
301{
302 struct xe_device *xe = file_inode(f)->i_private;
303 char buf[32];
304 int len = 0;
305
306 len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms);
307
308 return simple_read_from_buffer(ubuf, size, pos, buf, len);
309}
310
311static ssize_t atomic_svm_timeslice_ms_set(struct file *f,
312 const char __user *ubuf,
313 size_t size, loff_t *pos)
314{
315 struct xe_device *xe = file_inode(f)->i_private;
316 u32 atomic_svm_timeslice_ms;
317 ssize_t ret;
318
319 ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms);
320 if (ret)
321 return ret;
322
323 xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms;
324
325 return size;
326}
327
328static const struct file_operations atomic_svm_timeslice_ms_fops = {
329 .owner = THIS_MODULE,
330 .read = atomic_svm_timeslice_ms_show,
331 .write = atomic_svm_timeslice_ms_set,
332};
333
334static ssize_t disable_late_binding_show(struct file *f, char __user *ubuf,
335 size_t size, loff_t *pos)
336{
337 struct xe_device *xe = file_inode(f)->i_private;
338 struct xe_late_bind *late_bind = &xe->late_bind;
339 char buf[32];
340 int len;
341
342 len = scnprintf(buf, sizeof(buf), "%d\n", late_bind->disable);
343
344 return simple_read_from_buffer(ubuf, size, pos, buf, len);
345}
346
347static ssize_t disable_late_binding_set(struct file *f, const char __user *ubuf,
348 size_t size, loff_t *pos)
349{
350 struct xe_device *xe = file_inode(f)->i_private;
351 struct xe_late_bind *late_bind = &xe->late_bind;
352 u32 uval;
353 ssize_t ret;
354
355 ret = kstrtouint_from_user(ubuf, size, sizeof(uval), &uval);
356 if (ret)
357 return ret;
358
359 if (uval > 1)
360 return -EINVAL;
361
362 late_bind->disable = !!uval;
363 return size;
364}
365
366static const struct file_operations disable_late_binding_fops = {
367 .owner = THIS_MODULE,
368 .read = disable_late_binding_show,
369 .write = disable_late_binding_set,
370};
371
372void xe_debugfs_register(struct xe_device *xe)
373{
374 struct ttm_device *bdev = &xe->ttm;
375 struct drm_minor *minor = xe->drm.primary;
376 struct dentry *root = minor->debugfs_root;
377 struct ttm_resource_manager *man;
378 struct xe_tile *tile;
379 struct xe_gt *gt;
380 u32 mem_type;
381 u8 tile_id;
382 u8 id;
383
384 drm_debugfs_create_files(debugfs_list,
385 ARRAY_SIZE(debugfs_list),
386 root, minor);
387
388 if (xe->info.platform == XE_BATTLEMAGE && !IS_SRIOV_VF(xe)) {
389 drm_debugfs_create_files(debugfs_residencies,
390 ARRAY_SIZE(debugfs_residencies),
391 root, minor);
392 fault_create_debugfs_attr("inject_csc_hw_error", root,
393 &inject_csc_hw_error);
394 }
395
396 debugfs_create_file("forcewake_all", 0400, root, xe,
397 &forcewake_all_fops);
398
399 debugfs_create_file("wedged_mode", 0600, root, xe,
400 &wedged_mode_fops);
401
402 debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe,
403 &atomic_svm_timeslice_ms_fops);
404
405 debugfs_create_file("disable_late_binding", 0600, root, xe,
406 &disable_late_binding_fops);
407
408 for (mem_type = XE_PL_VRAM0; mem_type <= XE_PL_VRAM1; ++mem_type) {
409 man = ttm_manager_type(bdev, mem_type);
410
411 if (man) {
412 char name[16];
413
414 snprintf(name, sizeof(name), "vram%d_mm", mem_type - XE_PL_VRAM0);
415 ttm_resource_manager_create_debugfs(man, root, name);
416 }
417 }
418
419 man = ttm_manager_type(bdev, XE_PL_TT);
420 ttm_resource_manager_create_debugfs(man, root, "gtt_mm");
421
422 man = ttm_manager_type(bdev, XE_PL_STOLEN);
423 if (man)
424 ttm_resource_manager_create_debugfs(man, root, "stolen_mm");
425
426 for_each_tile(tile, xe, tile_id)
427 xe_tile_debugfs_register(tile);
428
429 for_each_gt(gt, xe, id)
430 xe_gt_debugfs_register(gt);
431
432 xe_pxp_debugfs_register(xe->pxp);
433
434 xe_psmi_debugfs_register(xe);
435
436 fault_create_debugfs_attr("fail_gt_reset", root, >_reset_failure);
437
438 if (IS_SRIOV_PF(xe))
439 xe_sriov_pf_debugfs_register(xe, root);
440 else if (IS_SRIOV_VF(xe))
441 xe_sriov_vf_debugfs_register(xe, root);
442}