Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
  1// SPDX-License-Identifier: MIT
  2
  3/*
  4 * Copyright © 2019 Intel Corporation
  5 */
  6
  7#include <linux/seq_file.h>
  8#include <linux/string_helpers.h>
  9
 10#include "i915_drv.h"
 11#include "i915_reg.h"
 12#include "intel_gt.h"
 13#include "intel_gt_clock_utils.h"
 14#include "intel_gt_debugfs.h"
 15#include "intel_gt_pm.h"
 16#include "intel_gt_pm_debugfs.h"
 17#include "intel_gt_regs.h"
 18#include "intel_llc.h"
 19#include "intel_mchbar_regs.h"
 20#include "intel_pcode.h"
 21#include "intel_rc6.h"
 22#include "intel_rps.h"
 23#include "intel_runtime_pm.h"
 24#include "intel_uncore.h"
 25#include "vlv_sideband.h"
 26
 27void intel_gt_pm_debugfs_forcewake_user_open(struct intel_gt *gt)
 28{
 29	atomic_inc(&gt->user_wakeref);
 30	intel_gt_pm_get_untracked(gt);
 31	if (GRAPHICS_VER(gt->i915) >= 6)
 32		intel_uncore_forcewake_user_get(gt->uncore);
 33}
 34
 35void intel_gt_pm_debugfs_forcewake_user_release(struct intel_gt *gt)
 36{
 37	if (GRAPHICS_VER(gt->i915) >= 6)
 38		intel_uncore_forcewake_user_put(gt->uncore);
 39	intel_gt_pm_put_untracked(gt);
 40	atomic_dec(&gt->user_wakeref);
 41}
 42
 43static int forcewake_user_open(struct inode *inode, struct file *file)
 44{
 45	struct intel_gt *gt = inode->i_private;
 46
 47	intel_gt_pm_debugfs_forcewake_user_open(gt);
 48
 49	return 0;
 50}
 51
 52static int forcewake_user_release(struct inode *inode, struct file *file)
 53{
 54	struct intel_gt *gt = inode->i_private;
 55
 56	intel_gt_pm_debugfs_forcewake_user_release(gt);
 57
 58	return 0;
 59}
 60
 61static const struct file_operations forcewake_user_fops = {
 62	.owner = THIS_MODULE,
 63	.open = forcewake_user_open,
 64	.release = forcewake_user_release,
 65};
 66
 67static int fw_domains_show(struct seq_file *m, void *data)
 68{
 69	struct intel_gt *gt = m->private;
 70	struct intel_uncore *uncore = gt->uncore;
 71	struct intel_uncore_forcewake_domain *fw_domain;
 72	unsigned int tmp;
 73
 74	spin_lock_irq(&uncore->lock);
 75
 76	seq_printf(m, "user.bypass_count = %u\n",
 77		   uncore->user_forcewake_count);
 78
 79	for_each_fw_domain(fw_domain, uncore, tmp)
 80		seq_printf(m, "%s.wake_count = %u\n",
 81			   intel_uncore_forcewake_domain_to_str(fw_domain->id),
 82			   READ_ONCE(fw_domain->wake_count));
 83
 84	spin_unlock_irq(&uncore->lock);
 85
 86	return 0;
 87}
 88DEFINE_INTEL_GT_DEBUGFS_ATTRIBUTE(fw_domains);
 89
 90static int vlv_drpc(struct seq_file *m)
 91{
 92	struct intel_gt *gt = m->private;
 93	struct intel_uncore *uncore = gt->uncore;
 94	u32 rcctl1, pw_status, mt_fwake_req;
 95
 96	mt_fwake_req = intel_uncore_read_fw(uncore, FORCEWAKE_MT);
 97	pw_status = intel_uncore_read(uncore, VLV_GTLC_PW_STATUS);
 98	rcctl1 = intel_uncore_read(uncore, GEN6_RC_CONTROL);
 99
100	seq_printf(m, "RC6 Enabled: %s\n",
101		   str_yes_no(rcctl1 & (GEN7_RC_CTL_TO_MODE |
102					GEN6_RC_CTL_EI_MODE(1))));
103	seq_printf(m, "Multi-threaded Forcewake Request: 0x%x\n", mt_fwake_req);
104	seq_printf(m, "Render Power Well: %s\n",
105		   (pw_status & VLV_GTLC_PW_RENDER_STATUS_MASK) ? "Up" : "Down");
106	seq_printf(m, "Media Power Well: %s\n",
107		   (pw_status & VLV_GTLC_PW_MEDIA_STATUS_MASK) ? "Up" : "Down");
108
109	intel_rc6_print_residency(m, "Render RC6 residency since boot:", INTEL_RC6_RES_RC6);
110	intel_rc6_print_residency(m, "Media RC6 residency since boot:", INTEL_RC6_RES_VLV_MEDIA);
111
112	return fw_domains_show(m, NULL);
113}
114
115static int gen6_drpc(struct seq_file *m)
116{
117	struct intel_gt *gt = m->private;
118	struct drm_i915_private *i915 = gt->i915;
119	struct intel_uncore *uncore = gt->uncore;
120	u32 gt_core_status, mt_fwake_req, rcctl1, rc6vids = 0;
121	u32 gen9_powergate_enable = 0, gen9_powergate_status = 0;
122
123	mt_fwake_req = intel_uncore_read_fw(uncore, FORCEWAKE_MT);
124	gt_core_status = intel_uncore_read_fw(uncore, GEN6_GT_CORE_STATUS);
125
126	rcctl1 = intel_uncore_read(uncore, GEN6_RC_CONTROL);
127	if (GRAPHICS_VER(i915) >= 9) {
128		gen9_powergate_enable =
129			intel_uncore_read(uncore, GEN9_PG_ENABLE);
130		gen9_powergate_status =
131			intel_uncore_read(uncore, GEN9_PWRGT_DOMAIN_STATUS);
132	}
133
134	if (GRAPHICS_VER(i915) <= 7)
135		snb_pcode_read(gt->uncore, GEN6_PCODE_READ_RC6VIDS, &rc6vids, NULL);
136
137	seq_printf(m, "RC1e Enabled: %s\n",
138		   str_yes_no(rcctl1 & GEN6_RC_CTL_RC1e_ENABLE));
139	seq_printf(m, "RC6 Enabled: %s\n",
140		   str_yes_no(rcctl1 & GEN6_RC_CTL_RC6_ENABLE));
141	if (GRAPHICS_VER(i915) >= 9) {
142		seq_printf(m, "Render Well Gating Enabled: %s\n",
143			   str_yes_no(gen9_powergate_enable & GEN9_RENDER_PG_ENABLE));
144		seq_printf(m, "Media Well Gating Enabled: %s\n",
145			   str_yes_no(gen9_powergate_enable & GEN9_MEDIA_PG_ENABLE));
146	}
147	seq_printf(m, "Deep RC6 Enabled: %s\n",
148		   str_yes_no(rcctl1 & GEN6_RC_CTL_RC6p_ENABLE));
149	seq_printf(m, "Deepest RC6 Enabled: %s\n",
150		   str_yes_no(rcctl1 & GEN6_RC_CTL_RC6pp_ENABLE));
151	seq_puts(m, "Current RC state: ");
152	switch (gt_core_status & GEN6_RCn_MASK) {
153	case GEN6_RC0:
154		if (gt_core_status & GEN6_CORE_CPD_STATE_MASK)
155			seq_puts(m, "Core Power Down\n");
156		else
157			seq_puts(m, "on\n");
158		break;
159	case GEN6_RC3:
160		seq_puts(m, "RC3\n");
161		break;
162	case GEN6_RC6:
163		seq_puts(m, "RC6\n");
164		break;
165	case GEN6_RC7:
166		seq_puts(m, "RC7\n");
167		break;
168	default:
169		seq_puts(m, "Unknown\n");
170		break;
171	}
172
173	seq_printf(m, "Core Power Down: %s\n",
174		   str_yes_no(gt_core_status & GEN6_CORE_CPD_STATE_MASK));
175	seq_printf(m, "Multi-threaded Forcewake Request: 0x%x\n", mt_fwake_req);
176	if (GRAPHICS_VER(i915) >= 9) {
177		seq_printf(m, "Render Power Well: %s\n",
178			   (gen9_powergate_status &
179			    GEN9_PWRGT_RENDER_STATUS_MASK) ? "Up" : "Down");
180		seq_printf(m, "Media Power Well: %s\n",
181			   (gen9_powergate_status &
182			    GEN9_PWRGT_MEDIA_STATUS_MASK) ? "Up" : "Down");
183	}
184
185	/* Not exactly sure what this is */
186	intel_rc6_print_residency(m, "RC6 \"Locked to RPn\" residency since boot:",
187				  INTEL_RC6_RES_RC6_LOCKED);
188	intel_rc6_print_residency(m, "RC6 residency since boot:", INTEL_RC6_RES_RC6);
189	intel_rc6_print_residency(m, "RC6+ residency since boot:", INTEL_RC6_RES_RC6p);
190	intel_rc6_print_residency(m, "RC6++ residency since boot:", INTEL_RC6_RES_RC6pp);
191
192	if (GRAPHICS_VER(i915) <= 7) {
193		seq_printf(m, "RC6   voltage: %dmV\n",
194			   GEN6_DECODE_RC6_VID(((rc6vids >> 0) & 0xff)));
195		seq_printf(m, "RC6+  voltage: %dmV\n",
196			   GEN6_DECODE_RC6_VID(((rc6vids >> 8) & 0xff)));
197		seq_printf(m, "RC6++ voltage: %dmV\n",
198			   GEN6_DECODE_RC6_VID(((rc6vids >> 16) & 0xff)));
199	}
200
201	return fw_domains_show(m, NULL);
202}
203
204static int ilk_drpc(struct seq_file *m)
205{
206	struct intel_gt *gt = m->private;
207	struct intel_uncore *uncore = gt->uncore;
208	u32 rgvmodectl, rstdbyctl;
209	u16 crstandvid;
210
211	rgvmodectl = intel_uncore_read(uncore, MEMMODECTL);
212	rstdbyctl = intel_uncore_read(uncore, RSTDBYCTL);
213	crstandvid = intel_uncore_read16(uncore, CRSTANDVID);
214
215	seq_printf(m, "HD boost: %s\n",
216		   str_yes_no(rgvmodectl & MEMMODE_BOOST_EN));
217	seq_printf(m, "Boost freq: %d\n",
218		   (rgvmodectl & MEMMODE_BOOST_FREQ_MASK) >>
219		   MEMMODE_BOOST_FREQ_SHIFT);
220	seq_printf(m, "HW control enabled: %s\n",
221		   str_yes_no(rgvmodectl & MEMMODE_HWIDLE_EN));
222	seq_printf(m, "SW control enabled: %s\n",
223		   str_yes_no(rgvmodectl & MEMMODE_SWMODE_EN));
224	seq_printf(m, "Gated voltage change: %s\n",
225		   str_yes_no(rgvmodectl & MEMMODE_RCLK_GATE));
226	seq_printf(m, "Starting frequency: P%d\n",
227		   (rgvmodectl & MEMMODE_FSTART_MASK) >> MEMMODE_FSTART_SHIFT);
228	seq_printf(m, "Max P-state: P%d\n",
229		   (rgvmodectl & MEMMODE_FMAX_MASK) >> MEMMODE_FMAX_SHIFT);
230	seq_printf(m, "Min P-state: P%d\n", (rgvmodectl & MEMMODE_FMIN_MASK));
231	seq_printf(m, "RS1 VID: %d\n", (crstandvid & 0x3f));
232	seq_printf(m, "RS2 VID: %d\n", ((crstandvid >> 8) & 0x3f));
233	seq_printf(m, "Render standby enabled: %s\n",
234		   str_yes_no(!(rstdbyctl & RCX_SW_EXIT)));
235	seq_puts(m, "Current RS state: ");
236	switch (rstdbyctl & RSX_STATUS_MASK) {
237	case RSX_STATUS_ON:
238		seq_puts(m, "on\n");
239		break;
240	case RSX_STATUS_RC1:
241		seq_puts(m, "RC1\n");
242		break;
243	case RSX_STATUS_RC1E:
244		seq_puts(m, "RC1E\n");
245		break;
246	case RSX_STATUS_RS1:
247		seq_puts(m, "RS1\n");
248		break;
249	case RSX_STATUS_RS2:
250		seq_puts(m, "RS2 (RC6)\n");
251		break;
252	case RSX_STATUS_RS3:
253		seq_puts(m, "RC3 (RC6+)\n");
254		break;
255	default:
256		seq_puts(m, "unknown\n");
257		break;
258	}
259
260	return 0;
261}
262
263static int mtl_drpc(struct seq_file *m)
264{
265	struct intel_gt *gt = m->private;
266	struct intel_uncore *uncore = gt->uncore;
267	u32 gt_core_status, rcctl1, mt_fwake_req;
268	u32 mtl_powergate_enable = 0, mtl_powergate_status = 0;
269
270	mt_fwake_req = intel_uncore_read_fw(uncore, FORCEWAKE_MT);
271	gt_core_status = intel_uncore_read(uncore, MTL_MIRROR_TARGET_WP1);
272
273	rcctl1 = intel_uncore_read(uncore, GEN6_RC_CONTROL);
274	mtl_powergate_enable = intel_uncore_read(uncore, GEN9_PG_ENABLE);
275	mtl_powergate_status = intel_uncore_read(uncore,
276						 GEN9_PWRGT_DOMAIN_STATUS);
277
278	seq_printf(m, "RC6 Enabled: %s\n",
279		   str_yes_no(rcctl1 & GEN6_RC_CTL_RC6_ENABLE));
280	if (gt->type == GT_MEDIA) {
281		seq_printf(m, "Media Well Gating Enabled: %s\n",
282			   str_yes_no(mtl_powergate_enable & GEN9_MEDIA_PG_ENABLE));
283	} else {
284		seq_printf(m, "Render Well Gating Enabled: %s\n",
285			   str_yes_no(mtl_powergate_enable & GEN9_RENDER_PG_ENABLE));
286	}
287
288	seq_puts(m, "Current RC state: ");
289	switch (REG_FIELD_GET(MTL_CC_MASK, gt_core_status)) {
290	case MTL_CC0:
291		seq_puts(m, "RC0\n");
292		break;
293	case MTL_CC6:
294		seq_puts(m, "RC6\n");
295		break;
296	default:
297		seq_puts(m, "Unknown\n");
298		break;
299	}
300
301	seq_printf(m, "Multi-threaded Forcewake Request: 0x%x\n", mt_fwake_req);
302	if (gt->type == GT_MEDIA)
303		seq_printf(m, "Media Power Well: %s\n",
304			   (mtl_powergate_status &
305			    GEN9_PWRGT_MEDIA_STATUS_MASK) ? "Up" : "Down");
306	else
307		seq_printf(m, "Render Power Well: %s\n",
308			   (mtl_powergate_status &
309			    GEN9_PWRGT_RENDER_STATUS_MASK) ? "Up" : "Down");
310
311	/* Works for both render and media gt's */
312	intel_rc6_print_residency(m, "RC6 residency since boot:", INTEL_RC6_RES_RC6);
313
314	return fw_domains_show(m, NULL);
315}
316
317static int drpc_show(struct seq_file *m, void *unused)
318{
319	struct intel_gt *gt = m->private;
320	struct drm_i915_private *i915 = gt->i915;
321	intel_wakeref_t wakeref;
322	int err = -ENODEV;
323
324	with_intel_runtime_pm(gt->uncore->rpm, wakeref) {
325		if (GRAPHICS_VER_FULL(i915) >= IP_VER(12, 70))
326			err = mtl_drpc(m);
327		else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))
328			err = vlv_drpc(m);
329		else if (GRAPHICS_VER(i915) >= 6)
330			err = gen6_drpc(m);
331		else
332			err = ilk_drpc(m);
333	}
334
335	return err;
336}
337DEFINE_INTEL_GT_DEBUGFS_ATTRIBUTE(drpc);
338
339void intel_gt_pm_frequency_dump(struct intel_gt *gt, struct drm_printer *p)
340{
341	struct drm_i915_private *i915 = gt->i915;
342	struct intel_uncore *uncore = gt->uncore;
343	struct intel_rps *rps = &gt->rps;
344	intel_wakeref_t wakeref;
345
346	wakeref = intel_runtime_pm_get(uncore->rpm);
347
348	if (GRAPHICS_VER(i915) == 5) {
349		u16 rgvswctl = intel_uncore_read16(uncore, MEMSWCTL);
350		u16 rgvstat = intel_uncore_read16(uncore, MEMSTAT_ILK);
351
352		drm_printf(p, "Requested P-state: %d\n", (rgvswctl >> 8) & 0xf);
353		drm_printf(p, "Requested VID: %d\n", rgvswctl & 0x3f);
354		drm_printf(p, "Current VID: %d\n", (rgvstat & MEMSTAT_VID_MASK) >>
355			   MEMSTAT_VID_SHIFT);
356		drm_printf(p, "Current P-state: %d\n",
357			   REG_FIELD_GET(MEMSTAT_PSTATE_MASK, rgvstat));
358	} else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) {
359		u32 rpmodectl, freq_sts;
360
361		rpmodectl = intel_uncore_read(uncore, GEN6_RP_CONTROL);
362		drm_printf(p, "Video Turbo Mode: %s\n",
363			   str_yes_no(rpmodectl & GEN6_RP_MEDIA_TURBO));
364		drm_printf(p, "HW control enabled: %s\n",
365			   str_yes_no(rpmodectl & GEN6_RP_ENABLE));
366		drm_printf(p, "SW control enabled: %s\n",
367			   str_yes_no((rpmodectl & GEN6_RP_MEDIA_MODE_MASK) == GEN6_RP_MEDIA_SW_MODE));
368
369		vlv_punit_get(i915);
370		freq_sts = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS);
371		vlv_punit_put(i915);
372
373		drm_printf(p, "PUNIT_REG_GPU_FREQ_STS: 0x%08x\n", freq_sts);
374
375		drm_printf(p, "actual GPU freq: %d MHz\n",
376			   intel_gpu_freq(rps, (freq_sts >> 8) & 0xff));
377
378		drm_printf(p, "current GPU freq: %d MHz\n",
379			   intel_gpu_freq(rps, rps->cur_freq));
380
381		drm_printf(p, "max GPU freq: %d MHz\n",
382			   intel_gpu_freq(rps, rps->max_freq));
383
384		drm_printf(p, "min GPU freq: %d MHz\n",
385			   intel_gpu_freq(rps, rps->min_freq));
386
387		drm_printf(p, "idle GPU freq: %d MHz\n",
388			   intel_gpu_freq(rps, rps->idle_freq));
389
390		drm_printf(p, "efficient (RPe) frequency: %d MHz\n",
391			   intel_gpu_freq(rps, rps->efficient_freq));
392	} else if (GRAPHICS_VER(i915) >= 6) {
393		gen6_rps_frequency_dump(rps, p);
394	} else {
395		drm_puts(p, "no P-state info available\n");
396	}
397
398	intel_runtime_pm_put(uncore->rpm, wakeref);
399}
400
401static int frequency_show(struct seq_file *m, void *unused)
402{
403	struct intel_gt *gt = m->private;
404	struct drm_printer p = drm_seq_file_printer(m);
405
406	intel_gt_pm_frequency_dump(gt, &p);
407
408	return 0;
409}
410DEFINE_INTEL_GT_DEBUGFS_ATTRIBUTE(frequency);
411
412static int llc_show(struct seq_file *m, void *data)
413{
414	struct intel_gt *gt = m->private;
415	struct drm_i915_private *i915 = gt->i915;
416	const bool edram = GRAPHICS_VER(i915) > 8;
417	struct intel_rps *rps = &gt->rps;
418	unsigned int max_gpu_freq, min_gpu_freq;
419	intel_wakeref_t wakeref;
420	int gpu_freq, ia_freq;
421
422	seq_printf(m, "LLC: %s\n", str_yes_no(HAS_LLC(i915)));
423	seq_printf(m, "%s: %uMB\n", edram ? "eDRAM" : "eLLC",
424		   i915->edram_size_mb);
425
426	min_gpu_freq = rps->min_freq;
427	max_gpu_freq = rps->max_freq;
428	if (IS_GEN9_BC(i915) || GRAPHICS_VER(i915) >= 11) {
429		/* Convert GT frequency to 50 HZ units */
430		min_gpu_freq /= GEN9_FREQ_SCALER;
431		max_gpu_freq /= GEN9_FREQ_SCALER;
432	}
433
434	seq_puts(m, "GPU freq (MHz)\tEffective GPU freq (MHz)\tEffective Ring freq (MHz)\n");
435
436	wakeref = intel_runtime_pm_get(gt->uncore->rpm);
437	for (gpu_freq = min_gpu_freq; gpu_freq <= max_gpu_freq; gpu_freq++) {
438		ia_freq = gpu_freq;
439		snb_pcode_read(gt->uncore, GEN6_PCODE_READ_MIN_FREQ_TABLE,
440			       &ia_freq, NULL);
441		seq_printf(m, "%d\t\t%d\t\t\t\t%d\n",
442			   intel_gpu_freq(rps,
443					  (gpu_freq *
444					   (IS_GEN9_BC(i915) ||
445					    GRAPHICS_VER(i915) >= 11 ?
446					    GEN9_FREQ_SCALER : 1))),
447			   ((ia_freq >> 0) & 0xff) * 100,
448			   ((ia_freq >> 8) & 0xff) * 100);
449	}
450	intel_runtime_pm_put(gt->uncore->rpm, wakeref);
451
452	return 0;
453}
454
455static bool llc_eval(void *data)
456{
457	struct intel_gt *gt = data;
458
459	return HAS_LLC(gt->i915);
460}
461
462DEFINE_INTEL_GT_DEBUGFS_ATTRIBUTE(llc);
463
464static const char *rps_power_to_str(unsigned int power)
465{
466	static const char * const strings[] = {
467		[LOW_POWER] = "low power",
468		[BETWEEN] = "mixed",
469		[HIGH_POWER] = "high power",
470	};
471
472	if (power >= ARRAY_SIZE(strings) || !strings[power])
473		return "unknown";
474
475	return strings[power];
476}
477
478static int rps_boost_show(struct seq_file *m, void *data)
479{
480	struct intel_gt *gt = m->private;
481	struct drm_i915_private *i915 = gt->i915;
482	struct intel_rps *rps = &gt->rps;
483
484	seq_printf(m, "RPS enabled? %s\n",
485		   str_yes_no(intel_rps_is_enabled(rps)));
486	seq_printf(m, "RPS active? %s\n",
487		   str_yes_no(intel_rps_is_active(rps)));
488	seq_printf(m, "GPU busy? %s, %llums\n",
489		   str_yes_no(gt->awake),
490		   ktime_to_ms(intel_gt_get_awake_time(gt)));
491	seq_printf(m, "Boosts outstanding? %d\n",
492		   atomic_read(&rps->num_waiters));
493	seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive));
494	seq_printf(m, "Frequency requested %d, actual %d\n",
495		   intel_gpu_freq(rps, rps->cur_freq),
496		   intel_rps_read_actual_frequency(rps));
497	seq_printf(m, "  min hard:%d, soft:%d; max soft:%d, hard:%d\n",
498		   intel_gpu_freq(rps, rps->min_freq),
499		   intel_gpu_freq(rps, rps->min_freq_softlimit),
500		   intel_gpu_freq(rps, rps->max_freq_softlimit),
501		   intel_gpu_freq(rps, rps->max_freq));
502	seq_printf(m, "  idle:%d, efficient:%d, boost:%d\n",
503		   intel_gpu_freq(rps, rps->idle_freq),
504		   intel_gpu_freq(rps, rps->efficient_freq),
505		   intel_gpu_freq(rps, rps->boost_freq));
506
507	seq_printf(m, "Wait boosts: %d\n", READ_ONCE(rps->boosts));
508
509	if (GRAPHICS_VER(i915) >= 6 && intel_rps_is_active(rps)) {
510		struct intel_uncore *uncore = gt->uncore;
511		u32 rpup, rpupei;
512		u32 rpdown, rpdownei;
513
514		intel_uncore_forcewake_get(uncore, FORCEWAKE_ALL);
515		rpup = intel_uncore_read_fw(uncore, GEN6_RP_CUR_UP) & GEN6_RP_EI_MASK;
516		rpupei = intel_uncore_read_fw(uncore, GEN6_RP_CUR_UP_EI) & GEN6_RP_EI_MASK;
517		rpdown = intel_uncore_read_fw(uncore, GEN6_RP_CUR_DOWN) & GEN6_RP_EI_MASK;
518		rpdownei = intel_uncore_read_fw(uncore, GEN6_RP_CUR_DOWN_EI) & GEN6_RP_EI_MASK;
519		intel_uncore_forcewake_put(uncore, FORCEWAKE_ALL);
520
521		seq_printf(m, "\nRPS Autotuning (current \"%s\" window):\n",
522			   rps_power_to_str(rps->power.mode));
523		seq_printf(m, "  Avg. up: %d%% [above threshold? %d%%]\n",
524			   rpup && rpupei ? 100 * rpup / rpupei : 0,
525			   rps->power.up_threshold);
526		seq_printf(m, "  Avg. down: %d%% [below threshold? %d%%]\n",
527			   rpdown && rpdownei ? 100 * rpdown / rpdownei : 0,
528			   rps->power.down_threshold);
529	} else {
530		seq_puts(m, "\nRPS Autotuning inactive\n");
531	}
532
533	return 0;
534}
535
536static bool rps_eval(void *data)
537{
538	struct intel_gt *gt = data;
539
540	if (intel_guc_slpc_is_used(gt_to_guc(gt)))
541		return false;
542	else
543		return HAS_RPS(gt->i915);
544}
545
546DEFINE_INTEL_GT_DEBUGFS_ATTRIBUTE(rps_boost);
547
548static int perf_limit_reasons_get(void *data, u64 *val)
549{
550	struct intel_gt *gt = data;
551	intel_wakeref_t wakeref;
552
553	with_intel_runtime_pm(gt->uncore->rpm, wakeref)
554		*val = intel_uncore_read(gt->uncore, intel_gt_perf_limit_reasons_reg(gt));
555
556	return 0;
557}
558
559static int perf_limit_reasons_clear(void *data, u64 val)
560{
561	struct intel_gt *gt = data;
562	intel_wakeref_t wakeref;
563
564	/*
565	 * Clear the upper 16 "log" bits, the lower 16 "status" bits are
566	 * read-only. The upper 16 "log" bits are identical to the lower 16
567	 * "status" bits except that the "log" bits remain set until cleared.
568	 */
569	with_intel_runtime_pm(gt->uncore->rpm, wakeref)
570		intel_uncore_rmw(gt->uncore, intel_gt_perf_limit_reasons_reg(gt),
571				 GT0_PERF_LIMIT_REASONS_LOG_MASK, 0);
572
573	return 0;
574}
575
576static bool perf_limit_reasons_eval(void *data)
577{
578	struct intel_gt *gt = data;
579
580	return i915_mmio_reg_valid(intel_gt_perf_limit_reasons_reg(gt));
581}
582
583DEFINE_SIMPLE_ATTRIBUTE(perf_limit_reasons_fops, perf_limit_reasons_get,
584			perf_limit_reasons_clear, "0x%llx\n");
585
586void intel_gt_pm_debugfs_register(struct intel_gt *gt, struct dentry *root)
587{
588	static const struct intel_gt_debugfs_file files[] = {
589		{ "drpc", &drpc_fops, NULL },
590		{ "frequency", &frequency_fops, NULL },
591		{ "forcewake", &fw_domains_fops, NULL },
592		{ "forcewake_user", &forcewake_user_fops, NULL},
593		{ "llc", &llc_fops, llc_eval },
594		{ "rps_boost", &rps_boost_fops, rps_eval },
595		{ "perf_limit_reasons", &perf_limit_reasons_fops, perf_limit_reasons_eval },
596	};
597
598	intel_gt_debugfs_register_files(root, files, ARRAY_SIZE(files), gt);
599}