Linux Audio

Check our new training course

Yocto / OpenEmbedded training

Feb 10-13, 2025
Register
Loading...
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Copyright (C) 2017 NVIDIA CORPORATION.  All rights reserved.
  4 */
  5
  6#include <linux/dma-mapping.h>
  7#include <linux/iommu.h>
  8#include <linux/interconnect.h>
  9
 10#include <drm/drm_atomic.h>
 11#include <drm/drm_atomic_helper.h>
 12#include <drm/drm_fourcc.h>
 13#include <drm/drm_framebuffer.h>
 14#include <drm/drm_gem_atomic_helper.h>
 
 15
 16#include "dc.h"
 17#include "plane.h"
 18
 19static void tegra_plane_destroy(struct drm_plane *plane)
 20{
 21	struct tegra_plane *p = to_tegra_plane(plane);
 22
 23	drm_plane_cleanup(plane);
 24	kfree(p);
 25}
 26
 27static void tegra_plane_reset(struct drm_plane *plane)
 28{
 29	struct tegra_plane *p = to_tegra_plane(plane);
 30	struct tegra_plane_state *state;
 31	unsigned int i;
 32
 33	if (plane->state)
 34		__drm_atomic_helper_plane_destroy_state(plane->state);
 35
 36	kfree(plane->state);
 37	plane->state = NULL;
 38
 39	state = kzalloc(sizeof(*state), GFP_KERNEL);
 40	if (state) {
 41		plane->state = &state->base;
 42		plane->state->plane = plane;
 43		plane->state->zpos = p->index;
 44		plane->state->normalized_zpos = p->index;
 45
 46		for (i = 0; i < 3; i++)
 47			state->iova[i] = DMA_MAPPING_ERROR;
 48	}
 49}
 50
 51static struct drm_plane_state *
 52tegra_plane_atomic_duplicate_state(struct drm_plane *plane)
 53{
 54	struct tegra_plane_state *state = to_tegra_plane_state(plane->state);
 55	struct tegra_plane_state *copy;
 56	unsigned int i;
 57
 58	copy = kmalloc(sizeof(*copy), GFP_KERNEL);
 59	if (!copy)
 60		return NULL;
 61
 62	__drm_atomic_helper_plane_duplicate_state(plane, &copy->base);
 63	copy->tiling = state->tiling;
 64	copy->format = state->format;
 65	copy->swap = state->swap;
 66	copy->reflect_x = state->reflect_x;
 67	copy->reflect_y = state->reflect_y;
 68	copy->opaque = state->opaque;
 69	copy->total_peak_memory_bandwidth = state->total_peak_memory_bandwidth;
 70	copy->peak_memory_bandwidth = state->peak_memory_bandwidth;
 71	copy->avg_memory_bandwidth = state->avg_memory_bandwidth;
 72
 73	for (i = 0; i < 2; i++)
 74		copy->blending[i] = state->blending[i];
 75
 76	for (i = 0; i < 3; i++) {
 77		copy->iova[i] = DMA_MAPPING_ERROR;
 78		copy->map[i] = NULL;
 79	}
 80
 81	return &copy->base;
 82}
 83
 84static void tegra_plane_atomic_destroy_state(struct drm_plane *plane,
 85					     struct drm_plane_state *state)
 86{
 87	__drm_atomic_helper_plane_destroy_state(state);
 88	kfree(state);
 89}
 90
 91static bool tegra_plane_supports_sector_layout(struct drm_plane *plane)
 92{
 93	struct drm_crtc *crtc;
 94
 95	drm_for_each_crtc(crtc, plane->dev) {
 96		if (plane->possible_crtcs & drm_crtc_mask(crtc)) {
 97			struct tegra_dc *dc = to_tegra_dc(crtc);
 98
 99			if (!dc->soc->supports_sector_layout)
100				return false;
101		}
102	}
103
104	return true;
105}
106
107static bool tegra_plane_format_mod_supported(struct drm_plane *plane,
108					     uint32_t format,
109					     uint64_t modifier)
110{
111	const struct drm_format_info *info = drm_format_info(format);
112
113	if (modifier == DRM_FORMAT_MOD_LINEAR)
114		return true;
115
116	/* check for the sector layout bit */
117	if (fourcc_mod_is_vendor(modifier, NVIDIA)) {
118		if (modifier & DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT) {
119			if (!tegra_plane_supports_sector_layout(plane))
120				return false;
121		}
122	}
123
124	if (info->num_planes == 1)
125		return true;
126
127	return false;
128}
129
130const struct drm_plane_funcs tegra_plane_funcs = {
131	.update_plane = drm_atomic_helper_update_plane,
132	.disable_plane = drm_atomic_helper_disable_plane,
133	.destroy = tegra_plane_destroy,
134	.reset = tegra_plane_reset,
135	.atomic_duplicate_state = tegra_plane_atomic_duplicate_state,
136	.atomic_destroy_state = tegra_plane_atomic_destroy_state,
137	.format_mod_supported = tegra_plane_format_mod_supported,
138};
139
140static int tegra_dc_pin(struct tegra_dc *dc, struct tegra_plane_state *state)
141{
 
142	unsigned int i;
143	int err;
144
145	for (i = 0; i < state->base.fb->format->num_planes; i++) {
146		struct tegra_bo *bo = tegra_fb_get_plane(state->base.fb, i);
147		struct host1x_bo_mapping *map;
 
148
149		map = host1x_bo_pin(dc->dev, &bo->base, DMA_TO_DEVICE, &dc->client.cache);
150		if (IS_ERR(map)) {
151			err = PTR_ERR(map);
 
 
 
 
 
 
 
 
 
 
 
 
 
152			goto unpin;
153		}
154
155		if (!dc->client.group) {
 
 
 
 
156			/*
157			 * The display controller needs contiguous memory, so
158			 * fail if the buffer is discontiguous and we fail to
159			 * map its SG table to a single contiguous chunk of
160			 * I/O virtual memory.
161			 */
162			if (map->chunks > 1) {
163				err = -EINVAL;
164				goto unpin;
165			}
166
167			state->iova[i] = map->phys;
 
168		} else {
169			state->iova[i] = bo->iova;
170		}
171
172		state->map[i] = map;
173	}
174
175	return 0;
176
177unpin:
178	dev_err(dc->dev, "failed to map plane %u: %d\n", i, err);
179
180	while (i--) {
181		host1x_bo_unpin(state->map[i]);
 
 
 
 
 
 
182		state->iova[i] = DMA_MAPPING_ERROR;
183		state->map[i] = NULL;
184	}
185
186	return err;
187}
188
189static void tegra_dc_unpin(struct tegra_dc *dc, struct tegra_plane_state *state)
190{
191	unsigned int i;
192
193	for (i = 0; i < state->base.fb->format->num_planes; i++) {
194		host1x_bo_unpin(state->map[i]);
 
 
 
 
 
 
195		state->iova[i] = DMA_MAPPING_ERROR;
196		state->map[i] = NULL;
197	}
198}
199
200int tegra_plane_prepare_fb(struct drm_plane *plane,
201			   struct drm_plane_state *state)
202{
203	struct tegra_dc *dc = to_tegra_dc(state->crtc);
204	int err;
205
206	if (!state->fb)
207		return 0;
208
209	err = drm_gem_plane_helper_prepare_fb(plane, state);
210	if (err < 0)
211		return err;
212
213	return tegra_dc_pin(dc, to_tegra_plane_state(state));
214}
215
216void tegra_plane_cleanup_fb(struct drm_plane *plane,
217			    struct drm_plane_state *state)
218{
219	struct tegra_dc *dc = to_tegra_dc(state->crtc);
220
221	if (dc)
222		tegra_dc_unpin(dc, to_tegra_plane_state(state));
223}
224
225static int tegra_plane_calculate_memory_bandwidth(struct drm_plane_state *state)
226{
227	struct tegra_plane_state *tegra_state = to_tegra_plane_state(state);
228	unsigned int i, bpp, dst_w, dst_h, src_w, src_h, mul;
229	const struct tegra_dc_soc_info *soc;
230	const struct drm_format_info *fmt;
231	struct drm_crtc_state *crtc_state;
232	u64 avg_bandwidth, peak_bandwidth;
233
234	if (!state->visible)
235		return 0;
236
237	crtc_state = drm_atomic_get_new_crtc_state(state->state, state->crtc);
238	if (!crtc_state)
239		return -EINVAL;
240
241	src_w = drm_rect_width(&state->src) >> 16;
242	src_h = drm_rect_height(&state->src) >> 16;
243	dst_w = drm_rect_width(&state->dst);
244	dst_h = drm_rect_height(&state->dst);
245
246	fmt = state->fb->format;
247	soc = to_tegra_dc(state->crtc)->soc;
248
249	/*
250	 * Note that real memory bandwidth vary depending on format and
251	 * memory layout, we are not taking that into account because small
252	 * estimation error isn't important since bandwidth is rounded up
253	 * anyway.
254	 */
255	for (i = 0, bpp = 0; i < fmt->num_planes; i++) {
256		unsigned int bpp_plane = fmt->cpp[i] * 8;
257
258		/*
259		 * Sub-sampling is relevant for chroma planes only and vertical
260		 * readouts are not cached, hence only horizontal sub-sampling
261		 * matters.
262		 */
263		if (i > 0)
264			bpp_plane /= fmt->hsub;
265
266		bpp += bpp_plane;
267	}
268
269	/* average bandwidth in kbytes/sec */
270	avg_bandwidth  = min(src_w, dst_w) * min(src_h, dst_h);
271	avg_bandwidth *= drm_mode_vrefresh(&crtc_state->adjusted_mode);
272	avg_bandwidth  = DIV_ROUND_UP(avg_bandwidth * bpp, 8) + 999;
273	do_div(avg_bandwidth, 1000);
274
275	/* mode.clock in kHz, peak bandwidth in kbytes/sec */
276	peak_bandwidth = DIV_ROUND_UP(crtc_state->adjusted_mode.clock * bpp, 8);
277
278	/*
279	 * Tegra30/114 Memory Controller can't interleave DC memory requests
280	 * for the tiled windows because DC uses 16-bytes atom, while DDR3
281	 * uses 32-bytes atom.  Hence there is x2 memory overfetch for tiled
282	 * framebuffer and DDR3 on these SoCs.
283	 */
284	if (soc->plane_tiled_memory_bandwidth_x2 &&
285	    tegra_state->tiling.mode == TEGRA_BO_TILING_MODE_TILED)
286		mul = 2;
287	else
288		mul = 1;
289
290	/* ICC bandwidth in kbytes/sec */
291	tegra_state->peak_memory_bandwidth = kBps_to_icc(peak_bandwidth) * mul;
292	tegra_state->avg_memory_bandwidth  = kBps_to_icc(avg_bandwidth)  * mul;
293
294	return 0;
295}
296
297int tegra_plane_state_add(struct tegra_plane *plane,
298			  struct drm_plane_state *state)
299{
300	struct drm_crtc_state *crtc_state;
301	struct tegra_dc_state *tegra;
302	int err;
303
304	/* Propagate errors from allocation or locking failures. */
305	crtc_state = drm_atomic_get_crtc_state(state->state, state->crtc);
306	if (IS_ERR(crtc_state))
307		return PTR_ERR(crtc_state);
308
309	/* Check plane state for visibility and calculate clipping bounds */
310	err = drm_atomic_helper_check_plane_state(state, crtc_state,
311						  0, INT_MAX, true, true);
312	if (err < 0)
313		return err;
314
315	err = tegra_plane_calculate_memory_bandwidth(state);
316	if (err < 0)
317		return err;
318
319	tegra = to_dc_state(crtc_state);
320
321	tegra->planes |= WIN_A_ACT_REQ << plane->index;
322
323	return 0;
324}
325
326int tegra_plane_format(u32 fourcc, u32 *format, u32 *swap)
327{
328	/* assume no swapping of fetched data */
329	if (swap)
330		*swap = BYTE_SWAP_NOSWAP;
331
332	switch (fourcc) {
333	case DRM_FORMAT_ARGB4444:
334		*format = WIN_COLOR_DEPTH_B4G4R4A4;
335		break;
336
337	case DRM_FORMAT_ARGB1555:
338		*format = WIN_COLOR_DEPTH_B5G5R5A1;
339		break;
340
341	case DRM_FORMAT_RGB565:
342		*format = WIN_COLOR_DEPTH_B5G6R5;
343		break;
344
345	case DRM_FORMAT_RGBA5551:
346		*format = WIN_COLOR_DEPTH_A1B5G5R5;
347		break;
348
349	case DRM_FORMAT_ARGB8888:
350		*format = WIN_COLOR_DEPTH_B8G8R8A8;
351		break;
352
353	case DRM_FORMAT_ABGR8888:
354		*format = WIN_COLOR_DEPTH_R8G8B8A8;
355		break;
356
357	case DRM_FORMAT_ABGR4444:
358		*format = WIN_COLOR_DEPTH_R4G4B4A4;
359		break;
360
361	case DRM_FORMAT_ABGR1555:
362		*format = WIN_COLOR_DEPTH_R5G5B5A;
363		break;
364
365	case DRM_FORMAT_BGRA5551:
366		*format = WIN_COLOR_DEPTH_AR5G5B5;
367		break;
368
369	case DRM_FORMAT_XRGB1555:
370		*format = WIN_COLOR_DEPTH_B5G5R5X1;
371		break;
372
373	case DRM_FORMAT_RGBX5551:
374		*format = WIN_COLOR_DEPTH_X1B5G5R5;
375		break;
376
377	case DRM_FORMAT_XBGR1555:
378		*format = WIN_COLOR_DEPTH_R5G5B5X1;
379		break;
380
381	case DRM_FORMAT_BGRX5551:
382		*format = WIN_COLOR_DEPTH_X1R5G5B5;
383		break;
384
385	case DRM_FORMAT_BGR565:
386		*format = WIN_COLOR_DEPTH_R5G6B5;
387		break;
388
389	case DRM_FORMAT_BGRA8888:
390		*format = WIN_COLOR_DEPTH_A8R8G8B8;
391		break;
392
393	case DRM_FORMAT_RGBA8888:
394		*format = WIN_COLOR_DEPTH_A8B8G8R8;
395		break;
396
397	case DRM_FORMAT_XRGB8888:
398		*format = WIN_COLOR_DEPTH_B8G8R8X8;
399		break;
400
401	case DRM_FORMAT_XBGR8888:
402		*format = WIN_COLOR_DEPTH_R8G8B8X8;
403		break;
404
405	case DRM_FORMAT_UYVY:
406		*format = WIN_COLOR_DEPTH_YCbCr422;
407		break;
408
409	case DRM_FORMAT_YUYV:
410		if (!swap)
411			return -EINVAL;
412
413		*format = WIN_COLOR_DEPTH_YCbCr422;
414		*swap = BYTE_SWAP_SWAP2;
415		break;
416
417	case DRM_FORMAT_YVYU:
418		if (!swap)
419			return -EINVAL;
420
421		*format = WIN_COLOR_DEPTH_YCbCr422;
422		*swap = BYTE_SWAP_SWAP4;
423		break;
424
425	case DRM_FORMAT_VYUY:
426		if (!swap)
427			return -EINVAL;
428
429		*format = WIN_COLOR_DEPTH_YCbCr422;
430		*swap = BYTE_SWAP_SWAP4HW;
431		break;
432
433	case DRM_FORMAT_YUV420:
434		*format = WIN_COLOR_DEPTH_YCbCr420P;
435		break;
436
437	case DRM_FORMAT_YUV422:
438		*format = WIN_COLOR_DEPTH_YCbCr422P;
439		break;
440
441	case DRM_FORMAT_YUV444:
442		*format = WIN_COLOR_DEPTH_YCbCr444P;
443		break;
444
445	case DRM_FORMAT_NV12:
446		*format = WIN_COLOR_DEPTH_YCbCr420SP;
447		break;
448
449	case DRM_FORMAT_NV21:
450		*format = WIN_COLOR_DEPTH_YCrCb420SP;
451		break;
452
453	case DRM_FORMAT_NV16:
454		*format = WIN_COLOR_DEPTH_YCbCr422SP;
455		break;
456
457	case DRM_FORMAT_NV61:
458		*format = WIN_COLOR_DEPTH_YCrCb422SP;
459		break;
460
461	case DRM_FORMAT_NV24:
462		*format = WIN_COLOR_DEPTH_YCbCr444SP;
463		break;
464
465	case DRM_FORMAT_NV42:
466		*format = WIN_COLOR_DEPTH_YCrCb444SP;
467		break;
468
469	default:
470		return -EINVAL;
471	}
472
473	return 0;
474}
475
476bool tegra_plane_format_is_indexed(unsigned int format)
477{
478	switch (format) {
479	case WIN_COLOR_DEPTH_P1:
480	case WIN_COLOR_DEPTH_P2:
481	case WIN_COLOR_DEPTH_P4:
482	case WIN_COLOR_DEPTH_P8:
483		return true;
484	}
485
486	return false;
487}
488
489bool tegra_plane_format_is_yuv(unsigned int format, unsigned int *planes, unsigned int *bpc)
490{
491	switch (format) {
492	case WIN_COLOR_DEPTH_YCbCr422:
493	case WIN_COLOR_DEPTH_YUV422:
494		if (planes)
495			*planes = 1;
496
497		if (bpc)
498			*bpc = 8;
499
500		return true;
501
502	case WIN_COLOR_DEPTH_YCbCr420P:
503	case WIN_COLOR_DEPTH_YUV420P:
504	case WIN_COLOR_DEPTH_YCbCr422P:
505	case WIN_COLOR_DEPTH_YUV422P:
506	case WIN_COLOR_DEPTH_YCbCr422R:
507	case WIN_COLOR_DEPTH_YUV422R:
508	case WIN_COLOR_DEPTH_YCbCr422RA:
509	case WIN_COLOR_DEPTH_YUV422RA:
510	case WIN_COLOR_DEPTH_YCbCr444P:
511		if (planes)
512			*planes = 3;
513
514		if (bpc)
515			*bpc = 8;
516
517		return true;
518
519	case WIN_COLOR_DEPTH_YCrCb420SP:
520	case WIN_COLOR_DEPTH_YCbCr420SP:
521	case WIN_COLOR_DEPTH_YCrCb422SP:
522	case WIN_COLOR_DEPTH_YCbCr422SP:
523	case WIN_COLOR_DEPTH_YCrCb444SP:
524	case WIN_COLOR_DEPTH_YCbCr444SP:
525		if (planes)
526			*planes = 2;
527
528		if (bpc)
529			*bpc = 8;
530
531		return true;
532	}
533
534	if (planes)
535		*planes = 1;
536
537	return false;
538}
539
540static bool __drm_format_has_alpha(u32 format)
541{
542	switch (format) {
543	case DRM_FORMAT_ARGB1555:
544	case DRM_FORMAT_RGBA5551:
545	case DRM_FORMAT_ABGR8888:
546	case DRM_FORMAT_ARGB8888:
547		return true;
548	}
549
550	return false;
551}
552
553static int tegra_plane_format_get_alpha(unsigned int opaque,
554					unsigned int *alpha)
555{
556	if (tegra_plane_format_is_yuv(opaque, NULL, NULL)) {
557		*alpha = opaque;
558		return 0;
559	}
560
561	switch (opaque) {
562	case WIN_COLOR_DEPTH_B5G5R5X1:
563		*alpha = WIN_COLOR_DEPTH_B5G5R5A1;
564		return 0;
565
566	case WIN_COLOR_DEPTH_X1B5G5R5:
567		*alpha = WIN_COLOR_DEPTH_A1B5G5R5;
568		return 0;
569
570	case WIN_COLOR_DEPTH_R8G8B8X8:
571		*alpha = WIN_COLOR_DEPTH_R8G8B8A8;
572		return 0;
573
574	case WIN_COLOR_DEPTH_B8G8R8X8:
575		*alpha = WIN_COLOR_DEPTH_B8G8R8A8;
576		return 0;
577
578	case WIN_COLOR_DEPTH_B5G6R5:
579		*alpha = opaque;
580		return 0;
581	}
582
583	return -EINVAL;
584}
585
586/*
587 * This is applicable to Tegra20 and Tegra30 only where the opaque formats can
588 * be emulated using the alpha formats and alpha blending disabled.
589 */
590static int tegra_plane_setup_opacity(struct tegra_plane *tegra,
591				     struct tegra_plane_state *state)
592{
593	unsigned int format;
594	int err;
595
596	switch (state->format) {
597	case WIN_COLOR_DEPTH_B5G5R5A1:
598	case WIN_COLOR_DEPTH_A1B5G5R5:
599	case WIN_COLOR_DEPTH_R8G8B8A8:
600	case WIN_COLOR_DEPTH_B8G8R8A8:
601		state->opaque = false;
602		break;
603
604	default:
605		err = tegra_plane_format_get_alpha(state->format, &format);
606		if (err < 0)
607			return err;
608
609		state->format = format;
610		state->opaque = true;
611		break;
612	}
613
614	return 0;
615}
616
617static int tegra_plane_check_transparency(struct tegra_plane *tegra,
618					  struct tegra_plane_state *state)
619{
620	struct drm_plane_state *old, *plane_state;
621	struct drm_plane *plane;
622
623	old = drm_atomic_get_old_plane_state(state->base.state, &tegra->base);
624
625	/* check if zpos / transparency changed */
626	if (old->normalized_zpos == state->base.normalized_zpos &&
627	    to_tegra_plane_state(old)->opaque == state->opaque)
628		return 0;
629
630	/* include all sibling planes into this commit */
631	drm_for_each_plane(plane, tegra->base.dev) {
632		struct tegra_plane *p = to_tegra_plane(plane);
633
634		/* skip this plane and planes on different CRTCs */
635		if (p == tegra || p->dc != tegra->dc)
636			continue;
637
638		plane_state = drm_atomic_get_plane_state(state->base.state,
639							 plane);
640		if (IS_ERR(plane_state))
641			return PTR_ERR(plane_state);
642	}
643
644	return 1;
645}
646
647static unsigned int tegra_plane_get_overlap_index(struct tegra_plane *plane,
648						  struct tegra_plane *other)
649{
650	unsigned int index = 0, i;
651
652	WARN_ON(plane == other);
653
654	for (i = 0; i < 3; i++) {
655		if (i == plane->index)
656			continue;
657
658		if (i == other->index)
659			break;
660
661		index++;
662	}
663
664	return index;
665}
666
667static void tegra_plane_update_transparency(struct tegra_plane *tegra,
668					    struct tegra_plane_state *state)
669{
670	struct drm_plane_state *new;
671	struct drm_plane *plane;
672	unsigned int i;
673
674	for_each_new_plane_in_state(state->base.state, plane, new, i) {
675		struct tegra_plane *p = to_tegra_plane(plane);
676		unsigned index;
677
678		/* skip this plane and planes on different CRTCs */
679		if (p == tegra || p->dc != tegra->dc)
680			continue;
681
682		index = tegra_plane_get_overlap_index(tegra, p);
683
684		if (new->fb && __drm_format_has_alpha(new->fb->format->format))
685			state->blending[index].alpha = true;
686		else
687			state->blending[index].alpha = false;
688
689		if (new->normalized_zpos > state->base.normalized_zpos)
690			state->blending[index].top = true;
691		else
692			state->blending[index].top = false;
693
694		/*
695		 * Missing framebuffer means that plane is disabled, in this
696		 * case mark B / C window as top to be able to differentiate
697		 * windows indices order in regards to zPos for the middle
698		 * window X / Y registers programming.
699		 */
700		if (!new->fb)
701			state->blending[index].top = (index == 1);
702	}
703}
704
705static int tegra_plane_setup_transparency(struct tegra_plane *tegra,
706					  struct tegra_plane_state *state)
707{
708	struct tegra_plane_state *tegra_state;
709	struct drm_plane_state *new;
710	struct drm_plane *plane;
711	int err;
712
713	/*
714	 * If planes zpos / transparency changed, sibling planes blending
715	 * state may require adjustment and in this case they will be included
716	 * into this atom commit, otherwise blending state is unchanged.
717	 */
718	err = tegra_plane_check_transparency(tegra, state);
719	if (err <= 0)
720		return err;
721
722	/*
723	 * All planes are now in the atomic state, walk them up and update
724	 * transparency state for each plane.
725	 */
726	drm_for_each_plane(plane, tegra->base.dev) {
727		struct tegra_plane *p = to_tegra_plane(plane);
728
729		/* skip planes on different CRTCs */
730		if (p->dc != tegra->dc)
731			continue;
732
733		new = drm_atomic_get_new_plane_state(state->base.state, plane);
734		tegra_state = to_tegra_plane_state(new);
735
736		/*
737		 * There is no need to update blending state for the disabled
738		 * plane.
739		 */
740		if (new->fb)
741			tegra_plane_update_transparency(p, tegra_state);
742	}
743
744	return 0;
745}
746
747int tegra_plane_setup_legacy_state(struct tegra_plane *tegra,
748				   struct tegra_plane_state *state)
749{
750	int err;
751
752	err = tegra_plane_setup_opacity(tegra, state);
753	if (err < 0)
754		return err;
755
756	err = tegra_plane_setup_transparency(tegra, state);
757	if (err < 0)
758		return err;
759
760	return 0;
761}
762
763static const char * const tegra_plane_icc_names[TEGRA_DC_LEGACY_PLANES_NUM] = {
764	"wina", "winb", "winc", NULL, NULL, NULL, "cursor",
765};
766
767int tegra_plane_interconnect_init(struct tegra_plane *plane)
768{
769	const char *icc_name = tegra_plane_icc_names[plane->index];
770	struct device *dev = plane->dc->dev;
771	struct tegra_dc *dc = plane->dc;
772	int err;
773
774	if (WARN_ON(plane->index >= TEGRA_DC_LEGACY_PLANES_NUM) ||
775	    WARN_ON(!tegra_plane_icc_names[plane->index]))
776		return -EINVAL;
777
778	plane->icc_mem = devm_of_icc_get(dev, icc_name);
779	err = PTR_ERR_OR_ZERO(plane->icc_mem);
780	if (err)
781		return dev_err_probe(dev, err, "failed to get %s interconnect\n",
782				     icc_name);
783
784	/* plane B on T20/30 has a dedicated memory client for a 6-tap vertical filter */
785	if (plane->index == 1 && dc->soc->has_win_b_vfilter_mem_client) {
786		plane->icc_mem_vfilter = devm_of_icc_get(dev, "winb-vfilter");
787		err = PTR_ERR_OR_ZERO(plane->icc_mem_vfilter);
788		if (err)
789			return dev_err_probe(dev, err, "failed to get %s interconnect\n",
790					     "winb-vfilter");
791	}
792
793	return 0;
794}
v5.14.15
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 * Copyright (C) 2017 NVIDIA CORPORATION.  All rights reserved.
  4 */
  5
 
  6#include <linux/iommu.h>
 
  7
  8#include <drm/drm_atomic.h>
  9#include <drm/drm_atomic_helper.h>
 10#include <drm/drm_fourcc.h>
 
 11#include <drm/drm_gem_atomic_helper.h>
 12#include <drm/drm_plane_helper.h>
 13
 14#include "dc.h"
 15#include "plane.h"
 16
 17static void tegra_plane_destroy(struct drm_plane *plane)
 18{
 19	struct tegra_plane *p = to_tegra_plane(plane);
 20
 21	drm_plane_cleanup(plane);
 22	kfree(p);
 23}
 24
 25static void tegra_plane_reset(struct drm_plane *plane)
 26{
 27	struct tegra_plane *p = to_tegra_plane(plane);
 28	struct tegra_plane_state *state;
 29	unsigned int i;
 30
 31	if (plane->state)
 32		__drm_atomic_helper_plane_destroy_state(plane->state);
 33
 34	kfree(plane->state);
 35	plane->state = NULL;
 36
 37	state = kzalloc(sizeof(*state), GFP_KERNEL);
 38	if (state) {
 39		plane->state = &state->base;
 40		plane->state->plane = plane;
 41		plane->state->zpos = p->index;
 42		plane->state->normalized_zpos = p->index;
 43
 44		for (i = 0; i < 3; i++)
 45			state->iova[i] = DMA_MAPPING_ERROR;
 46	}
 47}
 48
 49static struct drm_plane_state *
 50tegra_plane_atomic_duplicate_state(struct drm_plane *plane)
 51{
 52	struct tegra_plane_state *state = to_tegra_plane_state(plane->state);
 53	struct tegra_plane_state *copy;
 54	unsigned int i;
 55
 56	copy = kmalloc(sizeof(*copy), GFP_KERNEL);
 57	if (!copy)
 58		return NULL;
 59
 60	__drm_atomic_helper_plane_duplicate_state(plane, &copy->base);
 61	copy->tiling = state->tiling;
 62	copy->format = state->format;
 63	copy->swap = state->swap;
 64	copy->reflect_x = state->reflect_x;
 65	copy->reflect_y = state->reflect_y;
 66	copy->opaque = state->opaque;
 
 
 
 67
 68	for (i = 0; i < 2; i++)
 69		copy->blending[i] = state->blending[i];
 70
 71	for (i = 0; i < 3; i++) {
 72		copy->iova[i] = DMA_MAPPING_ERROR;
 73		copy->sgt[i] = NULL;
 74	}
 75
 76	return &copy->base;
 77}
 78
 79static void tegra_plane_atomic_destroy_state(struct drm_plane *plane,
 80					     struct drm_plane_state *state)
 81{
 82	__drm_atomic_helper_plane_destroy_state(state);
 83	kfree(state);
 84}
 85
 86static bool tegra_plane_supports_sector_layout(struct drm_plane *plane)
 87{
 88	struct drm_crtc *crtc;
 89
 90	drm_for_each_crtc(crtc, plane->dev) {
 91		if (plane->possible_crtcs & drm_crtc_mask(crtc)) {
 92			struct tegra_dc *dc = to_tegra_dc(crtc);
 93
 94			if (!dc->soc->supports_sector_layout)
 95				return false;
 96		}
 97	}
 98
 99	return true;
100}
101
102static bool tegra_plane_format_mod_supported(struct drm_plane *plane,
103					     uint32_t format,
104					     uint64_t modifier)
105{
106	const struct drm_format_info *info = drm_format_info(format);
107
108	if (modifier == DRM_FORMAT_MOD_LINEAR)
109		return true;
110
111	/* check for the sector layout bit */
112	if ((modifier >> 56) == DRM_FORMAT_MOD_VENDOR_NVIDIA) {
113		if (modifier & DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT) {
114			if (!tegra_plane_supports_sector_layout(plane))
115				return false;
116		}
117	}
118
119	if (info->num_planes == 1)
120		return true;
121
122	return false;
123}
124
125const struct drm_plane_funcs tegra_plane_funcs = {
126	.update_plane = drm_atomic_helper_update_plane,
127	.disable_plane = drm_atomic_helper_disable_plane,
128	.destroy = tegra_plane_destroy,
129	.reset = tegra_plane_reset,
130	.atomic_duplicate_state = tegra_plane_atomic_duplicate_state,
131	.atomic_destroy_state = tegra_plane_atomic_destroy_state,
132	.format_mod_supported = tegra_plane_format_mod_supported,
133};
134
135static int tegra_dc_pin(struct tegra_dc *dc, struct tegra_plane_state *state)
136{
137	struct iommu_domain *domain = iommu_get_domain_for_dev(dc->dev);
138	unsigned int i;
139	int err;
140
141	for (i = 0; i < state->base.fb->format->num_planes; i++) {
142		struct tegra_bo *bo = tegra_fb_get_plane(state->base.fb, i);
143		dma_addr_t phys_addr, *phys;
144		struct sg_table *sgt;
145
146		/*
147		 * If we're not attached to a domain, we already stored the
148		 * physical address when the buffer was allocated. If we're
149		 * part of a group that's shared between all display
150		 * controllers, we've also already mapped the framebuffer
151		 * through the SMMU. In both cases we can short-circuit the
152		 * code below and retrieve the stored IOV address.
153		 */
154		if (!domain || dc->client.group)
155			phys = &phys_addr;
156		else
157			phys = NULL;
158
159		sgt = host1x_bo_pin(dc->dev, &bo->base, phys);
160		if (IS_ERR(sgt)) {
161			err = PTR_ERR(sgt);
162			goto unpin;
163		}
164
165		if (sgt) {
166			err = dma_map_sgtable(dc->dev, sgt, DMA_TO_DEVICE, 0);
167			if (err)
168				goto unpin;
169
170			/*
171			 * The display controller needs contiguous memory, so
172			 * fail if the buffer is discontiguous and we fail to
173			 * map its SG table to a single contiguous chunk of
174			 * I/O virtual memory.
175			 */
176			if (sgt->nents > 1) {
177				err = -EINVAL;
178				goto unpin;
179			}
180
181			state->iova[i] = sg_dma_address(sgt->sgl);
182			state->sgt[i] = sgt;
183		} else {
184			state->iova[i] = phys_addr;
185		}
 
 
186	}
187
188	return 0;
189
190unpin:
191	dev_err(dc->dev, "failed to map plane %u: %d\n", i, err);
192
193	while (i--) {
194		struct tegra_bo *bo = tegra_fb_get_plane(state->base.fb, i);
195		struct sg_table *sgt = state->sgt[i];
196
197		if (sgt)
198			dma_unmap_sgtable(dc->dev, sgt, DMA_TO_DEVICE, 0);
199
200		host1x_bo_unpin(dc->dev, &bo->base, sgt);
201		state->iova[i] = DMA_MAPPING_ERROR;
202		state->sgt[i] = NULL;
203	}
204
205	return err;
206}
207
208static void tegra_dc_unpin(struct tegra_dc *dc, struct tegra_plane_state *state)
209{
210	unsigned int i;
211
212	for (i = 0; i < state->base.fb->format->num_planes; i++) {
213		struct tegra_bo *bo = tegra_fb_get_plane(state->base.fb, i);
214		struct sg_table *sgt = state->sgt[i];
215
216		if (sgt)
217			dma_unmap_sgtable(dc->dev, sgt, DMA_TO_DEVICE, 0);
218
219		host1x_bo_unpin(dc->dev, &bo->base, sgt);
220		state->iova[i] = DMA_MAPPING_ERROR;
221		state->sgt[i] = NULL;
222	}
223}
224
225int tegra_plane_prepare_fb(struct drm_plane *plane,
226			   struct drm_plane_state *state)
227{
228	struct tegra_dc *dc = to_tegra_dc(state->crtc);
 
229
230	if (!state->fb)
231		return 0;
232
233	drm_gem_plane_helper_prepare_fb(plane, state);
 
 
234
235	return tegra_dc_pin(dc, to_tegra_plane_state(state));
236}
237
238void tegra_plane_cleanup_fb(struct drm_plane *plane,
239			    struct drm_plane_state *state)
240{
241	struct tegra_dc *dc = to_tegra_dc(state->crtc);
242
243	if (dc)
244		tegra_dc_unpin(dc, to_tegra_plane_state(state));
245}
246
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
247int tegra_plane_state_add(struct tegra_plane *plane,
248			  struct drm_plane_state *state)
249{
250	struct drm_crtc_state *crtc_state;
251	struct tegra_dc_state *tegra;
252	int err;
253
254	/* Propagate errors from allocation or locking failures. */
255	crtc_state = drm_atomic_get_crtc_state(state->state, state->crtc);
256	if (IS_ERR(crtc_state))
257		return PTR_ERR(crtc_state);
258
259	/* Check plane state for visibility and calculate clipping bounds */
260	err = drm_atomic_helper_check_plane_state(state, crtc_state,
261						  0, INT_MAX, true, true);
262	if (err < 0)
263		return err;
264
 
 
 
 
265	tegra = to_dc_state(crtc_state);
266
267	tegra->planes |= WIN_A_ACT_REQ << plane->index;
268
269	return 0;
270}
271
272int tegra_plane_format(u32 fourcc, u32 *format, u32 *swap)
273{
274	/* assume no swapping of fetched data */
275	if (swap)
276		*swap = BYTE_SWAP_NOSWAP;
277
278	switch (fourcc) {
279	case DRM_FORMAT_ARGB4444:
280		*format = WIN_COLOR_DEPTH_B4G4R4A4;
281		break;
282
283	case DRM_FORMAT_ARGB1555:
284		*format = WIN_COLOR_DEPTH_B5G5R5A1;
285		break;
286
287	case DRM_FORMAT_RGB565:
288		*format = WIN_COLOR_DEPTH_B5G6R5;
289		break;
290
291	case DRM_FORMAT_RGBA5551:
292		*format = WIN_COLOR_DEPTH_A1B5G5R5;
293		break;
294
295	case DRM_FORMAT_ARGB8888:
296		*format = WIN_COLOR_DEPTH_B8G8R8A8;
297		break;
298
299	case DRM_FORMAT_ABGR8888:
300		*format = WIN_COLOR_DEPTH_R8G8B8A8;
301		break;
302
303	case DRM_FORMAT_ABGR4444:
304		*format = WIN_COLOR_DEPTH_R4G4B4A4;
305		break;
306
307	case DRM_FORMAT_ABGR1555:
308		*format = WIN_COLOR_DEPTH_R5G5B5A;
309		break;
310
311	case DRM_FORMAT_BGRA5551:
312		*format = WIN_COLOR_DEPTH_AR5G5B5;
313		break;
314
315	case DRM_FORMAT_XRGB1555:
316		*format = WIN_COLOR_DEPTH_B5G5R5X1;
317		break;
318
319	case DRM_FORMAT_RGBX5551:
320		*format = WIN_COLOR_DEPTH_X1B5G5R5;
321		break;
322
323	case DRM_FORMAT_XBGR1555:
324		*format = WIN_COLOR_DEPTH_R5G5B5X1;
325		break;
326
327	case DRM_FORMAT_BGRX5551:
328		*format = WIN_COLOR_DEPTH_X1R5G5B5;
329		break;
330
331	case DRM_FORMAT_BGR565:
332		*format = WIN_COLOR_DEPTH_R5G6B5;
333		break;
334
335	case DRM_FORMAT_BGRA8888:
336		*format = WIN_COLOR_DEPTH_A8R8G8B8;
337		break;
338
339	case DRM_FORMAT_RGBA8888:
340		*format = WIN_COLOR_DEPTH_A8B8G8R8;
341		break;
342
343	case DRM_FORMAT_XRGB8888:
344		*format = WIN_COLOR_DEPTH_B8G8R8X8;
345		break;
346
347	case DRM_FORMAT_XBGR8888:
348		*format = WIN_COLOR_DEPTH_R8G8B8X8;
349		break;
350
351	case DRM_FORMAT_UYVY:
352		*format = WIN_COLOR_DEPTH_YCbCr422;
353		break;
354
355	case DRM_FORMAT_YUYV:
356		if (!swap)
357			return -EINVAL;
358
359		*format = WIN_COLOR_DEPTH_YCbCr422;
360		*swap = BYTE_SWAP_SWAP2;
361		break;
362
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
363	case DRM_FORMAT_YUV420:
364		*format = WIN_COLOR_DEPTH_YCbCr420P;
365		break;
366
367	case DRM_FORMAT_YUV422:
368		*format = WIN_COLOR_DEPTH_YCbCr422P;
369		break;
370
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
371	default:
372		return -EINVAL;
373	}
374
375	return 0;
376}
377
378bool tegra_plane_format_is_indexed(unsigned int format)
379{
380	switch (format) {
381	case WIN_COLOR_DEPTH_P1:
382	case WIN_COLOR_DEPTH_P2:
383	case WIN_COLOR_DEPTH_P4:
384	case WIN_COLOR_DEPTH_P8:
385		return true;
386	}
387
388	return false;
389}
390
391bool tegra_plane_format_is_yuv(unsigned int format, bool *planar, unsigned int *bpc)
392{
393	switch (format) {
394	case WIN_COLOR_DEPTH_YCbCr422:
395	case WIN_COLOR_DEPTH_YUV422:
396		if (planar)
397			*planar = false;
398
399		if (bpc)
400			*bpc = 8;
401
402		return true;
403
404	case WIN_COLOR_DEPTH_YCbCr420P:
405	case WIN_COLOR_DEPTH_YUV420P:
406	case WIN_COLOR_DEPTH_YCbCr422P:
407	case WIN_COLOR_DEPTH_YUV422P:
408	case WIN_COLOR_DEPTH_YCbCr422R:
409	case WIN_COLOR_DEPTH_YUV422R:
410	case WIN_COLOR_DEPTH_YCbCr422RA:
411	case WIN_COLOR_DEPTH_YUV422RA:
412		if (planar)
413			*planar = true;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
414
415		if (bpc)
416			*bpc = 8;
417
418		return true;
419	}
420
421	if (planar)
422		*planar = false;
423
424	return false;
425}
426
427static bool __drm_format_has_alpha(u32 format)
428{
429	switch (format) {
430	case DRM_FORMAT_ARGB1555:
431	case DRM_FORMAT_RGBA5551:
432	case DRM_FORMAT_ABGR8888:
433	case DRM_FORMAT_ARGB8888:
434		return true;
435	}
436
437	return false;
438}
439
440static int tegra_plane_format_get_alpha(unsigned int opaque,
441					unsigned int *alpha)
442{
443	if (tegra_plane_format_is_yuv(opaque, NULL, NULL)) {
444		*alpha = opaque;
445		return 0;
446	}
447
448	switch (opaque) {
449	case WIN_COLOR_DEPTH_B5G5R5X1:
450		*alpha = WIN_COLOR_DEPTH_B5G5R5A1;
451		return 0;
452
453	case WIN_COLOR_DEPTH_X1B5G5R5:
454		*alpha = WIN_COLOR_DEPTH_A1B5G5R5;
455		return 0;
456
457	case WIN_COLOR_DEPTH_R8G8B8X8:
458		*alpha = WIN_COLOR_DEPTH_R8G8B8A8;
459		return 0;
460
461	case WIN_COLOR_DEPTH_B8G8R8X8:
462		*alpha = WIN_COLOR_DEPTH_B8G8R8A8;
463		return 0;
464
465	case WIN_COLOR_DEPTH_B5G6R5:
466		*alpha = opaque;
467		return 0;
468	}
469
470	return -EINVAL;
471}
472
473/*
474 * This is applicable to Tegra20 and Tegra30 only where the opaque formats can
475 * be emulated using the alpha formats and alpha blending disabled.
476 */
477static int tegra_plane_setup_opacity(struct tegra_plane *tegra,
478				     struct tegra_plane_state *state)
479{
480	unsigned int format;
481	int err;
482
483	switch (state->format) {
484	case WIN_COLOR_DEPTH_B5G5R5A1:
485	case WIN_COLOR_DEPTH_A1B5G5R5:
486	case WIN_COLOR_DEPTH_R8G8B8A8:
487	case WIN_COLOR_DEPTH_B8G8R8A8:
488		state->opaque = false;
489		break;
490
491	default:
492		err = tegra_plane_format_get_alpha(state->format, &format);
493		if (err < 0)
494			return err;
495
496		state->format = format;
497		state->opaque = true;
498		break;
499	}
500
501	return 0;
502}
503
504static int tegra_plane_check_transparency(struct tegra_plane *tegra,
505					  struct tegra_plane_state *state)
506{
507	struct drm_plane_state *old, *plane_state;
508	struct drm_plane *plane;
509
510	old = drm_atomic_get_old_plane_state(state->base.state, &tegra->base);
511
512	/* check if zpos / transparency changed */
513	if (old->normalized_zpos == state->base.normalized_zpos &&
514	    to_tegra_plane_state(old)->opaque == state->opaque)
515		return 0;
516
517	/* include all sibling planes into this commit */
518	drm_for_each_plane(plane, tegra->base.dev) {
519		struct tegra_plane *p = to_tegra_plane(plane);
520
521		/* skip this plane and planes on different CRTCs */
522		if (p == tegra || p->dc != tegra->dc)
523			continue;
524
525		plane_state = drm_atomic_get_plane_state(state->base.state,
526							 plane);
527		if (IS_ERR(plane_state))
528			return PTR_ERR(plane_state);
529	}
530
531	return 1;
532}
533
534static unsigned int tegra_plane_get_overlap_index(struct tegra_plane *plane,
535						  struct tegra_plane *other)
536{
537	unsigned int index = 0, i;
538
539	WARN_ON(plane == other);
540
541	for (i = 0; i < 3; i++) {
542		if (i == plane->index)
543			continue;
544
545		if (i == other->index)
546			break;
547
548		index++;
549	}
550
551	return index;
552}
553
554static void tegra_plane_update_transparency(struct tegra_plane *tegra,
555					    struct tegra_plane_state *state)
556{
557	struct drm_plane_state *new;
558	struct drm_plane *plane;
559	unsigned int i;
560
561	for_each_new_plane_in_state(state->base.state, plane, new, i) {
562		struct tegra_plane *p = to_tegra_plane(plane);
563		unsigned index;
564
565		/* skip this plane and planes on different CRTCs */
566		if (p == tegra || p->dc != tegra->dc)
567			continue;
568
569		index = tegra_plane_get_overlap_index(tegra, p);
570
571		if (new->fb && __drm_format_has_alpha(new->fb->format->format))
572			state->blending[index].alpha = true;
573		else
574			state->blending[index].alpha = false;
575
576		if (new->normalized_zpos > state->base.normalized_zpos)
577			state->blending[index].top = true;
578		else
579			state->blending[index].top = false;
580
581		/*
582		 * Missing framebuffer means that plane is disabled, in this
583		 * case mark B / C window as top to be able to differentiate
584		 * windows indices order in regards to zPos for the middle
585		 * window X / Y registers programming.
586		 */
587		if (!new->fb)
588			state->blending[index].top = (index == 1);
589	}
590}
591
592static int tegra_plane_setup_transparency(struct tegra_plane *tegra,
593					  struct tegra_plane_state *state)
594{
595	struct tegra_plane_state *tegra_state;
596	struct drm_plane_state *new;
597	struct drm_plane *plane;
598	int err;
599
600	/*
601	 * If planes zpos / transparency changed, sibling planes blending
602	 * state may require adjustment and in this case they will be included
603	 * into this atom commit, otherwise blending state is unchanged.
604	 */
605	err = tegra_plane_check_transparency(tegra, state);
606	if (err <= 0)
607		return err;
608
609	/*
610	 * All planes are now in the atomic state, walk them up and update
611	 * transparency state for each plane.
612	 */
613	drm_for_each_plane(plane, tegra->base.dev) {
614		struct tegra_plane *p = to_tegra_plane(plane);
615
616		/* skip planes on different CRTCs */
617		if (p->dc != tegra->dc)
618			continue;
619
620		new = drm_atomic_get_new_plane_state(state->base.state, plane);
621		tegra_state = to_tegra_plane_state(new);
622
623		/*
624		 * There is no need to update blending state for the disabled
625		 * plane.
626		 */
627		if (new->fb)
628			tegra_plane_update_transparency(p, tegra_state);
629	}
630
631	return 0;
632}
633
634int tegra_plane_setup_legacy_state(struct tegra_plane *tegra,
635				   struct tegra_plane_state *state)
636{
637	int err;
638
639	err = tegra_plane_setup_opacity(tegra, state);
640	if (err < 0)
641		return err;
642
643	err = tegra_plane_setup_transparency(tegra, state);
644	if (err < 0)
645		return err;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
646
647	return 0;
648}