Linux Audio

Check our new training course

Loading...
v6.13.7
  1/*
  2 * Copyright (C) 2008 Maarten Maathuis.
  3 * All Rights Reserved.
  4 *
  5 * Permission is hereby granted, free of charge, to any person obtaining
  6 * a copy of this software and associated documentation files (the
  7 * "Software"), to deal in the Software without restriction, including
  8 * without limitation the rights to use, copy, modify, merge, publish,
  9 * distribute, sublicense, and/or sell copies of the Software, and to
 10 * permit persons to whom the Software is furnished to do so, subject to
 11 * the following conditions:
 12 *
 13 * The above copyright notice and this permission notice (including the
 14 * next paragraph) shall be included in all copies or substantial
 15 * portions of the Software.
 16 *
 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 24 *
 25 */
 26
 27#include <acpi/video.h>
 28
 29#include <drm/drm_atomic.h>
 30#include <drm/drm_atomic_helper.h>
 31#include <drm/drm_client_event.h>
 32#include <drm/drm_crtc_helper.h>
 
 33#include <drm/drm_fourcc.h>
 34#include <drm/drm_gem_framebuffer_helper.h>
 35#include <drm/drm_probe_helper.h>
 36#include <drm/drm_vblank.h>
 37
 
 38#include "nouveau_crtc.h"
 39#include "nouveau_gem.h"
 40#include "nouveau_connector.h"
 41#include "nv50_display.h"
 42
 43#include <nvif/class.h>
 44#include <nvif/if0011.h>
 45#include <nvif/if0013.h>
 46#include <dispnv50/crc.h>
 47#include <dispnv50/tile.h>
 48
 49int
 50nouveau_display_vblank_enable(struct drm_crtc *crtc)
 51{
 52	struct nouveau_crtc *nv_crtc;
 53
 54	nv_crtc = nouveau_crtc(crtc);
 55	nvif_event_allow(&nv_crtc->vblank);
 56
 57	return 0;
 58}
 59
 60void
 61nouveau_display_vblank_disable(struct drm_crtc *crtc)
 62{
 63	struct nouveau_crtc *nv_crtc;
 64
 65	nv_crtc = nouveau_crtc(crtc);
 66	nvif_event_block(&nv_crtc->vblank);
 67}
 68
 69static inline int
 70calc(int blanks, int blanke, int total, int line)
 71{
 72	if (blanke >= blanks) {
 73		if (line >= blanks)
 74			line -= total;
 75	} else {
 76		if (line >= blanks)
 77			line -= total;
 78		line -= blanke + 1;
 79	}
 80	return line;
 81}
 82
 83static bool
 84nouveau_display_scanoutpos_head(struct drm_crtc *crtc, int *vpos, int *hpos,
 85				ktime_t *stime, ktime_t *etime)
 86{
 87	struct drm_vblank_crtc *vblank = drm_crtc_vblank_crtc(crtc);
 88	struct nvif_head *head = &nouveau_crtc(crtc)->head;
 89	struct nvif_head_scanoutpos_v0 args;
 
 
 
 
 
 
 90	int retry = 20;
 91	bool ret = false;
 92
 93	args.version = 0;
 94
 95	do {
 96		ret = nvif_mthd(&head->object, NVIF_HEAD_V0_SCANOUTPOS, &args, sizeof(args));
 97		if (ret != 0)
 98			return false;
 99
100		if (args.vline) {
101			ret = true;
102			break;
103		}
104
105		if (retry) ndelay(vblank->linedur_ns);
106	} while (retry--);
107
108	*hpos = args.hline;
109	*vpos = calc(args.vblanks, args.vblanke, args.vtotal, args.vline);
110	if (stime) *stime = ns_to_ktime(args.time[0]);
111	if (etime) *etime = ns_to_ktime(args.time[1]);
 
112
113	return ret;
114}
115
116bool
117nouveau_display_scanoutpos(struct drm_crtc *crtc,
118			   bool in_vblank_irq, int *vpos, int *hpos,
119			   ktime_t *stime, ktime_t *etime,
120			   const struct drm_display_mode *mode)
121{
122	return nouveau_display_scanoutpos_head(crtc, vpos, hpos,
123					       stime, etime);
124}
125
126static const struct drm_framebuffer_funcs nouveau_framebuffer_funcs = {
127	.destroy = drm_gem_fb_destroy,
128	.create_handle = drm_gem_fb_create_handle,
129};
130
131static void
132nouveau_decode_mod(struct nouveau_drm *drm,
133		   uint64_t modifier,
134		   uint32_t *tile_mode,
135		   uint8_t *kind)
136{
137	struct nouveau_display *disp = nouveau_display(drm->dev);
138	BUG_ON(!tile_mode || !kind);
139
140	if (modifier == DRM_FORMAT_MOD_LINEAR) {
141		/* tile_mode will not be used in this case */
142		*tile_mode = 0;
143		*kind = 0;
144	} else {
145		/*
146		 * Extract the block height and kind from the corresponding
147		 * modifier fields.  See drm_fourcc.h for details.
148		 */
149
150		if ((modifier & (0xffull << 12)) == 0ull) {
151			/* Legacy modifier.  Translate to this dev's 'kind.' */
152			modifier |= disp->format_modifiers[0] & (0xffull << 12);
153		}
154
155		*tile_mode = (uint32_t)(modifier & 0xF);
156		*kind = (uint8_t)((modifier >> 12) & 0xFF);
157
158		if (drm->client.device.info.chipset >= 0xc0)
159			*tile_mode <<= 4;
160	}
161}
162
163void
164nouveau_framebuffer_get_layout(struct drm_framebuffer *fb,
165			       uint32_t *tile_mode,
166			       uint8_t *kind)
167{
168	if (fb->flags & DRM_MODE_FB_MODIFIERS) {
169		struct nouveau_drm *drm = nouveau_drm(fb->dev);
170
171		nouveau_decode_mod(drm, fb->modifier, tile_mode, kind);
172	} else {
173		const struct nouveau_bo *nvbo = nouveau_gem_object(fb->obj[0]);
174
175		*tile_mode = nvbo->mode;
176		*kind = nvbo->kind;
177	}
178}
179
180static const u64 legacy_modifiers[] = {
181	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0),
182	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1),
183	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2),
184	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3),
185	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4),
186	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5),
187	DRM_FORMAT_MOD_INVALID
188};
189
190static int
191nouveau_validate_decode_mod(struct nouveau_drm *drm,
192			    uint64_t modifier,
193			    uint32_t *tile_mode,
194			    uint8_t *kind)
195{
196	struct nouveau_display *disp = nouveau_display(drm->dev);
197	int mod;
198
199	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
200		return -EINVAL;
201	}
202
203	BUG_ON(!disp->format_modifiers);
204
205	for (mod = 0;
206	     (disp->format_modifiers[mod] != DRM_FORMAT_MOD_INVALID) &&
207	     (disp->format_modifiers[mod] != modifier);
208	     mod++);
209
210	if (disp->format_modifiers[mod] == DRM_FORMAT_MOD_INVALID) {
211		for (mod = 0;
212		     (legacy_modifiers[mod] != DRM_FORMAT_MOD_INVALID) &&
213		     (legacy_modifiers[mod] != modifier);
214		     mod++);
215		if (legacy_modifiers[mod] == DRM_FORMAT_MOD_INVALID)
216			return -EINVAL;
217	}
218
219	nouveau_decode_mod(drm, modifier, tile_mode, kind);
220
221	return 0;
222}
223
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
224static int
225nouveau_check_bl_size(struct nouveau_drm *drm, struct nouveau_bo *nvbo,
226		      uint32_t offset, uint32_t stride, uint32_t h,
227		      uint32_t tile_mode)
228{
229	uint32_t gob_size, bw, bh, gobs_in_block;
230	uint64_t bl_size;
231
232	BUG_ON(drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA);
233
234	if (nouveau_check_tile_mode(tile_mode, drm->client.device.info.chipset))
 
 
 
 
 
 
235		return -EINVAL;
236
237	gobs_in_block = nouveau_get_gobs_in_block(tile_mode, drm->client.device.info.chipset);
 
 
 
 
238	bw = nouveau_get_width_in_blocks(stride);
239	bh = nouveau_get_height_in_blocks(h, gobs_in_block, drm->client.device.info.family);
240	gob_size = nouveau_get_gob_size(drm->client.device.info.family);
241
242	bl_size = bw * bh * gobs_in_block * gob_size;
243
244	DRM_DEBUG_KMS("offset=%u stride=%u h=%u gobs_in_block=%u bw=%u bh=%u gob_size=%u bl_size=%llu size=%zu\n",
245		      offset, stride, h, gobs_in_block, bw, bh, gob_size,
246		      bl_size, nvbo->bo.base.size);
247
248	if (bl_size + offset > nvbo->bo.base.size)
249		return -ERANGE;
250
251	return 0;
252}
253
254int
255nouveau_framebuffer_new(struct drm_device *dev,
256			const struct drm_mode_fb_cmd2 *mode_cmd,
257			struct drm_gem_object *gem,
258			struct drm_framebuffer **pfb)
259{
260	struct nouveau_drm *drm = nouveau_drm(dev);
261	struct nouveau_bo *nvbo = nouveau_gem_object(gem);
262	struct drm_framebuffer *fb;
263	const struct drm_format_info *info;
264	unsigned int height, i;
265	uint32_t tile_mode;
266	uint8_t kind;
267	int ret;
268
269        /* YUV overlays have special requirements pre-NV50 */
270	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA &&
271
272	    (mode_cmd->pixel_format == DRM_FORMAT_YUYV ||
273	     mode_cmd->pixel_format == DRM_FORMAT_UYVY ||
274	     mode_cmd->pixel_format == DRM_FORMAT_NV12 ||
275	     mode_cmd->pixel_format == DRM_FORMAT_NV21) &&
276	    (mode_cmd->pitches[0] & 0x3f || /* align 64 */
277	     mode_cmd->pitches[0] >= 0x10000 || /* at most 64k pitch */
278	     (mode_cmd->pitches[1] && /* pitches for planes must match */
279	      mode_cmd->pitches[0] != mode_cmd->pitches[1]))) {
280		DRM_DEBUG_KMS("Unsuitable framebuffer: format: %p4cc; pitches: 0x%x\n 0x%x\n",
281			      &mode_cmd->pixel_format,
282			      mode_cmd->pitches[0], mode_cmd->pitches[1]);
 
 
 
283		return -EINVAL;
284	}
285
286	if (mode_cmd->flags & DRM_MODE_FB_MODIFIERS) {
287		if (nouveau_validate_decode_mod(drm, mode_cmd->modifier[0],
288						&tile_mode, &kind)) {
289			DRM_DEBUG_KMS("Unsupported modifier: 0x%llx\n",
290				      mode_cmd->modifier[0]);
291			return -EINVAL;
292		}
293	} else {
294		tile_mode = nvbo->mode;
295		kind = nvbo->kind;
296	}
297
298	info = drm_get_format_info(dev, mode_cmd);
299
300	for (i = 0; i < info->num_planes; i++) {
 
 
 
301		height = drm_format_info_plane_height(info,
302						      mode_cmd->height,
303						      i);
304
305		if (kind) {
306			ret = nouveau_check_bl_size(drm, nvbo,
307						    mode_cmd->offsets[i],
308						    mode_cmd->pitches[i],
309						    height, tile_mode);
310			if (ret)
311				return ret;
312		} else {
313			uint32_t size = mode_cmd->pitches[i] * height;
314
315			if (size + mode_cmd->offsets[i] > nvbo->bo.base.size)
316				return -ERANGE;
317		}
318	}
319
320	if (!(fb = *pfb = kzalloc(sizeof(*fb), GFP_KERNEL)))
321		return -ENOMEM;
322
323	drm_helper_mode_fill_fb_struct(dev, fb, mode_cmd);
324	fb->obj[0] = gem;
325
326	ret = drm_framebuffer_init(dev, fb, &nouveau_framebuffer_funcs);
327	if (ret)
328		kfree(fb);
329	return ret;
330}
331
332struct drm_framebuffer *
333nouveau_user_framebuffer_create(struct drm_device *dev,
334				struct drm_file *file_priv,
335				const struct drm_mode_fb_cmd2 *mode_cmd)
336{
337	struct drm_framebuffer *fb;
338	struct drm_gem_object *gem;
339	int ret;
340
341	gem = drm_gem_object_lookup(file_priv, mode_cmd->handles[0]);
342	if (!gem)
343		return ERR_PTR(-ENOENT);
344
345	ret = nouveau_framebuffer_new(dev, mode_cmd, gem, &fb);
346	if (ret == 0)
347		return fb;
348
349	drm_gem_object_put(gem);
350	return ERR_PTR(ret);
351}
352
353static const struct drm_mode_config_funcs nouveau_mode_config_funcs = {
354	.fb_create = nouveau_user_framebuffer_create,
 
355};
356
357
358struct nouveau_drm_prop_enum_list {
359	u8 gen_mask;
360	int type;
361	char *name;
362};
363
364static struct nouveau_drm_prop_enum_list underscan[] = {
365	{ 6, UNDERSCAN_AUTO, "auto" },
366	{ 6, UNDERSCAN_OFF, "off" },
367	{ 6, UNDERSCAN_ON, "on" },
368	{}
369};
370
371static struct nouveau_drm_prop_enum_list dither_mode[] = {
372	{ 7, DITHERING_MODE_AUTO, "auto" },
373	{ 7, DITHERING_MODE_OFF, "off" },
374	{ 1, DITHERING_MODE_ON, "on" },
375	{ 6, DITHERING_MODE_STATIC2X2, "static 2x2" },
376	{ 6, DITHERING_MODE_DYNAMIC2X2, "dynamic 2x2" },
377	{ 4, DITHERING_MODE_TEMPORAL, "temporal" },
378	{}
379};
380
381static struct nouveau_drm_prop_enum_list dither_depth[] = {
382	{ 6, DITHERING_DEPTH_AUTO, "auto" },
383	{ 6, DITHERING_DEPTH_6BPC, "6 bpc" },
384	{ 6, DITHERING_DEPTH_8BPC, "8 bpc" },
385	{}
386};
387
388#define PROP_ENUM(p,gen,n,list) do {                                           \
389	struct nouveau_drm_prop_enum_list *l = (list);                         \
390	int c = 0;                                                             \
391	while (l->gen_mask) {                                                  \
392		if (l->gen_mask & (1 << (gen)))                                \
393			c++;                                                   \
394		l++;                                                           \
395	}                                                                      \
396	if (c) {                                                               \
397		p = drm_property_create(dev, DRM_MODE_PROP_ENUM, n, c);        \
398		l = (list);                                                    \
399		while (p && l->gen_mask) {                                     \
400			if (l->gen_mask & (1 << (gen))) {                      \
401				drm_property_add_enum(p, l->type, l->name);    \
402			}                                                      \
403			l++;                                                   \
404		}                                                              \
405	}                                                                      \
406} while(0)
407
408void
409nouveau_display_hpd_resume(struct nouveau_drm *drm)
410{
411	if (drm->headless)
412		return;
413
414	spin_lock_irq(&drm->hpd_lock);
415	drm->hpd_pending = ~0;
416	spin_unlock_irq(&drm->hpd_lock);
417
418	schedule_work(&drm->hpd_work);
419}
420
421static void
422nouveau_display_hpd_work(struct work_struct *work)
423{
424	struct nouveau_drm *drm = container_of(work, typeof(*drm), hpd_work);
425	struct drm_device *dev = drm->dev;
426	struct drm_connector *connector;
427	struct drm_connector_list_iter conn_iter;
428	u32 pending;
429	int changed = 0;
430	struct drm_connector *first_changed_connector = NULL;
431
432	pm_runtime_get_sync(dev->dev);
433
434	spin_lock_irq(&drm->hpd_lock);
435	pending = drm->hpd_pending;
436	drm->hpd_pending = 0;
437	spin_unlock_irq(&drm->hpd_lock);
438
439	/* Nothing to do, exit early without updating the last busy counter */
440	if (!pending)
441		goto noop;
442
443	mutex_lock(&dev->mode_config.mutex);
444	drm_connector_list_iter_begin(dev, &conn_iter);
445
446	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
447		struct nouveau_connector *nv_connector = nouveau_connector(connector);
448		enum drm_connector_status old_status = connector->status;
449		u64 bits, old_epoch_counter = connector->epoch_counter;
450
451		if (!(pending & drm_connector_mask(connector)))
452			continue;
453
454		spin_lock_irq(&drm->hpd_lock);
455		bits = nv_connector->hpd_pending;
456		nv_connector->hpd_pending = 0;
457		spin_unlock_irq(&drm->hpd_lock);
458
459		drm_dbg_kms(dev, "[CONNECTOR:%d:%s] plug:%d unplug:%d irq:%d\n",
460			    connector->base.id, connector->name,
461			    !!(bits & NVIF_CONN_EVENT_V0_PLUG),
462			    !!(bits & NVIF_CONN_EVENT_V0_UNPLUG),
463			    !!(bits & NVIF_CONN_EVENT_V0_IRQ));
464
465		if (bits & NVIF_CONN_EVENT_V0_IRQ) {
466			if (nouveau_dp_link_check(nv_connector))
467				continue;
468		}
469
470		connector->status = drm_helper_probe_detect(connector, NULL, false);
471		if (old_epoch_counter == connector->epoch_counter)
472			continue;
473
474		changed++;
475		if (!first_changed_connector) {
476			drm_connector_get(connector);
477			first_changed_connector = connector;
478		}
479
480		drm_dbg_kms(dev, "[CONNECTOR:%d:%s] status updated from %s to %s (epoch counter %llu->%llu)\n",
481			    connector->base.id, connector->name,
482			    drm_get_connector_status_name(old_status),
483			    drm_get_connector_status_name(connector->status),
484			    old_epoch_counter, connector->epoch_counter);
485	}
486
487	drm_connector_list_iter_end(&conn_iter);
488	mutex_unlock(&dev->mode_config.mutex);
489
490	if (changed == 1)
491		drm_kms_helper_connector_hotplug_event(first_changed_connector);
492	else if (changed > 0)
493		drm_kms_helper_hotplug_event(dev);
494
495	if (first_changed_connector)
496		drm_connector_put(first_changed_connector);
497
498	pm_runtime_mark_last_busy(drm->dev->dev);
499noop:
500	pm_runtime_put_autosuspend(dev->dev);
501}
502
503#ifdef CONFIG_ACPI
504
505static int
506nouveau_display_acpi_ntfy(struct notifier_block *nb, unsigned long val,
507			  void *data)
508{
509	struct nouveau_drm *drm = container_of(nb, typeof(*drm), acpi_nb);
510	struct acpi_bus_event *info = data;
511	int ret;
512
513	if (!strcmp(info->device_class, ACPI_VIDEO_CLASS)) {
514		if (info->type == ACPI_VIDEO_NOTIFY_PROBE) {
515			ret = pm_runtime_get(drm->dev->dev);
516			if (ret == 1 || ret == -EACCES) {
517				/* If the GPU is already awake, or in a state
518				 * where we can't wake it up, it can handle
519				 * it's own hotplug events.
520				 */
521				pm_runtime_put_autosuspend(drm->dev->dev);
522			} else if (ret == 0 || ret == -EINPROGRESS) {
523				/* We've started resuming the GPU already, so
524				 * it will handle scheduling a full reprobe
525				 * itself
526				 */
527				NV_DEBUG(drm, "ACPI requested connector reprobe\n");
 
528				pm_runtime_put_noidle(drm->dev->dev);
529			} else {
530				NV_WARN(drm, "Dropped ACPI reprobe event due to RPM error: %d\n",
531					ret);
532			}
533
534			/* acpi-video should not generate keypresses for this */
535			return NOTIFY_BAD;
536		}
537	}
538
539	return NOTIFY_DONE;
540}
541#endif
542
543int
544nouveau_display_init(struct drm_device *dev, bool resume, bool runtime)
545{
546	struct nouveau_display *disp = nouveau_display(dev);
547	struct drm_connector *connector;
548	struct drm_connector_list_iter conn_iter;
549	int ret;
550
551	/*
552	 * Enable hotplug interrupts (done as early as possible, since we need
553	 * them for MST)
554	 */
555	drm_connector_list_iter_begin(dev, &conn_iter);
556	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
557		struct nouveau_connector *conn = nouveau_connector(connector);
558		nvif_event_allow(&conn->hpd);
559		nvif_event_allow(&conn->irq);
560	}
561	drm_connector_list_iter_end(&conn_iter);
562
563	ret = disp->init(dev, resume, runtime);
564	if (ret)
565		return ret;
566
567	/* enable connector detection and polling for connectors without HPD
568	 * support
569	 */
570	drm_kms_helper_poll_enable(dev);
571
572	return ret;
573}
574
575void
576nouveau_display_fini(struct drm_device *dev, bool suspend, bool runtime)
577{
578	struct nouveau_display *disp = nouveau_display(dev);
579	struct nouveau_drm *drm = nouveau_drm(dev);
580	struct drm_connector *connector;
581	struct drm_connector_list_iter conn_iter;
582
583	if (!suspend) {
584		if (drm_drv_uses_atomic_modeset(dev))
585			drm_atomic_helper_shutdown(dev);
586		else
587			drm_helper_force_disable_all(dev);
588	}
589
590	/* disable hotplug interrupts */
591	drm_connector_list_iter_begin(dev, &conn_iter);
592	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
593		struct nouveau_connector *conn = nouveau_connector(connector);
594		nvif_event_block(&conn->irq);
595		nvif_event_block(&conn->hpd);
596	}
597	drm_connector_list_iter_end(&conn_iter);
598
599	if (!runtime && !drm->headless)
600		cancel_work_sync(&drm->hpd_work);
601
602	drm_kms_helper_poll_disable(dev);
603	disp->fini(dev, runtime, suspend);
604}
605
606static void
607nouveau_display_create_properties(struct drm_device *dev)
608{
609	struct nouveau_display *disp = nouveau_display(dev);
610	int gen;
611
612	if (disp->disp.object.oclass < NV50_DISP)
613		gen = 0;
614	else
615	if (disp->disp.object.oclass < GF110_DISP)
616		gen = 1;
617	else
618		gen = 2;
619
620	PROP_ENUM(disp->dithering_mode, gen, "dithering mode", dither_mode);
621	PROP_ENUM(disp->dithering_depth, gen, "dithering depth", dither_depth);
622	PROP_ENUM(disp->underscan_property, gen, "underscan", underscan);
623
624	disp->underscan_hborder_property =
625		drm_property_create_range(dev, 0, "underscan hborder", 0, 128);
626
627	disp->underscan_vborder_property =
628		drm_property_create_range(dev, 0, "underscan vborder", 0, 128);
629
630	if (gen < 1)
631		return;
632
633	/* -90..+90 */
634	disp->vibrant_hue_property =
635		drm_property_create_range(dev, 0, "vibrant hue", 0, 180);
636
637	/* -100..+100 */
638	disp->color_vibrance_property =
639		drm_property_create_range(dev, 0, "color vibrance", 0, 200);
640}
641
642int
643nouveau_display_create(struct drm_device *dev)
644{
645	struct nouveau_drm *drm = nouveau_drm(dev);
 
646	struct nouveau_display *disp;
647	int ret;
648
649	disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL);
650	if (!disp)
651		return -ENOMEM;
652
653	drm_mode_config_init(dev);
654	drm_mode_create_scaling_mode_property(dev);
655	drm_mode_create_dvi_i_properties(dev);
656
657	dev->mode_config.funcs = &nouveau_mode_config_funcs;
 
658
659	dev->mode_config.min_width = 0;
660	dev->mode_config.min_height = 0;
661	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_CELSIUS) {
662		dev->mode_config.max_width = 2048;
663		dev->mode_config.max_height = 2048;
664	} else
665	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
666		dev->mode_config.max_width = 4096;
667		dev->mode_config.max_height = 4096;
668	} else
669	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI) {
670		dev->mode_config.max_width = 8192;
671		dev->mode_config.max_height = 8192;
672	} else {
673		dev->mode_config.max_width = 16384;
674		dev->mode_config.max_height = 16384;
675	}
676
677	dev->mode_config.preferred_depth = 24;
678	dev->mode_config.prefer_shadow = 1;
 
679
680	if (drm->client.device.info.chipset < 0x11)
681		dev->mode_config.async_page_flip = false;
682	else
683		dev->mode_config.async_page_flip = true;
684
685	drm_kms_helper_poll_init(dev);
686	drm_kms_helper_poll_disable(dev);
687
688	if (nouveau_modeset != 2) {
689		ret = nvif_disp_ctor(&drm->client.device, "kmsDisp", 0, &disp->disp);
690		/* no display hw */
691		if (ret == -ENODEV) {
692			ret = 0;
693			drm->headless = true;
694			goto disp_create_err;
695		}
696
697		if (!ret && (disp->disp.outp_mask || drm->vbios.dcb.entries)) {
698			nouveau_display_create_properties(dev);
699			if (disp->disp.object.oclass < NV50_DISP) {
700				dev->mode_config.fb_modifiers_not_supported = true;
701				ret = nv04_display_create(dev);
702			} else {
703				ret = nv50_display_create(dev);
704			}
705		}
706	} else {
707		ret = 0;
708	}
709
710	if (ret)
711		goto disp_create_err;
712
713	drm_mode_config_reset(dev);
714
715	if (dev->mode_config.num_crtc) {
716		ret = drm_vblank_init(dev, dev->mode_config.num_crtc);
717		if (ret)
718			goto vblank_err;
719
720		if (disp->disp.object.oclass >= NV50_DISP)
721			nv50_crc_init(dev);
722	}
723
724	INIT_WORK(&drm->hpd_work, nouveau_display_hpd_work);
725	spin_lock_init(&drm->hpd_lock);
726#ifdef CONFIG_ACPI
727	drm->acpi_nb.notifier_call = nouveau_display_acpi_ntfy;
728	register_acpi_notifier(&drm->acpi_nb);
729#endif
730
731	return 0;
732
733vblank_err:
734	disp->dtor(dev);
735disp_create_err:
736	drm_kms_helper_poll_fini(dev);
737	drm_mode_config_cleanup(dev);
738	return ret;
739}
740
741void
742nouveau_display_destroy(struct drm_device *dev)
743{
744	struct nouveau_display *disp = nouveau_display(dev);
745	struct nouveau_drm *drm = nouveau_drm(dev);
746
747#ifdef CONFIG_ACPI
748	unregister_acpi_notifier(&drm->acpi_nb);
749#endif
750
751	drm_kms_helper_poll_fini(dev);
752	drm_mode_config_cleanup(dev);
753
754	if (disp->dtor)
755		disp->dtor(dev);
756
757	nvif_disp_dtor(&disp->disp);
758
759	drm->display = NULL;
760	kfree(disp);
761}
762
763int
764nouveau_display_suspend(struct drm_device *dev, bool runtime)
765{
766	struct nouveau_display *disp = nouveau_display(dev);
767
768	drm_client_dev_suspend(dev, false);
769
770	if (drm_drv_uses_atomic_modeset(dev)) {
771		if (!runtime) {
772			disp->suspend = drm_atomic_helper_suspend(dev);
773			if (IS_ERR(disp->suspend)) {
774				int ret = PTR_ERR(disp->suspend);
775				disp->suspend = NULL;
776				return ret;
777			}
778		}
779	}
780
781	nouveau_display_fini(dev, true, runtime);
782	return 0;
783}
784
785void
786nouveau_display_resume(struct drm_device *dev, bool runtime)
787{
788	struct nouveau_display *disp = nouveau_display(dev);
789
790	nouveau_display_init(dev, true, runtime);
791
792	if (drm_drv_uses_atomic_modeset(dev)) {
793		if (disp->suspend) {
794			drm_atomic_helper_resume(dev, disp->suspend);
795			disp->suspend = NULL;
796		}
 
797	}
798
799	drm_client_dev_resume(dev, false);
800}
801
802int
803nouveau_display_dumb_create(struct drm_file *file_priv, struct drm_device *dev,
804			    struct drm_mode_create_dumb *args)
805{
806	struct nouveau_cli *cli = nouveau_cli(file_priv);
807	struct nouveau_bo *bo;
808	uint32_t domain;
809	int ret;
810
811	args->pitch = roundup(args->width * (args->bpp / 8), 256);
812	args->size = args->pitch * args->height;
813	args->size = roundup(args->size, PAGE_SIZE);
814
815	/* Use VRAM if there is any ; otherwise fallback to system memory */
816	if (nouveau_drm(dev)->client.device.info.ram_size != 0)
817		domain = NOUVEAU_GEM_DOMAIN_VRAM;
818	else
819		domain = NOUVEAU_GEM_DOMAIN_GART;
820
821	ret = nouveau_gem_new(cli, args->size, 0, domain, 0, 0, &bo);
822	if (ret)
823		return ret;
824
825	ret = drm_gem_handle_create(file_priv, &bo->bo.base, &args->handle);
826	drm_gem_object_put(&bo->bo.base);
827	return ret;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
828}
v5.9
  1/*
  2 * Copyright (C) 2008 Maarten Maathuis.
  3 * All Rights Reserved.
  4 *
  5 * Permission is hereby granted, free of charge, to any person obtaining
  6 * a copy of this software and associated documentation files (the
  7 * "Software"), to deal in the Software without restriction, including
  8 * without limitation the rights to use, copy, modify, merge, publish,
  9 * distribute, sublicense, and/or sell copies of the Software, and to
 10 * permit persons to whom the Software is furnished to do so, subject to
 11 * the following conditions:
 12 *
 13 * The above copyright notice and this permission notice (including the
 14 * next paragraph) shall be included in all copies or substantial
 15 * portions of the Software.
 16 *
 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 24 *
 25 */
 26
 27#include <acpi/video.h>
 28
 29#include <drm/drm_atomic.h>
 30#include <drm/drm_atomic_helper.h>
 
 31#include <drm/drm_crtc_helper.h>
 32#include <drm/drm_fb_helper.h>
 33#include <drm/drm_fourcc.h>
 34#include <drm/drm_gem_framebuffer_helper.h>
 35#include <drm/drm_probe_helper.h>
 36#include <drm/drm_vblank.h>
 37
 38#include "nouveau_fbcon.h"
 39#include "nouveau_crtc.h"
 40#include "nouveau_gem.h"
 41#include "nouveau_connector.h"
 42#include "nv50_display.h"
 43
 44#include <nvif/class.h>
 45#include <nvif/cl0046.h>
 46#include <nvif/event.h>
 47#include <dispnv50/crc.h>
 
 48
 49int
 50nouveau_display_vblank_enable(struct drm_crtc *crtc)
 51{
 52	struct nouveau_crtc *nv_crtc;
 53
 54	nv_crtc = nouveau_crtc(crtc);
 55	nvif_notify_get(&nv_crtc->vblank);
 56
 57	return 0;
 58}
 59
 60void
 61nouveau_display_vblank_disable(struct drm_crtc *crtc)
 62{
 63	struct nouveau_crtc *nv_crtc;
 64
 65	nv_crtc = nouveau_crtc(crtc);
 66	nvif_notify_put(&nv_crtc->vblank);
 67}
 68
 69static inline int
 70calc(int blanks, int blanke, int total, int line)
 71{
 72	if (blanke >= blanks) {
 73		if (line >= blanks)
 74			line -= total;
 75	} else {
 76		if (line >= blanks)
 77			line -= total;
 78		line -= blanke + 1;
 79	}
 80	return line;
 81}
 82
 83static bool
 84nouveau_display_scanoutpos_head(struct drm_crtc *crtc, int *vpos, int *hpos,
 85				ktime_t *stime, ktime_t *etime)
 86{
 87	struct {
 88		struct nv04_disp_mthd_v0 base;
 89		struct nv04_disp_scanoutpos_v0 scan;
 90	} args = {
 91		.base.method = NV04_DISP_SCANOUTPOS,
 92		.base.head = nouveau_crtc(crtc)->index,
 93	};
 94	struct nouveau_display *disp = nouveau_display(crtc->dev);
 95	struct drm_vblank_crtc *vblank = &crtc->dev->vblank[drm_crtc_index(crtc)];
 96	int retry = 20;
 97	bool ret = false;
 98
 
 
 99	do {
100		ret = nvif_mthd(&disp->disp.object, 0, &args, sizeof(args));
101		if (ret != 0)
102			return false;
103
104		if (args.scan.vline) {
105			ret = true;
106			break;
107		}
108
109		if (retry) ndelay(vblank->linedur_ns);
110	} while (retry--);
111
112	*hpos = args.scan.hline;
113	*vpos = calc(args.scan.vblanks, args.scan.vblanke,
114		     args.scan.vtotal, args.scan.vline);
115	if (stime) *stime = ns_to_ktime(args.scan.time[0]);
116	if (etime) *etime = ns_to_ktime(args.scan.time[1]);
117
118	return ret;
119}
120
121bool
122nouveau_display_scanoutpos(struct drm_crtc *crtc,
123			   bool in_vblank_irq, int *vpos, int *hpos,
124			   ktime_t *stime, ktime_t *etime,
125			   const struct drm_display_mode *mode)
126{
127	return nouveau_display_scanoutpos_head(crtc, vpos, hpos,
128					       stime, etime);
129}
130
131static const struct drm_framebuffer_funcs nouveau_framebuffer_funcs = {
132	.destroy = drm_gem_fb_destroy,
133	.create_handle = drm_gem_fb_create_handle,
134};
135
136static void
137nouveau_decode_mod(struct nouveau_drm *drm,
138		   uint64_t modifier,
139		   uint32_t *tile_mode,
140		   uint8_t *kind)
141{
142	struct nouveau_display *disp = nouveau_display(drm->dev);
143	BUG_ON(!tile_mode || !kind);
144
145	if (modifier == DRM_FORMAT_MOD_LINEAR) {
146		/* tile_mode will not be used in this case */
147		*tile_mode = 0;
148		*kind = 0;
149	} else {
150		/*
151		 * Extract the block height and kind from the corresponding
152		 * modifier fields.  See drm_fourcc.h for details.
153		 */
154
155		if ((modifier & (0xffull << 12)) == 0ull) {
156			/* Legacy modifier.  Translate to this dev's 'kind.' */
157			modifier |= disp->format_modifiers[0] & (0xffull << 12);
158		}
159
160		*tile_mode = (uint32_t)(modifier & 0xF);
161		*kind = (uint8_t)((modifier >> 12) & 0xFF);
162
163		if (drm->client.device.info.chipset >= 0xc0)
164			*tile_mode <<= 4;
165	}
166}
167
168void
169nouveau_framebuffer_get_layout(struct drm_framebuffer *fb,
170			       uint32_t *tile_mode,
171			       uint8_t *kind)
172{
173	if (fb->flags & DRM_MODE_FB_MODIFIERS) {
174		struct nouveau_drm *drm = nouveau_drm(fb->dev);
175
176		nouveau_decode_mod(drm, fb->modifier, tile_mode, kind);
177	} else {
178		const struct nouveau_bo *nvbo = nouveau_gem_object(fb->obj[0]);
179
180		*tile_mode = nvbo->mode;
181		*kind = nvbo->kind;
182	}
183}
184
185static const u64 legacy_modifiers[] = {
186	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0),
187	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1),
188	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2),
189	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3),
190	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4),
191	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5),
192	DRM_FORMAT_MOD_INVALID
193};
194
195static int
196nouveau_validate_decode_mod(struct nouveau_drm *drm,
197			    uint64_t modifier,
198			    uint32_t *tile_mode,
199			    uint8_t *kind)
200{
201	struct nouveau_display *disp = nouveau_display(drm->dev);
202	int mod;
203
204	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
205		return -EINVAL;
206	}
207
208	BUG_ON(!disp->format_modifiers);
209
210	for (mod = 0;
211	     (disp->format_modifiers[mod] != DRM_FORMAT_MOD_INVALID) &&
212	     (disp->format_modifiers[mod] != modifier);
213	     mod++);
214
215	if (disp->format_modifiers[mod] == DRM_FORMAT_MOD_INVALID) {
216		for (mod = 0;
217		     (legacy_modifiers[mod] != DRM_FORMAT_MOD_INVALID) &&
218		     (legacy_modifiers[mod] != modifier);
219		     mod++);
220		if (legacy_modifiers[mod] == DRM_FORMAT_MOD_INVALID)
221			return -EINVAL;
222	}
223
224	nouveau_decode_mod(drm, modifier, tile_mode, kind);
225
226	return 0;
227}
228
229static inline uint32_t
230nouveau_get_width_in_blocks(uint32_t stride)
231{
232	/* GOBs per block in the x direction is always one, and GOBs are
233	 * 64 bytes wide
234	 */
235	static const uint32_t log_block_width = 6;
236
237	return (stride + (1 << log_block_width) - 1) >> log_block_width;
238}
239
240static inline uint32_t
241nouveau_get_height_in_blocks(struct nouveau_drm *drm,
242			     uint32_t height,
243			     uint32_t log_block_height_in_gobs)
244{
245	uint32_t log_gob_height;
246	uint32_t log_block_height;
247
248	BUG_ON(drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA);
249
250	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI)
251		log_gob_height = 2;
252	else
253		log_gob_height = 3;
254
255	log_block_height = log_block_height_in_gobs + log_gob_height;
256
257	return (height + (1 << log_block_height) - 1) >> log_block_height;
258}
259
260static int
261nouveau_check_bl_size(struct nouveau_drm *drm, struct nouveau_bo *nvbo,
262		      uint32_t offset, uint32_t stride, uint32_t h,
263		      uint32_t tile_mode)
264{
265	uint32_t gob_size, bw, bh;
266	uint64_t bl_size;
267
268	BUG_ON(drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA);
269
270	if (drm->client.device.info.chipset >= 0xc0) {
271		if (tile_mode & 0xF)
272			return -EINVAL;
273		tile_mode >>= 4;
274	}
275
276	if (tile_mode & 0xFFFFFFF0)
277		return -EINVAL;
278
279	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI)
280		gob_size = 256;
281	else
282		gob_size = 512;
283
284	bw = nouveau_get_width_in_blocks(stride);
285	bh = nouveau_get_height_in_blocks(drm, h, tile_mode);
 
286
287	bl_size = bw * bh * (1 << tile_mode) * gob_size;
288
289	DRM_DEBUG_KMS("offset=%u stride=%u h=%u tile_mode=0x%02x bw=%u bh=%u gob_size=%u bl_size=%llu size=%lu\n",
290		      offset, stride, h, tile_mode, bw, bh, gob_size, bl_size,
291		      nvbo->bo.mem.size);
292
293	if (bl_size + offset > nvbo->bo.mem.size)
294		return -ERANGE;
295
296	return 0;
297}
298
299int
300nouveau_framebuffer_new(struct drm_device *dev,
301			const struct drm_mode_fb_cmd2 *mode_cmd,
302			struct drm_gem_object *gem,
303			struct drm_framebuffer **pfb)
304{
305	struct nouveau_drm *drm = nouveau_drm(dev);
306	struct nouveau_bo *nvbo = nouveau_gem_object(gem);
307	struct drm_framebuffer *fb;
308	const struct drm_format_info *info;
309	unsigned int width, height, i;
310	uint32_t tile_mode;
311	uint8_t kind;
312	int ret;
313
314        /* YUV overlays have special requirements pre-NV50 */
315	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA &&
316
317	    (mode_cmd->pixel_format == DRM_FORMAT_YUYV ||
318	     mode_cmd->pixel_format == DRM_FORMAT_UYVY ||
319	     mode_cmd->pixel_format == DRM_FORMAT_NV12 ||
320	     mode_cmd->pixel_format == DRM_FORMAT_NV21) &&
321	    (mode_cmd->pitches[0] & 0x3f || /* align 64 */
322	     mode_cmd->pitches[0] >= 0x10000 || /* at most 64k pitch */
323	     (mode_cmd->pitches[1] && /* pitches for planes must match */
324	      mode_cmd->pitches[0] != mode_cmd->pitches[1]))) {
325		struct drm_format_name_buf format_name;
326		DRM_DEBUG_KMS("Unsuitable framebuffer: format: %s; pitches: 0x%x\n 0x%x\n",
327			      drm_get_format_name(mode_cmd->pixel_format,
328						  &format_name),
329			      mode_cmd->pitches[0],
330			      mode_cmd->pitches[1]);
331		return -EINVAL;
332	}
333
334	if (mode_cmd->flags & DRM_MODE_FB_MODIFIERS) {
335		if (nouveau_validate_decode_mod(drm, mode_cmd->modifier[0],
336						&tile_mode, &kind)) {
337			DRM_DEBUG_KMS("Unsupported modifier: 0x%llx\n",
338				      mode_cmd->modifier[0]);
339			return -EINVAL;
340		}
341	} else {
342		tile_mode = nvbo->mode;
343		kind = nvbo->kind;
344	}
345
346	info = drm_get_format_info(dev, mode_cmd);
347
348	for (i = 0; i < info->num_planes; i++) {
349		width = drm_format_info_plane_width(info,
350						    mode_cmd->width,
351						    i);
352		height = drm_format_info_plane_height(info,
353						      mode_cmd->height,
354						      i);
355
356		if (kind) {
357			ret = nouveau_check_bl_size(drm, nvbo,
358						    mode_cmd->offsets[i],
359						    mode_cmd->pitches[i],
360						    height, tile_mode);
361			if (ret)
362				return ret;
363		} else {
364			uint32_t size = mode_cmd->pitches[i] * height;
365
366			if (size + mode_cmd->offsets[i] > nvbo->bo.mem.size)
367				return -ERANGE;
368		}
369	}
370
371	if (!(fb = *pfb = kzalloc(sizeof(*fb), GFP_KERNEL)))
372		return -ENOMEM;
373
374	drm_helper_mode_fill_fb_struct(dev, fb, mode_cmd);
375	fb->obj[0] = gem;
376
377	ret = drm_framebuffer_init(dev, fb, &nouveau_framebuffer_funcs);
378	if (ret)
379		kfree(fb);
380	return ret;
381}
382
383struct drm_framebuffer *
384nouveau_user_framebuffer_create(struct drm_device *dev,
385				struct drm_file *file_priv,
386				const struct drm_mode_fb_cmd2 *mode_cmd)
387{
388	struct drm_framebuffer *fb;
389	struct drm_gem_object *gem;
390	int ret;
391
392	gem = drm_gem_object_lookup(file_priv, mode_cmd->handles[0]);
393	if (!gem)
394		return ERR_PTR(-ENOENT);
395
396	ret = nouveau_framebuffer_new(dev, mode_cmd, gem, &fb);
397	if (ret == 0)
398		return fb;
399
400	drm_gem_object_put(gem);
401	return ERR_PTR(ret);
402}
403
404static const struct drm_mode_config_funcs nouveau_mode_config_funcs = {
405	.fb_create = nouveau_user_framebuffer_create,
406	.output_poll_changed = nouveau_fbcon_output_poll_changed,
407};
408
409
410struct nouveau_drm_prop_enum_list {
411	u8 gen_mask;
412	int type;
413	char *name;
414};
415
416static struct nouveau_drm_prop_enum_list underscan[] = {
417	{ 6, UNDERSCAN_AUTO, "auto" },
418	{ 6, UNDERSCAN_OFF, "off" },
419	{ 6, UNDERSCAN_ON, "on" },
420	{}
421};
422
423static struct nouveau_drm_prop_enum_list dither_mode[] = {
424	{ 7, DITHERING_MODE_AUTO, "auto" },
425	{ 7, DITHERING_MODE_OFF, "off" },
426	{ 1, DITHERING_MODE_ON, "on" },
427	{ 6, DITHERING_MODE_STATIC2X2, "static 2x2" },
428	{ 6, DITHERING_MODE_DYNAMIC2X2, "dynamic 2x2" },
429	{ 4, DITHERING_MODE_TEMPORAL, "temporal" },
430	{}
431};
432
433static struct nouveau_drm_prop_enum_list dither_depth[] = {
434	{ 6, DITHERING_DEPTH_AUTO, "auto" },
435	{ 6, DITHERING_DEPTH_6BPC, "6 bpc" },
436	{ 6, DITHERING_DEPTH_8BPC, "8 bpc" },
437	{}
438};
439
440#define PROP_ENUM(p,gen,n,list) do {                                           \
441	struct nouveau_drm_prop_enum_list *l = (list);                         \
442	int c = 0;                                                             \
443	while (l->gen_mask) {                                                  \
444		if (l->gen_mask & (1 << (gen)))                                \
445			c++;                                                   \
446		l++;                                                           \
447	}                                                                      \
448	if (c) {                                                               \
449		p = drm_property_create(dev, DRM_MODE_PROP_ENUM, n, c);        \
450		l = (list);                                                    \
451		while (p && l->gen_mask) {                                     \
452			if (l->gen_mask & (1 << (gen))) {                      \
453				drm_property_add_enum(p, l->type, l->name);    \
454			}                                                      \
455			l++;                                                   \
456		}                                                              \
457	}                                                                      \
458} while(0)
459
 
 
 
 
 
 
 
 
 
 
 
 
 
460static void
461nouveau_display_hpd_work(struct work_struct *work)
462{
463	struct nouveau_drm *drm = container_of(work, typeof(*drm), hpd_work);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
464
465	pm_runtime_get_sync(drm->dev->dev);
 
466
467	drm_helper_hpd_irq_event(drm->dev);
 
 
 
 
 
 
468
469	pm_runtime_mark_last_busy(drm->dev->dev);
470	pm_runtime_put_sync(drm->dev->dev);
 
471}
472
473#ifdef CONFIG_ACPI
474
475static int
476nouveau_display_acpi_ntfy(struct notifier_block *nb, unsigned long val,
477			  void *data)
478{
479	struct nouveau_drm *drm = container_of(nb, typeof(*drm), acpi_nb);
480	struct acpi_bus_event *info = data;
481	int ret;
482
483	if (!strcmp(info->device_class, ACPI_VIDEO_CLASS)) {
484		if (info->type == ACPI_VIDEO_NOTIFY_PROBE) {
485			ret = pm_runtime_get(drm->dev->dev);
486			if (ret == 1 || ret == -EACCES) {
487				/* If the GPU is already awake, or in a state
488				 * where we can't wake it up, it can handle
489				 * it's own hotplug events.
490				 */
491				pm_runtime_put_autosuspend(drm->dev->dev);
492			} else if (ret == 0) {
493				/* This may be the only indication we receive
494				 * of a connector hotplug on a runtime
495				 * suspended GPU, schedule hpd_work to check.
496				 */
497				NV_DEBUG(drm, "ACPI requested connector reprobe\n");
498				schedule_work(&drm->hpd_work);
499				pm_runtime_put_noidle(drm->dev->dev);
500			} else {
501				NV_WARN(drm, "Dropped ACPI reprobe event due to RPM error: %d\n",
502					ret);
503			}
504
505			/* acpi-video should not generate keypresses for this */
506			return NOTIFY_BAD;
507		}
508	}
509
510	return NOTIFY_DONE;
511}
512#endif
513
514int
515nouveau_display_init(struct drm_device *dev, bool resume, bool runtime)
516{
517	struct nouveau_display *disp = nouveau_display(dev);
518	struct drm_connector *connector;
519	struct drm_connector_list_iter conn_iter;
520	int ret;
521
522	/*
523	 * Enable hotplug interrupts (done as early as possible, since we need
524	 * them for MST)
525	 */
526	drm_connector_list_iter_begin(dev, &conn_iter);
527	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
528		struct nouveau_connector *conn = nouveau_connector(connector);
529		nvif_notify_get(&conn->hpd);
 
530	}
531	drm_connector_list_iter_end(&conn_iter);
532
533	ret = disp->init(dev, resume, runtime);
534	if (ret)
535		return ret;
536
537	/* enable connector detection and polling for connectors without HPD
538	 * support
539	 */
540	drm_kms_helper_poll_enable(dev);
541
542	return ret;
543}
544
545void
546nouveau_display_fini(struct drm_device *dev, bool suspend, bool runtime)
547{
548	struct nouveau_display *disp = nouveau_display(dev);
549	struct nouveau_drm *drm = nouveau_drm(dev);
550	struct drm_connector *connector;
551	struct drm_connector_list_iter conn_iter;
552
553	if (!suspend) {
554		if (drm_drv_uses_atomic_modeset(dev))
555			drm_atomic_helper_shutdown(dev);
556		else
557			drm_helper_force_disable_all(dev);
558	}
559
560	/* disable hotplug interrupts */
561	drm_connector_list_iter_begin(dev, &conn_iter);
562	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
563		struct nouveau_connector *conn = nouveau_connector(connector);
564		nvif_notify_put(&conn->hpd);
 
565	}
566	drm_connector_list_iter_end(&conn_iter);
567
568	if (!runtime)
569		cancel_work_sync(&drm->hpd_work);
570
571	drm_kms_helper_poll_disable(dev);
572	disp->fini(dev, suspend);
573}
574
575static void
576nouveau_display_create_properties(struct drm_device *dev)
577{
578	struct nouveau_display *disp = nouveau_display(dev);
579	int gen;
580
581	if (disp->disp.object.oclass < NV50_DISP)
582		gen = 0;
583	else
584	if (disp->disp.object.oclass < GF110_DISP)
585		gen = 1;
586	else
587		gen = 2;
588
589	PROP_ENUM(disp->dithering_mode, gen, "dithering mode", dither_mode);
590	PROP_ENUM(disp->dithering_depth, gen, "dithering depth", dither_depth);
591	PROP_ENUM(disp->underscan_property, gen, "underscan", underscan);
592
593	disp->underscan_hborder_property =
594		drm_property_create_range(dev, 0, "underscan hborder", 0, 128);
595
596	disp->underscan_vborder_property =
597		drm_property_create_range(dev, 0, "underscan vborder", 0, 128);
598
599	if (gen < 1)
600		return;
601
602	/* -90..+90 */
603	disp->vibrant_hue_property =
604		drm_property_create_range(dev, 0, "vibrant hue", 0, 180);
605
606	/* -100..+100 */
607	disp->color_vibrance_property =
608		drm_property_create_range(dev, 0, "color vibrance", 0, 200);
609}
610
611int
612nouveau_display_create(struct drm_device *dev)
613{
614	struct nouveau_drm *drm = nouveau_drm(dev);
615	struct nvkm_device *device = nvxx_device(&drm->client.device);
616	struct nouveau_display *disp;
617	int ret;
618
619	disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL);
620	if (!disp)
621		return -ENOMEM;
622
623	drm_mode_config_init(dev);
624	drm_mode_create_scaling_mode_property(dev);
625	drm_mode_create_dvi_i_properties(dev);
626
627	dev->mode_config.funcs = &nouveau_mode_config_funcs;
628	dev->mode_config.fb_base = device->func->resource_addr(device, 1);
629
630	dev->mode_config.min_width = 0;
631	dev->mode_config.min_height = 0;
632	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_CELSIUS) {
633		dev->mode_config.max_width = 2048;
634		dev->mode_config.max_height = 2048;
635	} else
636	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
637		dev->mode_config.max_width = 4096;
638		dev->mode_config.max_height = 4096;
639	} else
640	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI) {
641		dev->mode_config.max_width = 8192;
642		dev->mode_config.max_height = 8192;
643	} else {
644		dev->mode_config.max_width = 16384;
645		dev->mode_config.max_height = 16384;
646	}
647
648	dev->mode_config.preferred_depth = 24;
649	dev->mode_config.prefer_shadow = 1;
650	dev->mode_config.allow_fb_modifiers = true;
651
652	if (drm->client.device.info.chipset < 0x11)
653		dev->mode_config.async_page_flip = false;
654	else
655		dev->mode_config.async_page_flip = true;
656
657	drm_kms_helper_poll_init(dev);
658	drm_kms_helper_poll_disable(dev);
659
660	if (nouveau_modeset != 2 && drm->vbios.dcb.entries) {
661		ret = nvif_disp_ctor(&drm->client.device, "kmsDisp", 0,
662				     &disp->disp);
663		if (ret == 0) {
 
 
 
 
 
 
664			nouveau_display_create_properties(dev);
665			if (disp->disp.object.oclass < NV50_DISP)
 
666				ret = nv04_display_create(dev);
667			else
668				ret = nv50_display_create(dev);
 
669		}
670	} else {
671		ret = 0;
672	}
673
674	if (ret)
675		goto disp_create_err;
676
677	drm_mode_config_reset(dev);
678
679	if (dev->mode_config.num_crtc) {
680		ret = drm_vblank_init(dev, dev->mode_config.num_crtc);
681		if (ret)
682			goto vblank_err;
683
684		if (disp->disp.object.oclass >= NV50_DISP)
685			nv50_crc_init(dev);
686	}
687
688	INIT_WORK(&drm->hpd_work, nouveau_display_hpd_work);
 
689#ifdef CONFIG_ACPI
690	drm->acpi_nb.notifier_call = nouveau_display_acpi_ntfy;
691	register_acpi_notifier(&drm->acpi_nb);
692#endif
693
694	return 0;
695
696vblank_err:
697	disp->dtor(dev);
698disp_create_err:
699	drm_kms_helper_poll_fini(dev);
700	drm_mode_config_cleanup(dev);
701	return ret;
702}
703
704void
705nouveau_display_destroy(struct drm_device *dev)
706{
707	struct nouveau_display *disp = nouveau_display(dev);
 
708
709#ifdef CONFIG_ACPI
710	unregister_acpi_notifier(&nouveau_drm(dev)->acpi_nb);
711#endif
712
713	drm_kms_helper_poll_fini(dev);
714	drm_mode_config_cleanup(dev);
715
716	if (disp->dtor)
717		disp->dtor(dev);
718
719	nvif_disp_dtor(&disp->disp);
720
721	nouveau_drm(dev)->display = NULL;
722	kfree(disp);
723}
724
725int
726nouveau_display_suspend(struct drm_device *dev, bool runtime)
727{
728	struct nouveau_display *disp = nouveau_display(dev);
729
 
 
730	if (drm_drv_uses_atomic_modeset(dev)) {
731		if (!runtime) {
732			disp->suspend = drm_atomic_helper_suspend(dev);
733			if (IS_ERR(disp->suspend)) {
734				int ret = PTR_ERR(disp->suspend);
735				disp->suspend = NULL;
736				return ret;
737			}
738		}
739	}
740
741	nouveau_display_fini(dev, true, runtime);
742	return 0;
743}
744
745void
746nouveau_display_resume(struct drm_device *dev, bool runtime)
747{
748	struct nouveau_display *disp = nouveau_display(dev);
749
750	nouveau_display_init(dev, true, runtime);
751
752	if (drm_drv_uses_atomic_modeset(dev)) {
753		if (disp->suspend) {
754			drm_atomic_helper_resume(dev, disp->suspend);
755			disp->suspend = NULL;
756		}
757		return;
758	}
 
 
759}
760
761int
762nouveau_display_dumb_create(struct drm_file *file_priv, struct drm_device *dev,
763			    struct drm_mode_create_dumb *args)
764{
765	struct nouveau_cli *cli = nouveau_cli(file_priv);
766	struct nouveau_bo *bo;
767	uint32_t domain;
768	int ret;
769
770	args->pitch = roundup(args->width * (args->bpp / 8), 256);
771	args->size = args->pitch * args->height;
772	args->size = roundup(args->size, PAGE_SIZE);
773
774	/* Use VRAM if there is any ; otherwise fallback to system memory */
775	if (nouveau_drm(dev)->client.device.info.ram_size != 0)
776		domain = NOUVEAU_GEM_DOMAIN_VRAM;
777	else
778		domain = NOUVEAU_GEM_DOMAIN_GART;
779
780	ret = nouveau_gem_new(cli, args->size, 0, domain, 0, 0, &bo);
781	if (ret)
782		return ret;
783
784	ret = drm_gem_handle_create(file_priv, &bo->bo.base, &args->handle);
785	drm_gem_object_put(&bo->bo.base);
786	return ret;
787}
788
789int
790nouveau_display_dumb_map_offset(struct drm_file *file_priv,
791				struct drm_device *dev,
792				uint32_t handle, uint64_t *poffset)
793{
794	struct drm_gem_object *gem;
795
796	gem = drm_gem_object_lookup(file_priv, handle);
797	if (gem) {
798		struct nouveau_bo *bo = nouveau_gem_object(gem);
799		*poffset = drm_vma_node_offset_addr(&bo->bo.base.vma_node);
800		drm_gem_object_put(gem);
801		return 0;
802	}
803
804	return -ENOENT;
805}