Linux Audio

Check our new training course

Loading...
v3.1
  1/*
  2 * Copyright (C) 2008 Maarten Maathuis.
  3 * All Rights Reserved.
  4 *
  5 * Permission is hereby granted, free of charge, to any person obtaining
  6 * a copy of this software and associated documentation files (the
  7 * "Software"), to deal in the Software without restriction, including
  8 * without limitation the rights to use, copy, modify, merge, publish,
  9 * distribute, sublicense, and/or sell copies of the Software, and to
 10 * permit persons to whom the Software is furnished to do so, subject to
 11 * the following conditions:
 12 *
 13 * The above copyright notice and this permission notice (including the
 14 * next paragraph) shall be included in all copies or substantial
 15 * portions of the Software.
 16 *
 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 24 *
 25 */
 26
 27#include "drmP.h"
 28#include "drm_crtc_helper.h"
 29#include "nouveau_drv.h"
 30#include "nouveau_fb.h"
 31#include "nouveau_fbcon.h"
 32#include "nouveau_hw.h"
 
 
 
 
 
 33#include "nouveau_crtc.h"
 34#include "nouveau_dma.h"
 
 35#include "nv50_display.h"
 36
 37static void
 38nouveau_user_framebuffer_destroy(struct drm_framebuffer *drm_fb)
 
 
 
 
 
 39{
 40	struct nouveau_framebuffer *fb = nouveau_framebuffer(drm_fb);
 41
 42	if (fb->nvbo)
 43		drm_gem_object_unreference_unlocked(fb->nvbo->gem);
 44
 45	drm_framebuffer_cleanup(drm_fb);
 46	kfree(fb);
 47}
 48
 49static int
 50nouveau_user_framebuffer_create_handle(struct drm_framebuffer *drm_fb,
 51				       struct drm_file *file_priv,
 52				       unsigned int *handle)
 53{
 54	struct nouveau_framebuffer *fb = nouveau_framebuffer(drm_fb);
 
 
 
 
 55
 56	return drm_gem_handle_create(file_priv, fb->nvbo->gem, handle);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 57}
 58
 59static const struct drm_framebuffer_funcs nouveau_framebuffer_funcs = {
 60	.destroy = nouveau_user_framebuffer_destroy,
 61	.create_handle = nouveau_user_framebuffer_create_handle,
 62};
 63
 64int
 65nouveau_framebuffer_init(struct drm_device *dev,
 66			 struct nouveau_framebuffer *nv_fb,
 67			 struct drm_mode_fb_cmd *mode_cmd,
 68			 struct nouveau_bo *nvbo)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 69{
 70	struct drm_nouveau_private *dev_priv = dev->dev_private;
 71	struct drm_framebuffer *fb = &nv_fb->base;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 72	int ret;
 73
 74	ret = drm_framebuffer_init(dev, fb, &nouveau_framebuffer_funcs);
 75	if (ret) {
 76		return ret;
 
 
 
 
 
 
 
 
 
 
 
 
 77	}
 78
 79	drm_helper_mode_fill_fb_struct(fb, mode_cmd);
 80	nv_fb->nvbo = nvbo;
 
 
 
 
 
 
 
 
 
 81
 82	if (dev_priv->card_type >= NV_50) {
 83		u32 tile_flags = nouveau_bo_tile_layout(nvbo);
 84		if (tile_flags == 0x7a00 ||
 85		    tile_flags == 0xfe00)
 86			nv_fb->r_dma = NvEvoFB32;
 87		else
 88		if (tile_flags == 0x7000)
 89			nv_fb->r_dma = NvEvoFB16;
 90		else
 91			nv_fb->r_dma = NvEvoVRAM_LP;
 
 
 
 
 
 
 92
 93		switch (fb->depth) {
 94		case  8: nv_fb->r_format = NV50_EVO_CRTC_FB_DEPTH_8; break;
 95		case 15: nv_fb->r_format = NV50_EVO_CRTC_FB_DEPTH_15; break;
 96		case 16: nv_fb->r_format = NV50_EVO_CRTC_FB_DEPTH_16; break;
 97		case 24:
 98		case 32: nv_fb->r_format = NV50_EVO_CRTC_FB_DEPTH_24; break;
 99		case 30: nv_fb->r_format = NV50_EVO_CRTC_FB_DEPTH_30; break;
100		default:
101			 NV_ERROR(dev, "unknown depth %d\n", fb->depth);
102			 return -EINVAL;
103		}
104
105		if (dev_priv->chipset == 0x50)
106			nv_fb->r_format |= (tile_flags << 8);
107
108		if (!tile_flags)
109			nv_fb->r_pitch = 0x00100000 | fb->pitch;
110		else {
111			u32 mode = nvbo->tile_mode;
112			if (dev_priv->card_type >= NV_C0)
113				mode >>= 4;
114			nv_fb->r_pitch = ((fb->pitch / 4) << 4) | mode;
115		}
116	}
117
118	return 0;
 
 
 
 
 
 
 
 
 
119}
120
121static struct drm_framebuffer *
122nouveau_user_framebuffer_create(struct drm_device *dev,
123				struct drm_file *file_priv,
124				struct drm_mode_fb_cmd *mode_cmd)
125{
126	struct nouveau_framebuffer *nouveau_fb;
127	struct drm_gem_object *gem;
128	int ret;
129
130	gem = drm_gem_object_lookup(dev, file_priv, mode_cmd->handle);
131	if (!gem)
132		return ERR_PTR(-ENOENT);
133
134	nouveau_fb = kzalloc(sizeof(struct nouveau_framebuffer), GFP_KERNEL);
135	if (!nouveau_fb)
136		return ERR_PTR(-ENOMEM);
137
138	ret = nouveau_framebuffer_init(dev, nouveau_fb, mode_cmd, nouveau_gem_object(gem));
139	if (ret) {
140		drm_gem_object_unreference(gem);
141		return ERR_PTR(ret);
142	}
143
144	return &nouveau_fb->base;
 
145}
146
147const struct drm_mode_config_funcs nouveau_mode_config_funcs = {
148	.fb_create = nouveau_user_framebuffer_create,
149	.output_poll_changed = nouveau_fbcon_output_poll_changed,
150};
151
152int
153nouveau_vblank_enable(struct drm_device *dev, int crtc)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
154{
155	struct drm_nouveau_private *dev_priv = dev->dev_private;
156
157	if (dev_priv->card_type >= NV_50)
158		nv_mask(dev, NV50_PDISPLAY_INTR_EN_1, 0,
159			NV50_PDISPLAY_INTR_EN_1_VBLANK_CRTC_(crtc));
160	else
161		NVWriteCRTC(dev, crtc, NV_PCRTC_INTR_EN_0,
162			    NV_PCRTC_INTR_0_VBLANK);
163
164	return 0;
165}
166
167void
168nouveau_vblank_disable(struct drm_device *dev, int crtc)
169{
170	struct drm_nouveau_private *dev_priv = dev->dev_private;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
171
172	if (dev_priv->card_type >= NV_50)
173		nv_mask(dev, NV50_PDISPLAY_INTR_EN_1,
174			NV50_PDISPLAY_INTR_EN_1_VBLANK_CRTC_(crtc), 0);
175	else
176		NVWriteCRTC(dev, crtc, NV_PCRTC_INTR_EN_0, 0);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
177}
178
 
 
179static int
180nouveau_page_flip_reserve(struct nouveau_bo *old_bo,
181			  struct nouveau_bo *new_bo)
182{
 
 
183	int ret;
184
185	ret = nouveau_bo_pin(new_bo, TTM_PL_FLAG_VRAM);
186	if (ret)
187		return ret;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
188
189	ret = ttm_bo_reserve(&new_bo->bo, false, false, false, 0);
190	if (ret)
191		goto fail;
 
 
 
 
 
192
193	ret = ttm_bo_reserve(&old_bo->bo, false, false, false, 0);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194	if (ret)
195		goto fail_unreserve;
196
197	return 0;
 
 
 
198
199fail_unreserve:
200	ttm_bo_unreserve(&new_bo->bo);
201fail:
202	nouveau_bo_unpin(new_bo);
203	return ret;
204}
205
206static void
207nouveau_page_flip_unreserve(struct nouveau_bo *old_bo,
208			    struct nouveau_bo *new_bo,
209			    struct nouveau_fence *fence)
210{
211	nouveau_bo_fence(new_bo, fence);
212	ttm_bo_unreserve(&new_bo->bo);
 
 
 
 
 
 
 
 
 
213
214	nouveau_bo_fence(old_bo, fence);
215	ttm_bo_unreserve(&old_bo->bo);
 
 
 
 
 
 
216
217	nouveau_bo_unpin(old_bo);
 
 
 
 
218}
219
220static int
221nouveau_page_flip_emit(struct nouveau_channel *chan,
222		       struct nouveau_bo *old_bo,
223		       struct nouveau_bo *new_bo,
224		       struct nouveau_page_flip_state *s,
225		       struct nouveau_fence **pfence)
226{
227	struct drm_nouveau_private *dev_priv = chan->dev->dev_private;
228	struct drm_device *dev = chan->dev;
229	unsigned long flags;
230	int ret;
231
232	/* Queue it to the pending list */
233	spin_lock_irqsave(&dev->event_lock, flags);
234	list_add_tail(&s->head, &chan->nvsw.flip);
235	spin_unlock_irqrestore(&dev->event_lock, flags);
 
 
 
236
237	/* Synchronize with the old framebuffer */
238	ret = nouveau_fence_sync(old_bo->bo.sync_obj, chan);
239	if (ret)
240		goto fail;
241
242	/* Emit the pageflip */
243	ret = RING_SPACE(chan, 2);
244	if (ret)
245		goto fail;
246
247	if (dev_priv->card_type < NV_C0)
248		BEGIN_RING(chan, NvSubSw, NV_SW_PAGE_FLIP, 1);
249	else
250		BEGIN_NVC0(chan, 2, NvSubM2MF, 0x0500, 1);
251	OUT_RING  (chan, 0);
252	FIRE_RING (chan);
253
254	ret = nouveau_fence_new(chan, pfence, true);
255	if (ret)
256		goto fail;
257
258	return 0;
259fail:
260	spin_lock_irqsave(&dev->event_lock, flags);
261	list_del(&s->head);
262	spin_unlock_irqrestore(&dev->event_lock, flags);
263	return ret;
 
264}
265
266int
267nouveau_crtc_page_flip(struct drm_crtc *crtc, struct drm_framebuffer *fb,
268		       struct drm_pending_vblank_event *event)
269{
270	struct drm_device *dev = crtc->dev;
271	struct drm_nouveau_private *dev_priv = dev->dev_private;
272	struct nouveau_bo *old_bo = nouveau_framebuffer(crtc->fb)->nvbo;
273	struct nouveau_bo *new_bo = nouveau_framebuffer(fb)->nvbo;
274	struct nouveau_page_flip_state *s;
275	struct nouveau_channel *chan;
276	struct nouveau_fence *fence;
277	int ret;
278
279	if (!dev_priv->channel)
280		return -ENODEV;
281
282	s = kzalloc(sizeof(*s), GFP_KERNEL);
283	if (!s)
284		return -ENOMEM;
285
286	/* Don't let the buffers go away while we flip */
287	ret = nouveau_page_flip_reserve(old_bo, new_bo);
288	if (ret)
289		goto fail_free;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
290
291	/* Initialize a page flip struct */
292	*s = (struct nouveau_page_flip_state)
293		{ { }, event, nouveau_crtc(crtc)->index,
294		  fb->bits_per_pixel, fb->pitch, crtc->x, crtc->y,
295		  new_bo->bo.offset };
296
297	/* Choose the channel the flip will be handled in */
298	chan = nouveau_fence_channel(new_bo->bo.sync_obj);
299	if (!chan)
300		chan = nouveau_channel_get_unlocked(dev_priv->channel);
301	mutex_lock(&chan->mutex);
302
303	/* Emit a page flip */
304	if (dev_priv->card_type >= NV_50) {
305		ret = nv50_display_flip_next(crtc, fb, chan);
306		if (ret) {
307			nouveau_channel_put(&chan);
308			goto fail_unreserve;
 
 
 
 
 
 
309		}
 
 
310	}
311
312	ret = nouveau_page_flip_emit(chan, old_bo, new_bo, s, &fence);
313	nouveau_channel_put(&chan);
314	if (ret)
315		goto fail_unreserve;
 
 
316
317	/* Update the crtc struct and cleanup */
318	crtc->fb = fb;
 
 
 
 
 
 
 
 
 
 
 
 
 
319
320	nouveau_page_flip_unreserve(old_bo, new_bo, fence);
321	nouveau_fence_unref(&fence);
322	return 0;
323
324fail_unreserve:
325	nouveau_page_flip_unreserve(old_bo, new_bo, NULL);
326fail_free:
327	kfree(s);
 
328	return ret;
329}
330
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
331int
332nouveau_finish_page_flip(struct nouveau_channel *chan,
333			 struct nouveau_page_flip_state *ps)
334{
335	struct drm_device *dev = chan->dev;
336	struct nouveau_page_flip_state *s;
337	unsigned long flags;
338
339	spin_lock_irqsave(&dev->event_lock, flags);
340
341	if (list_empty(&chan->nvsw.flip)) {
342		NV_ERROR(dev, "Unexpected pageflip in channel %d.\n", chan->id);
343		spin_unlock_irqrestore(&dev->event_lock, flags);
344		return -EINVAL;
345	}
346
347	s = list_first_entry(&chan->nvsw.flip,
348			     struct nouveau_page_flip_state, head);
349	if (s->event) {
350		struct drm_pending_vblank_event *e = s->event;
351		struct timeval now;
352
353		do_gettimeofday(&now);
354		e->event.sequence = 0;
355		e->event.tv_sec = now.tv_sec;
356		e->event.tv_usec = now.tv_usec;
357		list_add_tail(&e->base.link, &e->base.file_priv->event_list);
358		wake_up_interruptible(&e->base.file_priv->event_wait);
359	}
360
361	list_del(&s->head);
362	if (ps)
363		*ps = *s;
364	kfree(s);
365
366	spin_unlock_irqrestore(&dev->event_lock, flags);
367	return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
368}
v6.8
  1/*
  2 * Copyright (C) 2008 Maarten Maathuis.
  3 * All Rights Reserved.
  4 *
  5 * Permission is hereby granted, free of charge, to any person obtaining
  6 * a copy of this software and associated documentation files (the
  7 * "Software"), to deal in the Software without restriction, including
  8 * without limitation the rights to use, copy, modify, merge, publish,
  9 * distribute, sublicense, and/or sell copies of the Software, and to
 10 * permit persons to whom the Software is furnished to do so, subject to
 11 * the following conditions:
 12 *
 13 * The above copyright notice and this permission notice (including the
 14 * next paragraph) shall be included in all copies or substantial
 15 * portions of the Software.
 16 *
 17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 18 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 19 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
 20 * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
 21 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
 22 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
 23 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
 24 *
 25 */
 26
 27#include <acpi/video.h>
 28
 29#include <drm/drm_atomic.h>
 30#include <drm/drm_atomic_helper.h>
 31#include <drm/drm_crtc_helper.h>
 32#include <drm/drm_fb_helper.h>
 33#include <drm/drm_fourcc.h>
 34#include <drm/drm_gem_framebuffer_helper.h>
 35#include <drm/drm_probe_helper.h>
 36#include <drm/drm_vblank.h>
 37
 38#include "nouveau_crtc.h"
 39#include "nouveau_gem.h"
 40#include "nouveau_connector.h"
 41#include "nv50_display.h"
 42
 43#include <nvif/class.h>
 44#include <nvif/if0011.h>
 45#include <nvif/if0013.h>
 46#include <dispnv50/crc.h>
 47
 48int
 49nouveau_display_vblank_enable(struct drm_crtc *crtc)
 50{
 51	struct nouveau_crtc *nv_crtc;
 52
 53	nv_crtc = nouveau_crtc(crtc);
 54	nvif_event_allow(&nv_crtc->vblank);
 55
 56	return 0;
 
 57}
 58
 59void
 60nouveau_display_vblank_disable(struct drm_crtc *crtc)
 
 
 61{
 62	struct nouveau_crtc *nv_crtc;
 63
 64	nv_crtc = nouveau_crtc(crtc);
 65	nvif_event_block(&nv_crtc->vblank);
 66}
 67
 68static inline int
 69calc(int blanks, int blanke, int total, int line)
 70{
 71	if (blanke >= blanks) {
 72		if (line >= blanks)
 73			line -= total;
 74	} else {
 75		if (line >= blanks)
 76			line -= total;
 77		line -= blanke + 1;
 78	}
 79	return line;
 80}
 81
 82static bool
 83nouveau_display_scanoutpos_head(struct drm_crtc *crtc, int *vpos, int *hpos,
 84				ktime_t *stime, ktime_t *etime)
 85{
 86	struct drm_vblank_crtc *vblank = &crtc->dev->vblank[drm_crtc_index(crtc)];
 87	struct nvif_head *head = &nouveau_crtc(crtc)->head;
 88	struct nvif_head_scanoutpos_v0 args;
 89	int retry = 20;
 90	bool ret = false;
 91
 92	args.version = 0;
 93
 94	do {
 95		ret = nvif_mthd(&head->object, NVIF_HEAD_V0_SCANOUTPOS, &args, sizeof(args));
 96		if (ret != 0)
 97			return false;
 98
 99		if (args.vline) {
100			ret = true;
101			break;
102		}
103
104		if (retry) ndelay(vblank->linedur_ns);
105	} while (retry--);
106
107	*hpos = args.hline;
108	*vpos = calc(args.vblanks, args.vblanke, args.vtotal, args.vline);
109	if (stime) *stime = ns_to_ktime(args.time[0]);
110	if (etime) *etime = ns_to_ktime(args.time[1]);
111
112	return ret;
113}
114
115bool
116nouveau_display_scanoutpos(struct drm_crtc *crtc,
117			   bool in_vblank_irq, int *vpos, int *hpos,
118			   ktime_t *stime, ktime_t *etime,
119			   const struct drm_display_mode *mode)
120{
121	return nouveau_display_scanoutpos_head(crtc, vpos, hpos,
122					       stime, etime);
123}
124
125static const struct drm_framebuffer_funcs nouveau_framebuffer_funcs = {
126	.destroy = drm_gem_fb_destroy,
127	.create_handle = drm_gem_fb_create_handle,
128};
129
130static void
131nouveau_decode_mod(struct nouveau_drm *drm,
132		   uint64_t modifier,
133		   uint32_t *tile_mode,
134		   uint8_t *kind)
135{
136	struct nouveau_display *disp = nouveau_display(drm->dev);
137	BUG_ON(!tile_mode || !kind);
138
139	if (modifier == DRM_FORMAT_MOD_LINEAR) {
140		/* tile_mode will not be used in this case */
141		*tile_mode = 0;
142		*kind = 0;
143	} else {
144		/*
145		 * Extract the block height and kind from the corresponding
146		 * modifier fields.  See drm_fourcc.h for details.
147		 */
148
149		if ((modifier & (0xffull << 12)) == 0ull) {
150			/* Legacy modifier.  Translate to this dev's 'kind.' */
151			modifier |= disp->format_modifiers[0] & (0xffull << 12);
152		}
153
154		*tile_mode = (uint32_t)(modifier & 0xF);
155		*kind = (uint8_t)((modifier >> 12) & 0xFF);
156
157		if (drm->client.device.info.chipset >= 0xc0)
158			*tile_mode <<= 4;
159	}
160}
161
162void
163nouveau_framebuffer_get_layout(struct drm_framebuffer *fb,
164			       uint32_t *tile_mode,
165			       uint8_t *kind)
166{
167	if (fb->flags & DRM_MODE_FB_MODIFIERS) {
168		struct nouveau_drm *drm = nouveau_drm(fb->dev);
169
170		nouveau_decode_mod(drm, fb->modifier, tile_mode, kind);
171	} else {
172		const struct nouveau_bo *nvbo = nouveau_gem_object(fb->obj[0]);
173
174		*tile_mode = nvbo->mode;
175		*kind = nvbo->kind;
176	}
177}
178
179static const u64 legacy_modifiers[] = {
180	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0),
181	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1),
182	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2),
183	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3),
184	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4),
185	DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5),
186	DRM_FORMAT_MOD_INVALID
187};
188
189static int
190nouveau_validate_decode_mod(struct nouveau_drm *drm,
191			    uint64_t modifier,
192			    uint32_t *tile_mode,
193			    uint8_t *kind)
194{
195	struct nouveau_display *disp = nouveau_display(drm->dev);
196	int mod;
197
198	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
199		return -EINVAL;
200	}
201
202	BUG_ON(!disp->format_modifiers);
203
204	for (mod = 0;
205	     (disp->format_modifiers[mod] != DRM_FORMAT_MOD_INVALID) &&
206	     (disp->format_modifiers[mod] != modifier);
207	     mod++);
208
209	if (disp->format_modifiers[mod] == DRM_FORMAT_MOD_INVALID) {
210		for (mod = 0;
211		     (legacy_modifiers[mod] != DRM_FORMAT_MOD_INVALID) &&
212		     (legacy_modifiers[mod] != modifier);
213		     mod++);
214		if (legacy_modifiers[mod] == DRM_FORMAT_MOD_INVALID)
215			return -EINVAL;
216	}
217
218	nouveau_decode_mod(drm, modifier, tile_mode, kind);
219
220	return 0;
221}
222
223static inline uint32_t
224nouveau_get_width_in_blocks(uint32_t stride)
225{
226	/* GOBs per block in the x direction is always one, and GOBs are
227	 * 64 bytes wide
228	 */
229	static const uint32_t log_block_width = 6;
230
231	return (stride + (1 << log_block_width) - 1) >> log_block_width;
232}
233
234static inline uint32_t
235nouveau_get_height_in_blocks(struct nouveau_drm *drm,
236			     uint32_t height,
237			     uint32_t log_block_height_in_gobs)
238{
239	uint32_t log_gob_height;
240	uint32_t log_block_height;
241
242	BUG_ON(drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA);
243
244	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI)
245		log_gob_height = 2;
246	else
247		log_gob_height = 3;
248
249	log_block_height = log_block_height_in_gobs + log_gob_height;
250
251	return (height + (1 << log_block_height) - 1) >> log_block_height;
252}
253
254static int
255nouveau_check_bl_size(struct nouveau_drm *drm, struct nouveau_bo *nvbo,
256		      uint32_t offset, uint32_t stride, uint32_t h,
257		      uint32_t tile_mode)
258{
259	uint32_t gob_size, bw, bh;
260	uint64_t bl_size;
261
262	BUG_ON(drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA);
263
264	if (drm->client.device.info.chipset >= 0xc0) {
265		if (tile_mode & 0xF)
266			return -EINVAL;
267		tile_mode >>= 4;
268	}
269
270	if (tile_mode & 0xFFFFFFF0)
271		return -EINVAL;
272
273	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI)
274		gob_size = 256;
275	else
276		gob_size = 512;
277
278	bw = nouveau_get_width_in_blocks(stride);
279	bh = nouveau_get_height_in_blocks(drm, h, tile_mode);
280
281	bl_size = bw * bh * (1 << tile_mode) * gob_size;
282
283	DRM_DEBUG_KMS("offset=%u stride=%u h=%u tile_mode=0x%02x bw=%u bh=%u gob_size=%u bl_size=%llu size=%zu\n",
284		      offset, stride, h, tile_mode, bw, bh, gob_size, bl_size,
285		      nvbo->bo.base.size);
286
287	if (bl_size + offset > nvbo->bo.base.size)
288		return -ERANGE;
289
290	return 0;
291}
292
293int
294nouveau_framebuffer_new(struct drm_device *dev,
295			const struct drm_mode_fb_cmd2 *mode_cmd,
296			struct drm_gem_object *gem,
297			struct drm_framebuffer **pfb)
298{
299	struct nouveau_drm *drm = nouveau_drm(dev);
300	struct nouveau_bo *nvbo = nouveau_gem_object(gem);
301	struct drm_framebuffer *fb;
302	const struct drm_format_info *info;
303	unsigned int height, i;
304	uint32_t tile_mode;
305	uint8_t kind;
306	int ret;
307
308        /* YUV overlays have special requirements pre-NV50 */
309	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA &&
310
311	    (mode_cmd->pixel_format == DRM_FORMAT_YUYV ||
312	     mode_cmd->pixel_format == DRM_FORMAT_UYVY ||
313	     mode_cmd->pixel_format == DRM_FORMAT_NV12 ||
314	     mode_cmd->pixel_format == DRM_FORMAT_NV21) &&
315	    (mode_cmd->pitches[0] & 0x3f || /* align 64 */
316	     mode_cmd->pitches[0] >= 0x10000 || /* at most 64k pitch */
317	     (mode_cmd->pitches[1] && /* pitches for planes must match */
318	      mode_cmd->pitches[0] != mode_cmd->pitches[1]))) {
319		DRM_DEBUG_KMS("Unsuitable framebuffer: format: %p4cc; pitches: 0x%x\n 0x%x\n",
320			      &mode_cmd->pixel_format,
321			      mode_cmd->pitches[0], mode_cmd->pitches[1]);
322		return -EINVAL;
323	}
324
325	if (mode_cmd->flags & DRM_MODE_FB_MODIFIERS) {
326		if (nouveau_validate_decode_mod(drm, mode_cmd->modifier[0],
327						&tile_mode, &kind)) {
328			DRM_DEBUG_KMS("Unsupported modifier: 0x%llx\n",
329				      mode_cmd->modifier[0]);
330			return -EINVAL;
331		}
332	} else {
333		tile_mode = nvbo->mode;
334		kind = nvbo->kind;
335	}
336
337	info = drm_get_format_info(dev, mode_cmd);
338
339	for (i = 0; i < info->num_planes; i++) {
340		height = drm_format_info_plane_height(info,
341						      mode_cmd->height,
342						      i);
343
344		if (kind) {
345			ret = nouveau_check_bl_size(drm, nvbo,
346						    mode_cmd->offsets[i],
347						    mode_cmd->pitches[i],
348						    height, tile_mode);
349			if (ret)
350				return ret;
351		} else {
352			uint32_t size = mode_cmd->pitches[i] * height;
353
354			if (size + mode_cmd->offsets[i] > nvbo->bo.base.size)
355				return -ERANGE;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
356		}
357	}
358
359	if (!(fb = *pfb = kzalloc(sizeof(*fb), GFP_KERNEL)))
360		return -ENOMEM;
361
362	drm_helper_mode_fill_fb_struct(dev, fb, mode_cmd);
363	fb->obj[0] = gem;
364
365	ret = drm_framebuffer_init(dev, fb, &nouveau_framebuffer_funcs);
366	if (ret)
367		kfree(fb);
368	return ret;
369}
370
371struct drm_framebuffer *
372nouveau_user_framebuffer_create(struct drm_device *dev,
373				struct drm_file *file_priv,
374				const struct drm_mode_fb_cmd2 *mode_cmd)
375{
376	struct drm_framebuffer *fb;
377	struct drm_gem_object *gem;
378	int ret;
379
380	gem = drm_gem_object_lookup(file_priv, mode_cmd->handles[0]);
381	if (!gem)
382		return ERR_PTR(-ENOENT);
383
384	ret = nouveau_framebuffer_new(dev, mode_cmd, gem, &fb);
385	if (ret == 0)
386		return fb;
 
 
 
 
 
 
387
388	drm_gem_object_put(gem);
389	return ERR_PTR(ret);
390}
391
392static const struct drm_mode_config_funcs nouveau_mode_config_funcs = {
393	.fb_create = nouveau_user_framebuffer_create,
394	.output_poll_changed = drm_fb_helper_output_poll_changed,
395};
396
397
398struct nouveau_drm_prop_enum_list {
399	u8 gen_mask;
400	int type;
401	char *name;
402};
403
404static struct nouveau_drm_prop_enum_list underscan[] = {
405	{ 6, UNDERSCAN_AUTO, "auto" },
406	{ 6, UNDERSCAN_OFF, "off" },
407	{ 6, UNDERSCAN_ON, "on" },
408	{}
409};
410
411static struct nouveau_drm_prop_enum_list dither_mode[] = {
412	{ 7, DITHERING_MODE_AUTO, "auto" },
413	{ 7, DITHERING_MODE_OFF, "off" },
414	{ 1, DITHERING_MODE_ON, "on" },
415	{ 6, DITHERING_MODE_STATIC2X2, "static 2x2" },
416	{ 6, DITHERING_MODE_DYNAMIC2X2, "dynamic 2x2" },
417	{ 4, DITHERING_MODE_TEMPORAL, "temporal" },
418	{}
419};
420
421static struct nouveau_drm_prop_enum_list dither_depth[] = {
422	{ 6, DITHERING_DEPTH_AUTO, "auto" },
423	{ 6, DITHERING_DEPTH_6BPC, "6 bpc" },
424	{ 6, DITHERING_DEPTH_8BPC, "8 bpc" },
425	{}
426};
427
428#define PROP_ENUM(p,gen,n,list) do {                                           \
429	struct nouveau_drm_prop_enum_list *l = (list);                         \
430	int c = 0;                                                             \
431	while (l->gen_mask) {                                                  \
432		if (l->gen_mask & (1 << (gen)))                                \
433			c++;                                                   \
434		l++;                                                           \
435	}                                                                      \
436	if (c) {                                                               \
437		p = drm_property_create(dev, DRM_MODE_PROP_ENUM, n, c);        \
438		l = (list);                                                    \
439		while (p && l->gen_mask) {                                     \
440			if (l->gen_mask & (1 << (gen))) {                      \
441				drm_property_add_enum(p, l->type, l->name);    \
442			}                                                      \
443			l++;                                                   \
444		}                                                              \
445	}                                                                      \
446} while(0)
447
448void
449nouveau_display_hpd_resume(struct drm_device *dev)
450{
451	struct nouveau_drm *drm = nouveau_drm(dev);
452
453	spin_lock_irq(&drm->hpd_lock);
454	drm->hpd_pending = ~0;
455	spin_unlock_irq(&drm->hpd_lock);
 
 
 
456
457	schedule_work(&drm->hpd_work);
458}
459
460static void
461nouveau_display_hpd_work(struct work_struct *work)
462{
463	struct nouveau_drm *drm = container_of(work, typeof(*drm), hpd_work);
464	struct drm_device *dev = drm->dev;
465	struct drm_connector *connector;
466	struct drm_connector_list_iter conn_iter;
467	u32 pending;
468	int changed = 0;
469	struct drm_connector *first_changed_connector = NULL;
470
471	pm_runtime_get_sync(dev->dev);
472
473	spin_lock_irq(&drm->hpd_lock);
474	pending = drm->hpd_pending;
475	drm->hpd_pending = 0;
476	spin_unlock_irq(&drm->hpd_lock);
477
478	/* Nothing to do, exit early without updating the last busy counter */
479	if (!pending)
480		goto noop;
481
482	mutex_lock(&dev->mode_config.mutex);
483	drm_connector_list_iter_begin(dev, &conn_iter);
484
485	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
486		struct nouveau_connector *nv_connector = nouveau_connector(connector);
487		enum drm_connector_status old_status = connector->status;
488		u64 bits, old_epoch_counter = connector->epoch_counter;
489
490		if (!(pending & drm_connector_mask(connector)))
491			continue;
492
493		spin_lock_irq(&drm->hpd_lock);
494		bits = nv_connector->hpd_pending;
495		nv_connector->hpd_pending = 0;
496		spin_unlock_irq(&drm->hpd_lock);
497
498		drm_dbg_kms(dev, "[CONNECTOR:%d:%s] plug:%d unplug:%d irq:%d\n",
499			    connector->base.id, connector->name,
500			    !!(bits & NVIF_CONN_EVENT_V0_PLUG),
501			    !!(bits & NVIF_CONN_EVENT_V0_UNPLUG),
502			    !!(bits & NVIF_CONN_EVENT_V0_IRQ));
503
504		if (bits & NVIF_CONN_EVENT_V0_IRQ) {
505			if (nouveau_dp_link_check(nv_connector))
506				continue;
507		}
508
509		connector->status = drm_helper_probe_detect(connector, NULL, false);
510		if (old_epoch_counter == connector->epoch_counter)
511			continue;
512
513		changed++;
514		if (!first_changed_connector) {
515			drm_connector_get(connector);
516			first_changed_connector = connector;
517		}
518
519		drm_dbg_kms(dev, "[CONNECTOR:%d:%s] status updated from %s to %s (epoch counter %llu->%llu)\n",
520			    connector->base.id, connector->name,
521			    drm_get_connector_status_name(old_status),
522			    drm_get_connector_status_name(connector->status),
523			    old_epoch_counter, connector->epoch_counter);
524	}
525
526	drm_connector_list_iter_end(&conn_iter);
527	mutex_unlock(&dev->mode_config.mutex);
528
529	if (changed == 1)
530		drm_kms_helper_connector_hotplug_event(first_changed_connector);
531	else if (changed > 0)
532		drm_kms_helper_hotplug_event(dev);
533
534	if (first_changed_connector)
535		drm_connector_put(first_changed_connector);
536
537	pm_runtime_mark_last_busy(drm->dev->dev);
538noop:
539	pm_runtime_put_autosuspend(dev->dev);
540}
541
542#ifdef CONFIG_ACPI
543
544static int
545nouveau_display_acpi_ntfy(struct notifier_block *nb, unsigned long val,
546			  void *data)
547{
548	struct nouveau_drm *drm = container_of(nb, typeof(*drm), acpi_nb);
549	struct acpi_bus_event *info = data;
550	int ret;
551
552	if (!strcmp(info->device_class, ACPI_VIDEO_CLASS)) {
553		if (info->type == ACPI_VIDEO_NOTIFY_PROBE) {
554			ret = pm_runtime_get(drm->dev->dev);
555			if (ret == 1 || ret == -EACCES) {
556				/* If the GPU is already awake, or in a state
557				 * where we can't wake it up, it can handle
558				 * it's own hotplug events.
559				 */
560				pm_runtime_put_autosuspend(drm->dev->dev);
561			} else if (ret == 0 || ret == -EINPROGRESS) {
562				/* We've started resuming the GPU already, so
563				 * it will handle scheduling a full reprobe
564				 * itself
565				 */
566				NV_DEBUG(drm, "ACPI requested connector reprobe\n");
567				pm_runtime_put_noidle(drm->dev->dev);
568			} else {
569				NV_WARN(drm, "Dropped ACPI reprobe event due to RPM error: %d\n",
570					ret);
571			}
572
573			/* acpi-video should not generate keypresses for this */
574			return NOTIFY_BAD;
575		}
576	}
577
578	return NOTIFY_DONE;
579}
580#endif
581
582int
583nouveau_display_init(struct drm_device *dev, bool resume, bool runtime)
584{
585	struct nouveau_display *disp = nouveau_display(dev);
586	struct drm_connector *connector;
587	struct drm_connector_list_iter conn_iter;
588	int ret;
589
590	/*
591	 * Enable hotplug interrupts (done as early as possible, since we need
592	 * them for MST)
593	 */
594	drm_connector_list_iter_begin(dev, &conn_iter);
595	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
596		struct nouveau_connector *conn = nouveau_connector(connector);
597		nvif_event_allow(&conn->hpd);
598		nvif_event_allow(&conn->irq);
599	}
600	drm_connector_list_iter_end(&conn_iter);
601
602	ret = disp->init(dev, resume, runtime);
603	if (ret)
604		return ret;
605
606	/* enable connector detection and polling for connectors without HPD
607	 * support
608	 */
609	drm_kms_helper_poll_enable(dev);
610
 
 
 
 
611	return ret;
612}
613
614void
615nouveau_display_fini(struct drm_device *dev, bool suspend, bool runtime)
 
 
616{
617	struct nouveau_display *disp = nouveau_display(dev);
618	struct nouveau_drm *drm = nouveau_drm(dev);
619	struct drm_connector *connector;
620	struct drm_connector_list_iter conn_iter;
621
622	if (!suspend) {
623		if (drm_drv_uses_atomic_modeset(dev))
624			drm_atomic_helper_shutdown(dev);
625		else
626			drm_helper_force_disable_all(dev);
627	}
628
629	/* disable hotplug interrupts */
630	drm_connector_list_iter_begin(dev, &conn_iter);
631	nouveau_for_each_non_mst_connector_iter(connector, &conn_iter) {
632		struct nouveau_connector *conn = nouveau_connector(connector);
633		nvif_event_block(&conn->irq);
634		nvif_event_block(&conn->hpd);
635	}
636	drm_connector_list_iter_end(&conn_iter);
637
638	if (!runtime)
639		cancel_work_sync(&drm->hpd_work);
640
641	drm_kms_helper_poll_disable(dev);
642	disp->fini(dev, runtime, suspend);
643}
644
645static void
646nouveau_display_create_properties(struct drm_device *dev)
647{
648	struct nouveau_display *disp = nouveau_display(dev);
649	int gen;
 
 
 
 
 
 
650
651	if (disp->disp.object.oclass < NV50_DISP)
652		gen = 0;
653	else
654	if (disp->disp.object.oclass < GF110_DISP)
655		gen = 1;
656	else
657		gen = 2;
658
659	PROP_ENUM(disp->dithering_mode, gen, "dithering mode", dither_mode);
660	PROP_ENUM(disp->dithering_depth, gen, "dithering depth", dither_depth);
661	PROP_ENUM(disp->underscan_property, gen, "underscan", underscan);
 
662
663	disp->underscan_hborder_property =
664		drm_property_create_range(dev, 0, "underscan hborder", 0, 128);
 
 
665
666	disp->underscan_vborder_property =
667		drm_property_create_range(dev, 0, "underscan vborder", 0, 128);
 
 
 
 
668
669	if (gen < 1)
670		return;
 
671
672	/* -90..+90 */
673	disp->vibrant_hue_property =
674		drm_property_create_range(dev, 0, "vibrant hue", 0, 180);
675
676	/* -100..+100 */
677	disp->color_vibrance_property =
678		drm_property_create_range(dev, 0, "color vibrance", 0, 200);
679}
680
681int
682nouveau_display_create(struct drm_device *dev)
 
683{
684	struct nouveau_drm *drm = nouveau_drm(dev);
685	struct nouveau_display *disp;
 
 
 
 
 
686	int ret;
687
688	disp = drm->display = kzalloc(sizeof(*disp), GFP_KERNEL);
689	if (!disp)
 
 
 
690		return -ENOMEM;
691
692	drm_mode_config_init(dev);
693	drm_mode_create_scaling_mode_property(dev);
694	drm_mode_create_dvi_i_properties(dev);
695
696	dev->mode_config.funcs = &nouveau_mode_config_funcs;
697
698	dev->mode_config.min_width = 0;
699	dev->mode_config.min_height = 0;
700	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_CELSIUS) {
701		dev->mode_config.max_width = 2048;
702		dev->mode_config.max_height = 2048;
703	} else
704	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_TESLA) {
705		dev->mode_config.max_width = 4096;
706		dev->mode_config.max_height = 4096;
707	} else
708	if (drm->client.device.info.family < NV_DEVICE_INFO_V0_FERMI) {
709		dev->mode_config.max_width = 8192;
710		dev->mode_config.max_height = 8192;
711	} else {
712		dev->mode_config.max_width = 16384;
713		dev->mode_config.max_height = 16384;
714	}
715
716	dev->mode_config.preferred_depth = 24;
717	dev->mode_config.prefer_shadow = 1;
718
719	if (drm->client.device.info.chipset < 0x11)
720		dev->mode_config.async_page_flip = false;
721	else
722		dev->mode_config.async_page_flip = true;
723
724	drm_kms_helper_poll_init(dev);
725	drm_kms_helper_poll_disable(dev);
726
727	if (nouveau_modeset != 2) {
728		ret = nvif_disp_ctor(&drm->client.device, "kmsDisp", 0, &disp->disp);
729		/* no display hw */
730		if (ret == -ENODEV) {
731			ret = 0;
732			goto disp_create_err;
733		}
734
735		if (!ret && (disp->disp.outp_mask || drm->vbios.dcb.entries)) {
736			nouveau_display_create_properties(dev);
737			if (disp->disp.object.oclass < NV50_DISP) {
738				dev->mode_config.fb_modifiers_not_supported = true;
739				ret = nv04_display_create(dev);
740			} else {
741				ret = nv50_display_create(dev);
742			}
743		}
744	} else {
745		ret = 0;
746	}
747
 
 
748	if (ret)
749		goto disp_create_err;
750
751	drm_mode_config_reset(dev);
752
753	if (dev->mode_config.num_crtc) {
754		ret = drm_vblank_init(dev, dev->mode_config.num_crtc);
755		if (ret)
756			goto vblank_err;
757
758		if (disp->disp.object.oclass >= NV50_DISP)
759			nv50_crc_init(dev);
760	}
761
762	INIT_WORK(&drm->hpd_work, nouveau_display_hpd_work);
763	spin_lock_init(&drm->hpd_lock);
764#ifdef CONFIG_ACPI
765	drm->acpi_nb.notifier_call = nouveau_display_acpi_ntfy;
766	register_acpi_notifier(&drm->acpi_nb);
767#endif
768
 
 
769	return 0;
770
771vblank_err:
772	disp->dtor(dev);
773disp_create_err:
774	drm_kms_helper_poll_fini(dev);
775	drm_mode_config_cleanup(dev);
776	return ret;
777}
778
779void
780nouveau_display_destroy(struct drm_device *dev)
781{
782	struct nouveau_display *disp = nouveau_display(dev);
783	struct nouveau_drm *drm = nouveau_drm(dev);
784
785#ifdef CONFIG_ACPI
786	unregister_acpi_notifier(&drm->acpi_nb);
787#endif
788
789	drm_kms_helper_poll_fini(dev);
790	drm_mode_config_cleanup(dev);
791
792	if (disp->dtor)
793		disp->dtor(dev);
794
795	nvif_disp_dtor(&disp->disp);
796
797	drm->display = NULL;
798	kfree(disp);
799}
800
801int
802nouveau_display_suspend(struct drm_device *dev, bool runtime)
 
803{
804	struct nouveau_display *disp = nouveau_display(dev);
805
806	/* Disable console. */
807	drm_fb_helper_set_suspend_unlocked(dev->fb_helper, true);
 
 
 
 
 
 
 
808
809	if (drm_drv_uses_atomic_modeset(dev)) {
810		if (!runtime) {
811			disp->suspend = drm_atomic_helper_suspend(dev);
812			if (IS_ERR(disp->suspend)) {
813				int ret = PTR_ERR(disp->suspend);
814				disp->suspend = NULL;
815				return ret;
816			}
817		}
818	}
 
 
 
 
 
 
 
 
819
820	nouveau_display_fini(dev, true, runtime);
821	return 0;
822}
823
824void
825nouveau_display_resume(struct drm_device *dev, bool runtime)
826{
827	struct nouveau_display *disp = nouveau_display(dev);
828
829	nouveau_display_init(dev, true, runtime);
830
831	if (drm_drv_uses_atomic_modeset(dev)) {
832		if (disp->suspend) {
833			drm_atomic_helper_resume(dev, disp->suspend);
834			disp->suspend = NULL;
835		}
836	}
837
838	/* Enable console. */
839	drm_fb_helper_set_suspend_unlocked(dev->fb_helper, false);
840}
841
842int
843nouveau_display_dumb_create(struct drm_file *file_priv, struct drm_device *dev,
844			    struct drm_mode_create_dumb *args)
845{
846	struct nouveau_cli *cli = nouveau_cli(file_priv);
847	struct nouveau_bo *bo;
848	uint32_t domain;
849	int ret;
850
851	args->pitch = roundup(args->width * (args->bpp / 8), 256);
852	args->size = args->pitch * args->height;
853	args->size = roundup(args->size, PAGE_SIZE);
854
855	/* Use VRAM if there is any ; otherwise fallback to system memory */
856	if (nouveau_drm(dev)->client.device.info.ram_size != 0)
857		domain = NOUVEAU_GEM_DOMAIN_VRAM;
858	else
859		domain = NOUVEAU_GEM_DOMAIN_GART;
860
861	ret = nouveau_gem_new(cli, args->size, 0, domain, 0, 0, &bo);
862	if (ret)
863		return ret;
864
865	ret = drm_gem_handle_create(file_priv, &bo->bo.base, &args->handle);
866	drm_gem_object_put(&bo->bo.base);
867	return ret;
868}