Loading...
1/*
2 * Copyright 2008 Advanced Micro Devices, Inc.
3 * Copyright 2008 Red Hat Inc.
4 * Copyright 2009 Jerome Glisse.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22 * OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors: Dave Airlie
25 * Alex Deucher
26 * Jerome Glisse
27 */
28#include <linux/firmware.h>
29#include <linux/platform_device.h>
30#include <linux/slab.h>
31#include "drmP.h"
32#include "radeon.h"
33#include "radeon_asic.h"
34#include "radeon_drm.h"
35#include "rv770d.h"
36#include "atom.h"
37#include "avivod.h"
38
39#define R700_PFP_UCODE_SIZE 848
40#define R700_PM4_UCODE_SIZE 1360
41
42static void rv770_gpu_init(struct radeon_device *rdev);
43void rv770_fini(struct radeon_device *rdev);
44static void rv770_pcie_gen2_enable(struct radeon_device *rdev);
45
46u32 rv770_page_flip(struct radeon_device *rdev, int crtc_id, u64 crtc_base)
47{
48 struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
49 u32 tmp = RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset);
50
51 /* Lock the graphics update lock */
52 tmp |= AVIVO_D1GRPH_UPDATE_LOCK;
53 WREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset, tmp);
54
55 /* update the scanout addresses */
56 if (radeon_crtc->crtc_id) {
57 WREG32(D2GRPH_SECONDARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
58 WREG32(D2GRPH_PRIMARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
59 } else {
60 WREG32(D1GRPH_SECONDARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
61 WREG32(D1GRPH_PRIMARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
62 }
63 WREG32(D1GRPH_SECONDARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset,
64 (u32)crtc_base);
65 WREG32(D1GRPH_PRIMARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset,
66 (u32)crtc_base);
67
68 /* Wait for update_pending to go high. */
69 while (!(RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset) & AVIVO_D1GRPH_SURFACE_UPDATE_PENDING));
70 DRM_DEBUG("Update pending now high. Unlocking vupdate_lock.\n");
71
72 /* Unlock the lock, so double-buffering can take place inside vblank */
73 tmp &= ~AVIVO_D1GRPH_UPDATE_LOCK;
74 WREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset, tmp);
75
76 /* Return current update_pending status: */
77 return RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset) & AVIVO_D1GRPH_SURFACE_UPDATE_PENDING;
78}
79
80/* get temperature in millidegrees */
81int rv770_get_temp(struct radeon_device *rdev)
82{
83 u32 temp = (RREG32(CG_MULT_THERMAL_STATUS) & ASIC_T_MASK) >>
84 ASIC_T_SHIFT;
85 int actual_temp;
86
87 if (temp & 0x400)
88 actual_temp = -256;
89 else if (temp & 0x200)
90 actual_temp = 255;
91 else if (temp & 0x100) {
92 actual_temp = temp & 0x1ff;
93 actual_temp |= ~0x1ff;
94 } else
95 actual_temp = temp & 0xff;
96
97 return (actual_temp * 1000) / 2;
98}
99
100void rv770_pm_misc(struct radeon_device *rdev)
101{
102 int req_ps_idx = rdev->pm.requested_power_state_index;
103 int req_cm_idx = rdev->pm.requested_clock_mode_index;
104 struct radeon_power_state *ps = &rdev->pm.power_state[req_ps_idx];
105 struct radeon_voltage *voltage = &ps->clock_info[req_cm_idx].voltage;
106
107 if ((voltage->type == VOLTAGE_SW) && voltage->voltage) {
108 /* 0xff01 is a flag rather then an actual voltage */
109 if (voltage->voltage == 0xff01)
110 return;
111 if (voltage->voltage != rdev->pm.current_vddc) {
112 radeon_atom_set_voltage(rdev, voltage->voltage, SET_VOLTAGE_TYPE_ASIC_VDDC);
113 rdev->pm.current_vddc = voltage->voltage;
114 DRM_DEBUG("Setting: v: %d\n", voltage->voltage);
115 }
116 }
117}
118
119/*
120 * GART
121 */
122int rv770_pcie_gart_enable(struct radeon_device *rdev)
123{
124 u32 tmp;
125 int r, i;
126
127 if (rdev->gart.table.vram.robj == NULL) {
128 dev_err(rdev->dev, "No VRAM object for PCIE GART.\n");
129 return -EINVAL;
130 }
131 r = radeon_gart_table_vram_pin(rdev);
132 if (r)
133 return r;
134 radeon_gart_restore(rdev);
135 /* Setup L2 cache */
136 WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
137 ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
138 EFFECTIVE_L2_QUEUE_SIZE(7));
139 WREG32(VM_L2_CNTL2, 0);
140 WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
141 /* Setup TLB control */
142 tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
143 SYSTEM_ACCESS_MODE_NOT_IN_SYS |
144 SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
145 EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
146 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
147 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
148 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
149 WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
150 WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
151 WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
152 WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
153 WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR, rdev->mc.gtt_start >> 12);
154 WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR, rdev->mc.gtt_end >> 12);
155 WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12);
156 WREG32(VM_CONTEXT0_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) |
157 RANGE_PROTECTION_FAULT_ENABLE_DEFAULT);
158 WREG32(VM_CONTEXT0_PROTECTION_FAULT_DEFAULT_ADDR,
159 (u32)(rdev->dummy_page.addr >> 12));
160 for (i = 1; i < 7; i++)
161 WREG32(VM_CONTEXT0_CNTL + (i * 4), 0);
162
163 r600_pcie_gart_tlb_flush(rdev);
164 rdev->gart.ready = true;
165 return 0;
166}
167
168void rv770_pcie_gart_disable(struct radeon_device *rdev)
169{
170 u32 tmp;
171 int i, r;
172
173 /* Disable all tables */
174 for (i = 0; i < 7; i++)
175 WREG32(VM_CONTEXT0_CNTL + (i * 4), 0);
176
177 /* Setup L2 cache */
178 WREG32(VM_L2_CNTL, ENABLE_L2_FRAGMENT_PROCESSING |
179 EFFECTIVE_L2_QUEUE_SIZE(7));
180 WREG32(VM_L2_CNTL2, 0);
181 WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
182 /* Setup TLB control */
183 tmp = EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
184 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
185 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
186 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
187 WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
188 WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
189 WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
190 WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
191 if (rdev->gart.table.vram.robj) {
192 r = radeon_bo_reserve(rdev->gart.table.vram.robj, false);
193 if (likely(r == 0)) {
194 radeon_bo_kunmap(rdev->gart.table.vram.robj);
195 radeon_bo_unpin(rdev->gart.table.vram.robj);
196 radeon_bo_unreserve(rdev->gart.table.vram.robj);
197 }
198 }
199}
200
201void rv770_pcie_gart_fini(struct radeon_device *rdev)
202{
203 radeon_gart_fini(rdev);
204 rv770_pcie_gart_disable(rdev);
205 radeon_gart_table_vram_free(rdev);
206}
207
208
209void rv770_agp_enable(struct radeon_device *rdev)
210{
211 u32 tmp;
212 int i;
213
214 /* Setup L2 cache */
215 WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
216 ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
217 EFFECTIVE_L2_QUEUE_SIZE(7));
218 WREG32(VM_L2_CNTL2, 0);
219 WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
220 /* Setup TLB control */
221 tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
222 SYSTEM_ACCESS_MODE_NOT_IN_SYS |
223 SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
224 EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
225 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
226 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
227 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
228 WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
229 WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
230 WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
231 WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
232 for (i = 0; i < 7; i++)
233 WREG32(VM_CONTEXT0_CNTL + (i * 4), 0);
234}
235
236static void rv770_mc_program(struct radeon_device *rdev)
237{
238 struct rv515_mc_save save;
239 u32 tmp;
240 int i, j;
241
242 /* Initialize HDP */
243 for (i = 0, j = 0; i < 32; i++, j += 0x18) {
244 WREG32((0x2c14 + j), 0x00000000);
245 WREG32((0x2c18 + j), 0x00000000);
246 WREG32((0x2c1c + j), 0x00000000);
247 WREG32((0x2c20 + j), 0x00000000);
248 WREG32((0x2c24 + j), 0x00000000);
249 }
250 /* r7xx hw bug. Read from HDP_DEBUG1 rather
251 * than writing to HDP_REG_COHERENCY_FLUSH_CNTL
252 */
253 tmp = RREG32(HDP_DEBUG1);
254
255 rv515_mc_stop(rdev, &save);
256 if (r600_mc_wait_for_idle(rdev)) {
257 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
258 }
259 /* Lockout access through VGA aperture*/
260 WREG32(VGA_HDP_CONTROL, VGA_MEMORY_DISABLE);
261 /* Update configuration */
262 if (rdev->flags & RADEON_IS_AGP) {
263 if (rdev->mc.vram_start < rdev->mc.gtt_start) {
264 /* VRAM before AGP */
265 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
266 rdev->mc.vram_start >> 12);
267 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
268 rdev->mc.gtt_end >> 12);
269 } else {
270 /* VRAM after AGP */
271 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
272 rdev->mc.gtt_start >> 12);
273 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
274 rdev->mc.vram_end >> 12);
275 }
276 } else {
277 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
278 rdev->mc.vram_start >> 12);
279 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
280 rdev->mc.vram_end >> 12);
281 }
282 WREG32(MC_VM_SYSTEM_APERTURE_DEFAULT_ADDR, 0);
283 tmp = ((rdev->mc.vram_end >> 24) & 0xFFFF) << 16;
284 tmp |= ((rdev->mc.vram_start >> 24) & 0xFFFF);
285 WREG32(MC_VM_FB_LOCATION, tmp);
286 WREG32(HDP_NONSURFACE_BASE, (rdev->mc.vram_start >> 8));
287 WREG32(HDP_NONSURFACE_INFO, (2 << 7));
288 WREG32(HDP_NONSURFACE_SIZE, 0x3FFFFFFF);
289 if (rdev->flags & RADEON_IS_AGP) {
290 WREG32(MC_VM_AGP_TOP, rdev->mc.gtt_end >> 16);
291 WREG32(MC_VM_AGP_BOT, rdev->mc.gtt_start >> 16);
292 WREG32(MC_VM_AGP_BASE, rdev->mc.agp_base >> 22);
293 } else {
294 WREG32(MC_VM_AGP_BASE, 0);
295 WREG32(MC_VM_AGP_TOP, 0x0FFFFFFF);
296 WREG32(MC_VM_AGP_BOT, 0x0FFFFFFF);
297 }
298 if (r600_mc_wait_for_idle(rdev)) {
299 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
300 }
301 rv515_mc_resume(rdev, &save);
302 /* we need to own VRAM, so turn off the VGA renderer here
303 * to stop it overwriting our objects */
304 rv515_vga_render_disable(rdev);
305}
306
307
308/*
309 * CP.
310 */
311void r700_cp_stop(struct radeon_device *rdev)
312{
313 radeon_ttm_set_active_vram_size(rdev, rdev->mc.visible_vram_size);
314 WREG32(CP_ME_CNTL, (CP_ME_HALT | CP_PFP_HALT));
315 WREG32(SCRATCH_UMSK, 0);
316}
317
318static int rv770_cp_load_microcode(struct radeon_device *rdev)
319{
320 const __be32 *fw_data;
321 int i;
322
323 if (!rdev->me_fw || !rdev->pfp_fw)
324 return -EINVAL;
325
326 r700_cp_stop(rdev);
327 WREG32(CP_RB_CNTL,
328#ifdef __BIG_ENDIAN
329 BUF_SWAP_32BIT |
330#endif
331 RB_NO_UPDATE | RB_BLKSZ(15) | RB_BUFSZ(3));
332
333 /* Reset cp */
334 WREG32(GRBM_SOFT_RESET, SOFT_RESET_CP);
335 RREG32(GRBM_SOFT_RESET);
336 mdelay(15);
337 WREG32(GRBM_SOFT_RESET, 0);
338
339 fw_data = (const __be32 *)rdev->pfp_fw->data;
340 WREG32(CP_PFP_UCODE_ADDR, 0);
341 for (i = 0; i < R700_PFP_UCODE_SIZE; i++)
342 WREG32(CP_PFP_UCODE_DATA, be32_to_cpup(fw_data++));
343 WREG32(CP_PFP_UCODE_ADDR, 0);
344
345 fw_data = (const __be32 *)rdev->me_fw->data;
346 WREG32(CP_ME_RAM_WADDR, 0);
347 for (i = 0; i < R700_PM4_UCODE_SIZE; i++)
348 WREG32(CP_ME_RAM_DATA, be32_to_cpup(fw_data++));
349
350 WREG32(CP_PFP_UCODE_ADDR, 0);
351 WREG32(CP_ME_RAM_WADDR, 0);
352 WREG32(CP_ME_RAM_RADDR, 0);
353 return 0;
354}
355
356void r700_cp_fini(struct radeon_device *rdev)
357{
358 r700_cp_stop(rdev);
359 radeon_ring_fini(rdev);
360}
361
362/*
363 * Core functions
364 */
365static u32 r700_get_tile_pipe_to_backend_map(struct radeon_device *rdev,
366 u32 num_tile_pipes,
367 u32 num_backends,
368 u32 backend_disable_mask)
369{
370 u32 backend_map = 0;
371 u32 enabled_backends_mask;
372 u32 enabled_backends_count;
373 u32 cur_pipe;
374 u32 swizzle_pipe[R7XX_MAX_PIPES];
375 u32 cur_backend;
376 u32 i;
377 bool force_no_swizzle;
378
379 if (num_tile_pipes > R7XX_MAX_PIPES)
380 num_tile_pipes = R7XX_MAX_PIPES;
381 if (num_tile_pipes < 1)
382 num_tile_pipes = 1;
383 if (num_backends > R7XX_MAX_BACKENDS)
384 num_backends = R7XX_MAX_BACKENDS;
385 if (num_backends < 1)
386 num_backends = 1;
387
388 enabled_backends_mask = 0;
389 enabled_backends_count = 0;
390 for (i = 0; i < R7XX_MAX_BACKENDS; ++i) {
391 if (((backend_disable_mask >> i) & 1) == 0) {
392 enabled_backends_mask |= (1 << i);
393 ++enabled_backends_count;
394 }
395 if (enabled_backends_count == num_backends)
396 break;
397 }
398
399 if (enabled_backends_count == 0) {
400 enabled_backends_mask = 1;
401 enabled_backends_count = 1;
402 }
403
404 if (enabled_backends_count != num_backends)
405 num_backends = enabled_backends_count;
406
407 switch (rdev->family) {
408 case CHIP_RV770:
409 case CHIP_RV730:
410 force_no_swizzle = false;
411 break;
412 case CHIP_RV710:
413 case CHIP_RV740:
414 default:
415 force_no_swizzle = true;
416 break;
417 }
418
419 memset((uint8_t *)&swizzle_pipe[0], 0, sizeof(u32) * R7XX_MAX_PIPES);
420 switch (num_tile_pipes) {
421 case 1:
422 swizzle_pipe[0] = 0;
423 break;
424 case 2:
425 swizzle_pipe[0] = 0;
426 swizzle_pipe[1] = 1;
427 break;
428 case 3:
429 if (force_no_swizzle) {
430 swizzle_pipe[0] = 0;
431 swizzle_pipe[1] = 1;
432 swizzle_pipe[2] = 2;
433 } else {
434 swizzle_pipe[0] = 0;
435 swizzle_pipe[1] = 2;
436 swizzle_pipe[2] = 1;
437 }
438 break;
439 case 4:
440 if (force_no_swizzle) {
441 swizzle_pipe[0] = 0;
442 swizzle_pipe[1] = 1;
443 swizzle_pipe[2] = 2;
444 swizzle_pipe[3] = 3;
445 } else {
446 swizzle_pipe[0] = 0;
447 swizzle_pipe[1] = 2;
448 swizzle_pipe[2] = 3;
449 swizzle_pipe[3] = 1;
450 }
451 break;
452 case 5:
453 if (force_no_swizzle) {
454 swizzle_pipe[0] = 0;
455 swizzle_pipe[1] = 1;
456 swizzle_pipe[2] = 2;
457 swizzle_pipe[3] = 3;
458 swizzle_pipe[4] = 4;
459 } else {
460 swizzle_pipe[0] = 0;
461 swizzle_pipe[1] = 2;
462 swizzle_pipe[2] = 4;
463 swizzle_pipe[3] = 1;
464 swizzle_pipe[4] = 3;
465 }
466 break;
467 case 6:
468 if (force_no_swizzle) {
469 swizzle_pipe[0] = 0;
470 swizzle_pipe[1] = 1;
471 swizzle_pipe[2] = 2;
472 swizzle_pipe[3] = 3;
473 swizzle_pipe[4] = 4;
474 swizzle_pipe[5] = 5;
475 } else {
476 swizzle_pipe[0] = 0;
477 swizzle_pipe[1] = 2;
478 swizzle_pipe[2] = 4;
479 swizzle_pipe[3] = 5;
480 swizzle_pipe[4] = 3;
481 swizzle_pipe[5] = 1;
482 }
483 break;
484 case 7:
485 if (force_no_swizzle) {
486 swizzle_pipe[0] = 0;
487 swizzle_pipe[1] = 1;
488 swizzle_pipe[2] = 2;
489 swizzle_pipe[3] = 3;
490 swizzle_pipe[4] = 4;
491 swizzle_pipe[5] = 5;
492 swizzle_pipe[6] = 6;
493 } else {
494 swizzle_pipe[0] = 0;
495 swizzle_pipe[1] = 2;
496 swizzle_pipe[2] = 4;
497 swizzle_pipe[3] = 6;
498 swizzle_pipe[4] = 3;
499 swizzle_pipe[5] = 1;
500 swizzle_pipe[6] = 5;
501 }
502 break;
503 case 8:
504 if (force_no_swizzle) {
505 swizzle_pipe[0] = 0;
506 swizzle_pipe[1] = 1;
507 swizzle_pipe[2] = 2;
508 swizzle_pipe[3] = 3;
509 swizzle_pipe[4] = 4;
510 swizzle_pipe[5] = 5;
511 swizzle_pipe[6] = 6;
512 swizzle_pipe[7] = 7;
513 } else {
514 swizzle_pipe[0] = 0;
515 swizzle_pipe[1] = 2;
516 swizzle_pipe[2] = 4;
517 swizzle_pipe[3] = 6;
518 swizzle_pipe[4] = 3;
519 swizzle_pipe[5] = 1;
520 swizzle_pipe[6] = 7;
521 swizzle_pipe[7] = 5;
522 }
523 break;
524 }
525
526 cur_backend = 0;
527 for (cur_pipe = 0; cur_pipe < num_tile_pipes; ++cur_pipe) {
528 while (((1 << cur_backend) & enabled_backends_mask) == 0)
529 cur_backend = (cur_backend + 1) % R7XX_MAX_BACKENDS;
530
531 backend_map |= (u32)(((cur_backend & 3) << (swizzle_pipe[cur_pipe] * 2)));
532
533 cur_backend = (cur_backend + 1) % R7XX_MAX_BACKENDS;
534 }
535
536 return backend_map;
537}
538
539static void rv770_gpu_init(struct radeon_device *rdev)
540{
541 int i, j, num_qd_pipes;
542 u32 ta_aux_cntl;
543 u32 sx_debug_1;
544 u32 smx_dc_ctl0;
545 u32 db_debug3;
546 u32 num_gs_verts_per_thread;
547 u32 vgt_gs_per_es;
548 u32 gs_prim_buffer_depth = 0;
549 u32 sq_ms_fifo_sizes;
550 u32 sq_config;
551 u32 sq_thread_resource_mgmt;
552 u32 hdp_host_path_cntl;
553 u32 sq_dyn_gpr_size_simd_ab_0;
554 u32 backend_map;
555 u32 gb_tiling_config = 0;
556 u32 cc_rb_backend_disable = 0;
557 u32 cc_gc_shader_pipe_config = 0;
558 u32 mc_arb_ramcfg;
559 u32 db_debug4;
560
561 /* setup chip specs */
562 switch (rdev->family) {
563 case CHIP_RV770:
564 rdev->config.rv770.max_pipes = 4;
565 rdev->config.rv770.max_tile_pipes = 8;
566 rdev->config.rv770.max_simds = 10;
567 rdev->config.rv770.max_backends = 4;
568 rdev->config.rv770.max_gprs = 256;
569 rdev->config.rv770.max_threads = 248;
570 rdev->config.rv770.max_stack_entries = 512;
571 rdev->config.rv770.max_hw_contexts = 8;
572 rdev->config.rv770.max_gs_threads = 16 * 2;
573 rdev->config.rv770.sx_max_export_size = 128;
574 rdev->config.rv770.sx_max_export_pos_size = 16;
575 rdev->config.rv770.sx_max_export_smx_size = 112;
576 rdev->config.rv770.sq_num_cf_insts = 2;
577
578 rdev->config.rv770.sx_num_of_sets = 7;
579 rdev->config.rv770.sc_prim_fifo_size = 0xF9;
580 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
581 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
582 break;
583 case CHIP_RV730:
584 rdev->config.rv770.max_pipes = 2;
585 rdev->config.rv770.max_tile_pipes = 4;
586 rdev->config.rv770.max_simds = 8;
587 rdev->config.rv770.max_backends = 2;
588 rdev->config.rv770.max_gprs = 128;
589 rdev->config.rv770.max_threads = 248;
590 rdev->config.rv770.max_stack_entries = 256;
591 rdev->config.rv770.max_hw_contexts = 8;
592 rdev->config.rv770.max_gs_threads = 16 * 2;
593 rdev->config.rv770.sx_max_export_size = 256;
594 rdev->config.rv770.sx_max_export_pos_size = 32;
595 rdev->config.rv770.sx_max_export_smx_size = 224;
596 rdev->config.rv770.sq_num_cf_insts = 2;
597
598 rdev->config.rv770.sx_num_of_sets = 7;
599 rdev->config.rv770.sc_prim_fifo_size = 0xf9;
600 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
601 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
602 if (rdev->config.rv770.sx_max_export_pos_size > 16) {
603 rdev->config.rv770.sx_max_export_pos_size -= 16;
604 rdev->config.rv770.sx_max_export_smx_size += 16;
605 }
606 break;
607 case CHIP_RV710:
608 rdev->config.rv770.max_pipes = 2;
609 rdev->config.rv770.max_tile_pipes = 2;
610 rdev->config.rv770.max_simds = 2;
611 rdev->config.rv770.max_backends = 1;
612 rdev->config.rv770.max_gprs = 256;
613 rdev->config.rv770.max_threads = 192;
614 rdev->config.rv770.max_stack_entries = 256;
615 rdev->config.rv770.max_hw_contexts = 4;
616 rdev->config.rv770.max_gs_threads = 8 * 2;
617 rdev->config.rv770.sx_max_export_size = 128;
618 rdev->config.rv770.sx_max_export_pos_size = 16;
619 rdev->config.rv770.sx_max_export_smx_size = 112;
620 rdev->config.rv770.sq_num_cf_insts = 1;
621
622 rdev->config.rv770.sx_num_of_sets = 7;
623 rdev->config.rv770.sc_prim_fifo_size = 0x40;
624 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
625 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
626 break;
627 case CHIP_RV740:
628 rdev->config.rv770.max_pipes = 4;
629 rdev->config.rv770.max_tile_pipes = 4;
630 rdev->config.rv770.max_simds = 8;
631 rdev->config.rv770.max_backends = 4;
632 rdev->config.rv770.max_gprs = 256;
633 rdev->config.rv770.max_threads = 248;
634 rdev->config.rv770.max_stack_entries = 512;
635 rdev->config.rv770.max_hw_contexts = 8;
636 rdev->config.rv770.max_gs_threads = 16 * 2;
637 rdev->config.rv770.sx_max_export_size = 256;
638 rdev->config.rv770.sx_max_export_pos_size = 32;
639 rdev->config.rv770.sx_max_export_smx_size = 224;
640 rdev->config.rv770.sq_num_cf_insts = 2;
641
642 rdev->config.rv770.sx_num_of_sets = 7;
643 rdev->config.rv770.sc_prim_fifo_size = 0x100;
644 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
645 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
646
647 if (rdev->config.rv770.sx_max_export_pos_size > 16) {
648 rdev->config.rv770.sx_max_export_pos_size -= 16;
649 rdev->config.rv770.sx_max_export_smx_size += 16;
650 }
651 break;
652 default:
653 break;
654 }
655
656 /* Initialize HDP */
657 j = 0;
658 for (i = 0; i < 32; i++) {
659 WREG32((0x2c14 + j), 0x00000000);
660 WREG32((0x2c18 + j), 0x00000000);
661 WREG32((0x2c1c + j), 0x00000000);
662 WREG32((0x2c20 + j), 0x00000000);
663 WREG32((0x2c24 + j), 0x00000000);
664 j += 0x18;
665 }
666
667 WREG32(GRBM_CNTL, GRBM_READ_TIMEOUT(0xff));
668
669 /* setup tiling, simd, pipe config */
670 mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG);
671
672 switch (rdev->config.rv770.max_tile_pipes) {
673 case 1:
674 default:
675 gb_tiling_config |= PIPE_TILING(0);
676 break;
677 case 2:
678 gb_tiling_config |= PIPE_TILING(1);
679 break;
680 case 4:
681 gb_tiling_config |= PIPE_TILING(2);
682 break;
683 case 8:
684 gb_tiling_config |= PIPE_TILING(3);
685 break;
686 }
687 rdev->config.rv770.tiling_npipes = rdev->config.rv770.max_tile_pipes;
688
689 if (rdev->family == CHIP_RV770)
690 gb_tiling_config |= BANK_TILING(1);
691 else
692 gb_tiling_config |= BANK_TILING((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT);
693 rdev->config.rv770.tiling_nbanks = 4 << ((gb_tiling_config >> 4) & 0x3);
694 gb_tiling_config |= GROUP_SIZE((mc_arb_ramcfg & BURSTLENGTH_MASK) >> BURSTLENGTH_SHIFT);
695 if ((mc_arb_ramcfg & BURSTLENGTH_MASK) >> BURSTLENGTH_SHIFT)
696 rdev->config.rv770.tiling_group_size = 512;
697 else
698 rdev->config.rv770.tiling_group_size = 256;
699 if (((mc_arb_ramcfg & NOOFROWS_MASK) >> NOOFROWS_SHIFT) > 3) {
700 gb_tiling_config |= ROW_TILING(3);
701 gb_tiling_config |= SAMPLE_SPLIT(3);
702 } else {
703 gb_tiling_config |=
704 ROW_TILING(((mc_arb_ramcfg & NOOFROWS_MASK) >> NOOFROWS_SHIFT));
705 gb_tiling_config |=
706 SAMPLE_SPLIT(((mc_arb_ramcfg & NOOFROWS_MASK) >> NOOFROWS_SHIFT));
707 }
708
709 gb_tiling_config |= BANK_SWAPS(1);
710
711 cc_rb_backend_disable = RREG32(CC_RB_BACKEND_DISABLE) & 0x00ff0000;
712 cc_rb_backend_disable |=
713 BACKEND_DISABLE((R7XX_MAX_BACKENDS_MASK << rdev->config.rv770.max_backends) & R7XX_MAX_BACKENDS_MASK);
714
715 cc_gc_shader_pipe_config = RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffffff00;
716 cc_gc_shader_pipe_config |=
717 INACTIVE_QD_PIPES((R7XX_MAX_PIPES_MASK << rdev->config.rv770.max_pipes) & R7XX_MAX_PIPES_MASK);
718 cc_gc_shader_pipe_config |=
719 INACTIVE_SIMDS((R7XX_MAX_SIMDS_MASK << rdev->config.rv770.max_simds) & R7XX_MAX_SIMDS_MASK);
720
721 if (rdev->family == CHIP_RV740)
722 backend_map = 0x28;
723 else
724 backend_map = r700_get_tile_pipe_to_backend_map(rdev,
725 rdev->config.rv770.max_tile_pipes,
726 (R7XX_MAX_BACKENDS -
727 r600_count_pipe_bits((cc_rb_backend_disable &
728 R7XX_MAX_BACKENDS_MASK) >> 16)),
729 (cc_rb_backend_disable >> 16));
730
731 rdev->config.rv770.tile_config = gb_tiling_config;
732 rdev->config.rv770.backend_map = backend_map;
733 gb_tiling_config |= BACKEND_MAP(backend_map);
734
735 WREG32(GB_TILING_CONFIG, gb_tiling_config);
736 WREG32(DCP_TILING_CONFIG, (gb_tiling_config & 0xffff));
737 WREG32(HDP_TILING_CONFIG, (gb_tiling_config & 0xffff));
738
739 WREG32(CC_RB_BACKEND_DISABLE, cc_rb_backend_disable);
740 WREG32(CC_GC_SHADER_PIPE_CONFIG, cc_gc_shader_pipe_config);
741 WREG32(GC_USER_SHADER_PIPE_CONFIG, cc_gc_shader_pipe_config);
742 WREG32(CC_SYS_RB_BACKEND_DISABLE, cc_rb_backend_disable);
743
744 WREG32(CGTS_SYS_TCC_DISABLE, 0);
745 WREG32(CGTS_TCC_DISABLE, 0);
746 WREG32(CGTS_USER_SYS_TCC_DISABLE, 0);
747 WREG32(CGTS_USER_TCC_DISABLE, 0);
748
749 num_qd_pipes =
750 R7XX_MAX_PIPES - r600_count_pipe_bits((cc_gc_shader_pipe_config & INACTIVE_QD_PIPES_MASK) >> 8);
751 WREG32(VGT_OUT_DEALLOC_CNTL, (num_qd_pipes * 4) & DEALLOC_DIST_MASK);
752 WREG32(VGT_VERTEX_REUSE_BLOCK_CNTL, ((num_qd_pipes * 4) - 2) & VTX_REUSE_DEPTH_MASK);
753
754 /* set HW defaults for 3D engine */
755 WREG32(CP_QUEUE_THRESHOLDS, (ROQ_IB1_START(0x16) |
756 ROQ_IB2_START(0x2b)));
757
758 WREG32(CP_MEQ_THRESHOLDS, STQ_SPLIT(0x30));
759
760 ta_aux_cntl = RREG32(TA_CNTL_AUX);
761 WREG32(TA_CNTL_AUX, ta_aux_cntl | DISABLE_CUBE_ANISO);
762
763 sx_debug_1 = RREG32(SX_DEBUG_1);
764 sx_debug_1 |= ENABLE_NEW_SMX_ADDRESS;
765 WREG32(SX_DEBUG_1, sx_debug_1);
766
767 smx_dc_ctl0 = RREG32(SMX_DC_CTL0);
768 smx_dc_ctl0 &= ~CACHE_DEPTH(0x1ff);
769 smx_dc_ctl0 |= CACHE_DEPTH((rdev->config.rv770.sx_num_of_sets * 64) - 1);
770 WREG32(SMX_DC_CTL0, smx_dc_ctl0);
771
772 if (rdev->family != CHIP_RV740)
773 WREG32(SMX_EVENT_CTL, (ES_FLUSH_CTL(4) |
774 GS_FLUSH_CTL(4) |
775 ACK_FLUSH_CTL(3) |
776 SYNC_FLUSH_CTL));
777
778 db_debug3 = RREG32(DB_DEBUG3);
779 db_debug3 &= ~DB_CLK_OFF_DELAY(0x1f);
780 switch (rdev->family) {
781 case CHIP_RV770:
782 case CHIP_RV740:
783 db_debug3 |= DB_CLK_OFF_DELAY(0x1f);
784 break;
785 case CHIP_RV710:
786 case CHIP_RV730:
787 default:
788 db_debug3 |= DB_CLK_OFF_DELAY(2);
789 break;
790 }
791 WREG32(DB_DEBUG3, db_debug3);
792
793 if (rdev->family != CHIP_RV770) {
794 db_debug4 = RREG32(DB_DEBUG4);
795 db_debug4 |= DISABLE_TILE_COVERED_FOR_PS_ITER;
796 WREG32(DB_DEBUG4, db_debug4);
797 }
798
799 WREG32(SX_EXPORT_BUFFER_SIZES, (COLOR_BUFFER_SIZE((rdev->config.rv770.sx_max_export_size / 4) - 1) |
800 POSITION_BUFFER_SIZE((rdev->config.rv770.sx_max_export_pos_size / 4) - 1) |
801 SMX_BUFFER_SIZE((rdev->config.rv770.sx_max_export_smx_size / 4) - 1)));
802
803 WREG32(PA_SC_FIFO_SIZE, (SC_PRIM_FIFO_SIZE(rdev->config.rv770.sc_prim_fifo_size) |
804 SC_HIZ_TILE_FIFO_SIZE(rdev->config.rv770.sc_hiz_tile_fifo_size) |
805 SC_EARLYZ_TILE_FIFO_SIZE(rdev->config.rv770.sc_earlyz_tile_fifo_fize)));
806
807 WREG32(PA_SC_MULTI_CHIP_CNTL, 0);
808
809 WREG32(VGT_NUM_INSTANCES, 1);
810
811 WREG32(SPI_CONFIG_CNTL, GPR_WRITE_PRIORITY(0));
812
813 WREG32(SPI_CONFIG_CNTL_1, VTX_DONE_DELAY(4));
814
815 WREG32(CP_PERFMON_CNTL, 0);
816
817 sq_ms_fifo_sizes = (CACHE_FIFO_SIZE(16 * rdev->config.rv770.sq_num_cf_insts) |
818 DONE_FIFO_HIWATER(0xe0) |
819 ALU_UPDATE_FIFO_HIWATER(0x8));
820 switch (rdev->family) {
821 case CHIP_RV770:
822 case CHIP_RV730:
823 case CHIP_RV710:
824 sq_ms_fifo_sizes |= FETCH_FIFO_HIWATER(0x1);
825 break;
826 case CHIP_RV740:
827 default:
828 sq_ms_fifo_sizes |= FETCH_FIFO_HIWATER(0x4);
829 break;
830 }
831 WREG32(SQ_MS_FIFO_SIZES, sq_ms_fifo_sizes);
832
833 /* SQ_CONFIG, SQ_GPR_RESOURCE_MGMT, SQ_THREAD_RESOURCE_MGMT, SQ_STACK_RESOURCE_MGMT
834 * should be adjusted as needed by the 2D/3D drivers. This just sets default values
835 */
836 sq_config = RREG32(SQ_CONFIG);
837 sq_config &= ~(PS_PRIO(3) |
838 VS_PRIO(3) |
839 GS_PRIO(3) |
840 ES_PRIO(3));
841 sq_config |= (DX9_CONSTS |
842 VC_ENABLE |
843 EXPORT_SRC_C |
844 PS_PRIO(0) |
845 VS_PRIO(1) |
846 GS_PRIO(2) |
847 ES_PRIO(3));
848 if (rdev->family == CHIP_RV710)
849 /* no vertex cache */
850 sq_config &= ~VC_ENABLE;
851
852 WREG32(SQ_CONFIG, sq_config);
853
854 WREG32(SQ_GPR_RESOURCE_MGMT_1, (NUM_PS_GPRS((rdev->config.rv770.max_gprs * 24)/64) |
855 NUM_VS_GPRS((rdev->config.rv770.max_gprs * 24)/64) |
856 NUM_CLAUSE_TEMP_GPRS(((rdev->config.rv770.max_gprs * 24)/64)/2)));
857
858 WREG32(SQ_GPR_RESOURCE_MGMT_2, (NUM_GS_GPRS((rdev->config.rv770.max_gprs * 7)/64) |
859 NUM_ES_GPRS((rdev->config.rv770.max_gprs * 7)/64)));
860
861 sq_thread_resource_mgmt = (NUM_PS_THREADS((rdev->config.rv770.max_threads * 4)/8) |
862 NUM_VS_THREADS((rdev->config.rv770.max_threads * 2)/8) |
863 NUM_ES_THREADS((rdev->config.rv770.max_threads * 1)/8));
864 if (((rdev->config.rv770.max_threads * 1) / 8) > rdev->config.rv770.max_gs_threads)
865 sq_thread_resource_mgmt |= NUM_GS_THREADS(rdev->config.rv770.max_gs_threads);
866 else
867 sq_thread_resource_mgmt |= NUM_GS_THREADS((rdev->config.rv770.max_gs_threads * 1)/8);
868 WREG32(SQ_THREAD_RESOURCE_MGMT, sq_thread_resource_mgmt);
869
870 WREG32(SQ_STACK_RESOURCE_MGMT_1, (NUM_PS_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4) |
871 NUM_VS_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4)));
872
873 WREG32(SQ_STACK_RESOURCE_MGMT_2, (NUM_GS_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4) |
874 NUM_ES_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4)));
875
876 sq_dyn_gpr_size_simd_ab_0 = (SIMDA_RING0((rdev->config.rv770.max_gprs * 38)/64) |
877 SIMDA_RING1((rdev->config.rv770.max_gprs * 38)/64) |
878 SIMDB_RING0((rdev->config.rv770.max_gprs * 38)/64) |
879 SIMDB_RING1((rdev->config.rv770.max_gprs * 38)/64));
880
881 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_0, sq_dyn_gpr_size_simd_ab_0);
882 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_1, sq_dyn_gpr_size_simd_ab_0);
883 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_2, sq_dyn_gpr_size_simd_ab_0);
884 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_3, sq_dyn_gpr_size_simd_ab_0);
885 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_4, sq_dyn_gpr_size_simd_ab_0);
886 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_5, sq_dyn_gpr_size_simd_ab_0);
887 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_6, sq_dyn_gpr_size_simd_ab_0);
888 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_7, sq_dyn_gpr_size_simd_ab_0);
889
890 WREG32(PA_SC_FORCE_EOV_MAX_CNTS, (FORCE_EOV_MAX_CLK_CNT(4095) |
891 FORCE_EOV_MAX_REZ_CNT(255)));
892
893 if (rdev->family == CHIP_RV710)
894 WREG32(VGT_CACHE_INVALIDATION, (CACHE_INVALIDATION(TC_ONLY) |
895 AUTO_INVLD_EN(ES_AND_GS_AUTO)));
896 else
897 WREG32(VGT_CACHE_INVALIDATION, (CACHE_INVALIDATION(VC_AND_TC) |
898 AUTO_INVLD_EN(ES_AND_GS_AUTO)));
899
900 switch (rdev->family) {
901 case CHIP_RV770:
902 case CHIP_RV730:
903 case CHIP_RV740:
904 gs_prim_buffer_depth = 384;
905 break;
906 case CHIP_RV710:
907 gs_prim_buffer_depth = 128;
908 break;
909 default:
910 break;
911 }
912
913 num_gs_verts_per_thread = rdev->config.rv770.max_pipes * 16;
914 vgt_gs_per_es = gs_prim_buffer_depth + num_gs_verts_per_thread;
915 /* Max value for this is 256 */
916 if (vgt_gs_per_es > 256)
917 vgt_gs_per_es = 256;
918
919 WREG32(VGT_ES_PER_GS, 128);
920 WREG32(VGT_GS_PER_ES, vgt_gs_per_es);
921 WREG32(VGT_GS_PER_VS, 2);
922
923 /* more default values. 2D/3D driver should adjust as needed */
924 WREG32(VGT_GS_VERTEX_REUSE, 16);
925 WREG32(PA_SC_LINE_STIPPLE_STATE, 0);
926 WREG32(VGT_STRMOUT_EN, 0);
927 WREG32(SX_MISC, 0);
928 WREG32(PA_SC_MODE_CNTL, 0);
929 WREG32(PA_SC_EDGERULE, 0xaaaaaaaa);
930 WREG32(PA_SC_AA_CONFIG, 0);
931 WREG32(PA_SC_CLIPRECT_RULE, 0xffff);
932 WREG32(PA_SC_LINE_STIPPLE, 0);
933 WREG32(SPI_INPUT_Z, 0);
934 WREG32(SPI_PS_IN_CONTROL_0, NUM_INTERP(2));
935 WREG32(CB_COLOR7_FRAG, 0);
936
937 /* clear render buffer base addresses */
938 WREG32(CB_COLOR0_BASE, 0);
939 WREG32(CB_COLOR1_BASE, 0);
940 WREG32(CB_COLOR2_BASE, 0);
941 WREG32(CB_COLOR3_BASE, 0);
942 WREG32(CB_COLOR4_BASE, 0);
943 WREG32(CB_COLOR5_BASE, 0);
944 WREG32(CB_COLOR6_BASE, 0);
945 WREG32(CB_COLOR7_BASE, 0);
946
947 WREG32(TCP_CNTL, 0);
948
949 hdp_host_path_cntl = RREG32(HDP_HOST_PATH_CNTL);
950 WREG32(HDP_HOST_PATH_CNTL, hdp_host_path_cntl);
951
952 WREG32(PA_SC_MULTI_CHIP_CNTL, 0);
953
954 WREG32(PA_CL_ENHANCE, (CLIP_VTX_REORDER_ENA |
955 NUM_CLIP_SEQ(3)));
956
957}
958
959static int rv770_vram_scratch_init(struct radeon_device *rdev)
960{
961 int r;
962 u64 gpu_addr;
963
964 if (rdev->vram_scratch.robj == NULL) {
965 r = radeon_bo_create(rdev, RADEON_GPU_PAGE_SIZE,
966 PAGE_SIZE, true, RADEON_GEM_DOMAIN_VRAM,
967 &rdev->vram_scratch.robj);
968 if (r) {
969 return r;
970 }
971 }
972
973 r = radeon_bo_reserve(rdev->vram_scratch.robj, false);
974 if (unlikely(r != 0))
975 return r;
976 r = radeon_bo_pin(rdev->vram_scratch.robj,
977 RADEON_GEM_DOMAIN_VRAM, &gpu_addr);
978 if (r) {
979 radeon_bo_unreserve(rdev->vram_scratch.robj);
980 return r;
981 }
982 r = radeon_bo_kmap(rdev->vram_scratch.robj,
983 (void **)&rdev->vram_scratch.ptr);
984 if (r)
985 radeon_bo_unpin(rdev->vram_scratch.robj);
986 radeon_bo_unreserve(rdev->vram_scratch.robj);
987
988 return r;
989}
990
991static void rv770_vram_scratch_fini(struct radeon_device *rdev)
992{
993 int r;
994
995 if (rdev->vram_scratch.robj == NULL) {
996 return;
997 }
998 r = radeon_bo_reserve(rdev->vram_scratch.robj, false);
999 if (likely(r == 0)) {
1000 radeon_bo_kunmap(rdev->vram_scratch.robj);
1001 radeon_bo_unpin(rdev->vram_scratch.robj);
1002 radeon_bo_unreserve(rdev->vram_scratch.robj);
1003 }
1004 radeon_bo_unref(&rdev->vram_scratch.robj);
1005}
1006
1007void r700_vram_gtt_location(struct radeon_device *rdev, struct radeon_mc *mc)
1008{
1009 u64 size_bf, size_af;
1010
1011 if (mc->mc_vram_size > 0xE0000000) {
1012 /* leave room for at least 512M GTT */
1013 dev_warn(rdev->dev, "limiting VRAM\n");
1014 mc->real_vram_size = 0xE0000000;
1015 mc->mc_vram_size = 0xE0000000;
1016 }
1017 if (rdev->flags & RADEON_IS_AGP) {
1018 size_bf = mc->gtt_start;
1019 size_af = 0xFFFFFFFF - mc->gtt_end + 1;
1020 if (size_bf > size_af) {
1021 if (mc->mc_vram_size > size_bf) {
1022 dev_warn(rdev->dev, "limiting VRAM\n");
1023 mc->real_vram_size = size_bf;
1024 mc->mc_vram_size = size_bf;
1025 }
1026 mc->vram_start = mc->gtt_start - mc->mc_vram_size;
1027 } else {
1028 if (mc->mc_vram_size > size_af) {
1029 dev_warn(rdev->dev, "limiting VRAM\n");
1030 mc->real_vram_size = size_af;
1031 mc->mc_vram_size = size_af;
1032 }
1033 mc->vram_start = mc->gtt_end;
1034 }
1035 mc->vram_end = mc->vram_start + mc->mc_vram_size - 1;
1036 dev_info(rdev->dev, "VRAM: %lluM 0x%08llX - 0x%08llX (%lluM used)\n",
1037 mc->mc_vram_size >> 20, mc->vram_start,
1038 mc->vram_end, mc->real_vram_size >> 20);
1039 } else {
1040 radeon_vram_location(rdev, &rdev->mc, 0);
1041 rdev->mc.gtt_base_align = 0;
1042 radeon_gtt_location(rdev, mc);
1043 }
1044}
1045
1046int rv770_mc_init(struct radeon_device *rdev)
1047{
1048 u32 tmp;
1049 int chansize, numchan;
1050
1051 /* Get VRAM informations */
1052 rdev->mc.vram_is_ddr = true;
1053 tmp = RREG32(MC_ARB_RAMCFG);
1054 if (tmp & CHANSIZE_OVERRIDE) {
1055 chansize = 16;
1056 } else if (tmp & CHANSIZE_MASK) {
1057 chansize = 64;
1058 } else {
1059 chansize = 32;
1060 }
1061 tmp = RREG32(MC_SHARED_CHMAP);
1062 switch ((tmp & NOOFCHAN_MASK) >> NOOFCHAN_SHIFT) {
1063 case 0:
1064 default:
1065 numchan = 1;
1066 break;
1067 case 1:
1068 numchan = 2;
1069 break;
1070 case 2:
1071 numchan = 4;
1072 break;
1073 case 3:
1074 numchan = 8;
1075 break;
1076 }
1077 rdev->mc.vram_width = numchan * chansize;
1078 /* Could aper size report 0 ? */
1079 rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0);
1080 rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0);
1081 /* Setup GPU memory space */
1082 rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE);
1083 rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE);
1084 rdev->mc.visible_vram_size = rdev->mc.aper_size;
1085 r700_vram_gtt_location(rdev, &rdev->mc);
1086 radeon_update_bandwidth_info(rdev);
1087
1088 return 0;
1089}
1090
1091static int rv770_startup(struct radeon_device *rdev)
1092{
1093 int r;
1094
1095 /* enable pcie gen2 link */
1096 rv770_pcie_gen2_enable(rdev);
1097
1098 if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) {
1099 r = r600_init_microcode(rdev);
1100 if (r) {
1101 DRM_ERROR("Failed to load firmware!\n");
1102 return r;
1103 }
1104 }
1105
1106 rv770_mc_program(rdev);
1107 if (rdev->flags & RADEON_IS_AGP) {
1108 rv770_agp_enable(rdev);
1109 } else {
1110 r = rv770_pcie_gart_enable(rdev);
1111 if (r)
1112 return r;
1113 }
1114 r = rv770_vram_scratch_init(rdev);
1115 if (r)
1116 return r;
1117 rv770_gpu_init(rdev);
1118 r = r600_blit_init(rdev);
1119 if (r) {
1120 r600_blit_fini(rdev);
1121 rdev->asic->copy = NULL;
1122 dev_warn(rdev->dev, "failed blitter (%d) falling back to memcpy\n", r);
1123 }
1124
1125 /* allocate wb buffer */
1126 r = radeon_wb_init(rdev);
1127 if (r)
1128 return r;
1129
1130 /* Enable IRQ */
1131 r = r600_irq_init(rdev);
1132 if (r) {
1133 DRM_ERROR("radeon: IH init failed (%d).\n", r);
1134 radeon_irq_kms_fini(rdev);
1135 return r;
1136 }
1137 r600_irq_set(rdev);
1138
1139 r = radeon_ring_init(rdev, rdev->cp.ring_size);
1140 if (r)
1141 return r;
1142 r = rv770_cp_load_microcode(rdev);
1143 if (r)
1144 return r;
1145 r = r600_cp_resume(rdev);
1146 if (r)
1147 return r;
1148
1149 return 0;
1150}
1151
1152int rv770_resume(struct radeon_device *rdev)
1153{
1154 int r;
1155
1156 /* Do not reset GPU before posting, on rv770 hw unlike on r500 hw,
1157 * posting will perform necessary task to bring back GPU into good
1158 * shape.
1159 */
1160 /* post card */
1161 atom_asic_init(rdev->mode_info.atom_context);
1162
1163 r = rv770_startup(rdev);
1164 if (r) {
1165 DRM_ERROR("r600 startup failed on resume\n");
1166 return r;
1167 }
1168
1169 r = r600_ib_test(rdev);
1170 if (r) {
1171 DRM_ERROR("radeon: failed testing IB (%d).\n", r);
1172 return r;
1173 }
1174
1175 r = r600_audio_init(rdev);
1176 if (r) {
1177 dev_err(rdev->dev, "radeon: audio init failed\n");
1178 return r;
1179 }
1180
1181 return r;
1182
1183}
1184
1185int rv770_suspend(struct radeon_device *rdev)
1186{
1187 int r;
1188
1189 r600_audio_fini(rdev);
1190 /* FIXME: we should wait for ring to be empty */
1191 r700_cp_stop(rdev);
1192 rdev->cp.ready = false;
1193 r600_irq_suspend(rdev);
1194 radeon_wb_disable(rdev);
1195 rv770_pcie_gart_disable(rdev);
1196 /* unpin shaders bo */
1197 if (rdev->r600_blit.shader_obj) {
1198 r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false);
1199 if (likely(r == 0)) {
1200 radeon_bo_unpin(rdev->r600_blit.shader_obj);
1201 radeon_bo_unreserve(rdev->r600_blit.shader_obj);
1202 }
1203 }
1204 return 0;
1205}
1206
1207/* Plan is to move initialization in that function and use
1208 * helper function so that radeon_device_init pretty much
1209 * do nothing more than calling asic specific function. This
1210 * should also allow to remove a bunch of callback function
1211 * like vram_info.
1212 */
1213int rv770_init(struct radeon_device *rdev)
1214{
1215 int r;
1216
1217 /* This don't do much */
1218 r = radeon_gem_init(rdev);
1219 if (r)
1220 return r;
1221 /* Read BIOS */
1222 if (!radeon_get_bios(rdev)) {
1223 if (ASIC_IS_AVIVO(rdev))
1224 return -EINVAL;
1225 }
1226 /* Must be an ATOMBIOS */
1227 if (!rdev->is_atom_bios) {
1228 dev_err(rdev->dev, "Expecting atombios for R600 GPU\n");
1229 return -EINVAL;
1230 }
1231 r = radeon_atombios_init(rdev);
1232 if (r)
1233 return r;
1234 /* Post card if necessary */
1235 if (!radeon_card_posted(rdev)) {
1236 if (!rdev->bios) {
1237 dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n");
1238 return -EINVAL;
1239 }
1240 DRM_INFO("GPU not posted. posting now...\n");
1241 atom_asic_init(rdev->mode_info.atom_context);
1242 }
1243 /* Initialize scratch registers */
1244 r600_scratch_init(rdev);
1245 /* Initialize surface registers */
1246 radeon_surface_init(rdev);
1247 /* Initialize clocks */
1248 radeon_get_clock_info(rdev->ddev);
1249 /* Fence driver */
1250 r = radeon_fence_driver_init(rdev);
1251 if (r)
1252 return r;
1253 /* initialize AGP */
1254 if (rdev->flags & RADEON_IS_AGP) {
1255 r = radeon_agp_init(rdev);
1256 if (r)
1257 radeon_agp_disable(rdev);
1258 }
1259 r = rv770_mc_init(rdev);
1260 if (r)
1261 return r;
1262 /* Memory manager */
1263 r = radeon_bo_init(rdev);
1264 if (r)
1265 return r;
1266
1267 r = radeon_irq_kms_init(rdev);
1268 if (r)
1269 return r;
1270
1271 rdev->cp.ring_obj = NULL;
1272 r600_ring_init(rdev, 1024 * 1024);
1273
1274 rdev->ih.ring_obj = NULL;
1275 r600_ih_ring_init(rdev, 64 * 1024);
1276
1277 r = r600_pcie_gart_init(rdev);
1278 if (r)
1279 return r;
1280
1281 rdev->accel_working = true;
1282 r = rv770_startup(rdev);
1283 if (r) {
1284 dev_err(rdev->dev, "disabling GPU acceleration\n");
1285 r700_cp_fini(rdev);
1286 r600_irq_fini(rdev);
1287 radeon_wb_fini(rdev);
1288 radeon_irq_kms_fini(rdev);
1289 rv770_pcie_gart_fini(rdev);
1290 rdev->accel_working = false;
1291 }
1292 if (rdev->accel_working) {
1293 r = radeon_ib_pool_init(rdev);
1294 if (r) {
1295 dev_err(rdev->dev, "IB initialization failed (%d).\n", r);
1296 rdev->accel_working = false;
1297 } else {
1298 r = r600_ib_test(rdev);
1299 if (r) {
1300 dev_err(rdev->dev, "IB test failed (%d).\n", r);
1301 rdev->accel_working = false;
1302 }
1303 }
1304 }
1305
1306 r = r600_audio_init(rdev);
1307 if (r) {
1308 dev_err(rdev->dev, "radeon: audio init failed\n");
1309 return r;
1310 }
1311
1312 return 0;
1313}
1314
1315void rv770_fini(struct radeon_device *rdev)
1316{
1317 r600_blit_fini(rdev);
1318 r700_cp_fini(rdev);
1319 r600_irq_fini(rdev);
1320 radeon_wb_fini(rdev);
1321 radeon_ib_pool_fini(rdev);
1322 radeon_irq_kms_fini(rdev);
1323 rv770_pcie_gart_fini(rdev);
1324 rv770_vram_scratch_fini(rdev);
1325 radeon_gem_fini(rdev);
1326 radeon_fence_driver_fini(rdev);
1327 radeon_agp_fini(rdev);
1328 radeon_bo_fini(rdev);
1329 radeon_atombios_fini(rdev);
1330 kfree(rdev->bios);
1331 rdev->bios = NULL;
1332}
1333
1334static void rv770_pcie_gen2_enable(struct radeon_device *rdev)
1335{
1336 u32 link_width_cntl, lanes, speed_cntl, tmp;
1337 u16 link_cntl2;
1338
1339 if (radeon_pcie_gen2 == 0)
1340 return;
1341
1342 if (rdev->flags & RADEON_IS_IGP)
1343 return;
1344
1345 if (!(rdev->flags & RADEON_IS_PCIE))
1346 return;
1347
1348 /* x2 cards have a special sequence */
1349 if (ASIC_IS_X2(rdev))
1350 return;
1351
1352 /* advertise upconfig capability */
1353 link_width_cntl = RREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL);
1354 link_width_cntl &= ~LC_UPCONFIGURE_DIS;
1355 WREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
1356 link_width_cntl = RREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL);
1357 if (link_width_cntl & LC_RENEGOTIATION_SUPPORT) {
1358 lanes = (link_width_cntl & LC_LINK_WIDTH_RD_MASK) >> LC_LINK_WIDTH_RD_SHIFT;
1359 link_width_cntl &= ~(LC_LINK_WIDTH_MASK |
1360 LC_RECONFIG_ARC_MISSING_ESCAPE);
1361 link_width_cntl |= lanes | LC_RECONFIG_NOW |
1362 LC_RENEGOTIATE_EN | LC_UPCONFIGURE_SUPPORT;
1363 WREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
1364 } else {
1365 link_width_cntl |= LC_UPCONFIGURE_DIS;
1366 WREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
1367 }
1368
1369 speed_cntl = RREG32_PCIE_P(PCIE_LC_SPEED_CNTL);
1370 if ((speed_cntl & LC_OTHER_SIDE_EVER_SENT_GEN2) &&
1371 (speed_cntl & LC_OTHER_SIDE_SUPPORTS_GEN2)) {
1372
1373 tmp = RREG32(0x541c);
1374 WREG32(0x541c, tmp | 0x8);
1375 WREG32(MM_CFGREGS_CNTL, MM_WR_TO_CFG_EN);
1376 link_cntl2 = RREG16(0x4088);
1377 link_cntl2 &= ~TARGET_LINK_SPEED_MASK;
1378 link_cntl2 |= 0x2;
1379 WREG16(0x4088, link_cntl2);
1380 WREG32(MM_CFGREGS_CNTL, 0);
1381
1382 speed_cntl = RREG32_PCIE_P(PCIE_LC_SPEED_CNTL);
1383 speed_cntl &= ~LC_TARGET_LINK_SPEED_OVERRIDE_EN;
1384 WREG32_PCIE_P(PCIE_LC_SPEED_CNTL, speed_cntl);
1385
1386 speed_cntl = RREG32_PCIE_P(PCIE_LC_SPEED_CNTL);
1387 speed_cntl |= LC_CLR_FAILED_SPD_CHANGE_CNT;
1388 WREG32_PCIE_P(PCIE_LC_SPEED_CNTL, speed_cntl);
1389
1390 speed_cntl = RREG32_PCIE_P(PCIE_LC_SPEED_CNTL);
1391 speed_cntl &= ~LC_CLR_FAILED_SPD_CHANGE_CNT;
1392 WREG32_PCIE_P(PCIE_LC_SPEED_CNTL, speed_cntl);
1393
1394 speed_cntl = RREG32_PCIE_P(PCIE_LC_SPEED_CNTL);
1395 speed_cntl |= LC_GEN2_EN_STRAP;
1396 WREG32_PCIE_P(PCIE_LC_SPEED_CNTL, speed_cntl);
1397
1398 } else {
1399 link_width_cntl = RREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL);
1400 /* XXX: only disable it if gen1 bridge vendor == 0x111d or 0x1106 */
1401 if (1)
1402 link_width_cntl |= LC_UPCONFIGURE_DIS;
1403 else
1404 link_width_cntl &= ~LC_UPCONFIGURE_DIS;
1405 WREG32_PCIE_P(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
1406 }
1407}
1/*
2 * Copyright 2008 Advanced Micro Devices, Inc.
3 * Copyright 2008 Red Hat Inc.
4 * Copyright 2009 Jerome Glisse.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22 * OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors: Dave Airlie
25 * Alex Deucher
26 * Jerome Glisse
27 */
28
29#include <linux/firmware.h>
30#include <linux/pci.h>
31#include <linux/slab.h>
32
33#include <drm/drm_device.h>
34#include <drm/radeon_drm.h>
35
36#include "atom.h"
37#include "avivod.h"
38#include "radeon.h"
39#include "radeon_asic.h"
40#include "radeon_audio.h"
41#include "rv770d.h"
42
43#define R700_PFP_UCODE_SIZE 848
44#define R700_PM4_UCODE_SIZE 1360
45
46static void rv770_gpu_init(struct radeon_device *rdev);
47void rv770_fini(struct radeon_device *rdev);
48static void rv770_pcie_gen2_enable(struct radeon_device *rdev);
49int evergreen_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk);
50
51int rv770_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk)
52{
53 unsigned fb_div = 0, vclk_div = 0, dclk_div = 0;
54 int r;
55
56 /* RV740 uses evergreen uvd clk programming */
57 if (rdev->family == CHIP_RV740)
58 return evergreen_set_uvd_clocks(rdev, vclk, dclk);
59
60 /* bypass vclk and dclk with bclk */
61 WREG32_P(CG_UPLL_FUNC_CNTL_2,
62 VCLK_SRC_SEL(1) | DCLK_SRC_SEL(1),
63 ~(VCLK_SRC_SEL_MASK | DCLK_SRC_SEL_MASK));
64
65 if (!vclk || !dclk) {
66 /* keep the Bypass mode, put PLL to sleep */
67 WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_SLEEP_MASK, ~UPLL_SLEEP_MASK);
68 return 0;
69 }
70
71 r = radeon_uvd_calc_upll_dividers(rdev, vclk, dclk, 50000, 160000,
72 43663, 0x03FFFFFE, 1, 30, ~0,
73 &fb_div, &vclk_div, &dclk_div);
74 if (r)
75 return r;
76
77 fb_div |= 1;
78 vclk_div -= 1;
79 dclk_div -= 1;
80
81 /* set UPLL_FB_DIV to 0x50000 */
82 WREG32_P(CG_UPLL_FUNC_CNTL_3, UPLL_FB_DIV(0x50000), ~UPLL_FB_DIV_MASK);
83
84 /* deassert UPLL_RESET and UPLL_SLEEP */
85 WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~(UPLL_RESET_MASK | UPLL_SLEEP_MASK));
86
87 /* assert BYPASS EN and FB_DIV[0] <- ??? why? */
88 WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_BYPASS_EN_MASK, ~UPLL_BYPASS_EN_MASK);
89 WREG32_P(CG_UPLL_FUNC_CNTL_3, UPLL_FB_DIV(1), ~UPLL_FB_DIV(1));
90
91 r = radeon_uvd_send_upll_ctlreq(rdev, CG_UPLL_FUNC_CNTL);
92 if (r)
93 return r;
94
95 /* assert PLL_RESET */
96 WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_RESET_MASK, ~UPLL_RESET_MASK);
97
98 /* set the required FB_DIV, REF_DIV, Post divder values */
99 WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_REF_DIV(1), ~UPLL_REF_DIV_MASK);
100 WREG32_P(CG_UPLL_FUNC_CNTL_2,
101 UPLL_SW_HILEN(vclk_div >> 1) |
102 UPLL_SW_LOLEN((vclk_div >> 1) + (vclk_div & 1)) |
103 UPLL_SW_HILEN2(dclk_div >> 1) |
104 UPLL_SW_LOLEN2((dclk_div >> 1) + (dclk_div & 1)),
105 ~UPLL_SW_MASK);
106
107 WREG32_P(CG_UPLL_FUNC_CNTL_3, UPLL_FB_DIV(fb_div),
108 ~UPLL_FB_DIV_MASK);
109
110 /* give the PLL some time to settle */
111 mdelay(15);
112
113 /* deassert PLL_RESET */
114 WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_RESET_MASK);
115
116 mdelay(15);
117
118 /* deassert BYPASS EN and FB_DIV[0] <- ??? why? */
119 WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_BYPASS_EN_MASK);
120 WREG32_P(CG_UPLL_FUNC_CNTL_3, 0, ~UPLL_FB_DIV(1));
121
122 r = radeon_uvd_send_upll_ctlreq(rdev, CG_UPLL_FUNC_CNTL);
123 if (r)
124 return r;
125
126 /* switch VCLK and DCLK selection */
127 WREG32_P(CG_UPLL_FUNC_CNTL_2,
128 VCLK_SRC_SEL(2) | DCLK_SRC_SEL(2),
129 ~(VCLK_SRC_SEL_MASK | DCLK_SRC_SEL_MASK));
130
131 mdelay(100);
132
133 return 0;
134}
135
136static const u32 r7xx_golden_registers[] =
137{
138 0x8d00, 0xffffffff, 0x0e0e0074,
139 0x8d04, 0xffffffff, 0x013a2b34,
140 0x9508, 0xffffffff, 0x00000002,
141 0x8b20, 0xffffffff, 0,
142 0x88c4, 0xffffffff, 0x000000c2,
143 0x28350, 0xffffffff, 0,
144 0x9058, 0xffffffff, 0x0fffc40f,
145 0x240c, 0xffffffff, 0x00000380,
146 0x733c, 0xffffffff, 0x00000002,
147 0x2650, 0x00040000, 0,
148 0x20bc, 0x00040000, 0,
149 0x7300, 0xffffffff, 0x001000f0
150};
151
152static const u32 r7xx_golden_dyn_gpr_registers[] =
153{
154 0x8db0, 0xffffffff, 0x98989898,
155 0x8db4, 0xffffffff, 0x98989898,
156 0x8db8, 0xffffffff, 0x98989898,
157 0x8dbc, 0xffffffff, 0x98989898,
158 0x8dc0, 0xffffffff, 0x98989898,
159 0x8dc4, 0xffffffff, 0x98989898,
160 0x8dc8, 0xffffffff, 0x98989898,
161 0x8dcc, 0xffffffff, 0x98989898,
162 0x88c4, 0xffffffff, 0x00000082
163};
164
165static const u32 rv770_golden_registers[] =
166{
167 0x562c, 0xffffffff, 0,
168 0x3f90, 0xffffffff, 0,
169 0x9148, 0xffffffff, 0,
170 0x3f94, 0xffffffff, 0,
171 0x914c, 0xffffffff, 0,
172 0x9698, 0x18000000, 0x18000000
173};
174
175static const u32 rv770ce_golden_registers[] =
176{
177 0x562c, 0xffffffff, 0,
178 0x3f90, 0xffffffff, 0x00cc0000,
179 0x9148, 0xffffffff, 0x00cc0000,
180 0x3f94, 0xffffffff, 0x00cc0000,
181 0x914c, 0xffffffff, 0x00cc0000,
182 0x9b7c, 0xffffffff, 0x00fa0000,
183 0x3f8c, 0xffffffff, 0x00fa0000,
184 0x9698, 0x18000000, 0x18000000
185};
186
187static const u32 rv770_mgcg_init[] =
188{
189 0x8bcc, 0xffffffff, 0x130300f9,
190 0x5448, 0xffffffff, 0x100,
191 0x55e4, 0xffffffff, 0x100,
192 0x160c, 0xffffffff, 0x100,
193 0x5644, 0xffffffff, 0x100,
194 0xc164, 0xffffffff, 0x100,
195 0x8a18, 0xffffffff, 0x100,
196 0x897c, 0xffffffff, 0x8000100,
197 0x8b28, 0xffffffff, 0x3c000100,
198 0x9144, 0xffffffff, 0x100,
199 0x9a1c, 0xffffffff, 0x10000,
200 0x9a50, 0xffffffff, 0x100,
201 0x9a1c, 0xffffffff, 0x10001,
202 0x9a50, 0xffffffff, 0x100,
203 0x9a1c, 0xffffffff, 0x10002,
204 0x9a50, 0xffffffff, 0x100,
205 0x9a1c, 0xffffffff, 0x10003,
206 0x9a50, 0xffffffff, 0x100,
207 0x9a1c, 0xffffffff, 0x0,
208 0x9870, 0xffffffff, 0x100,
209 0x8d58, 0xffffffff, 0x100,
210 0x9500, 0xffffffff, 0x0,
211 0x9510, 0xffffffff, 0x100,
212 0x9500, 0xffffffff, 0x1,
213 0x9510, 0xffffffff, 0x100,
214 0x9500, 0xffffffff, 0x2,
215 0x9510, 0xffffffff, 0x100,
216 0x9500, 0xffffffff, 0x3,
217 0x9510, 0xffffffff, 0x100,
218 0x9500, 0xffffffff, 0x4,
219 0x9510, 0xffffffff, 0x100,
220 0x9500, 0xffffffff, 0x5,
221 0x9510, 0xffffffff, 0x100,
222 0x9500, 0xffffffff, 0x6,
223 0x9510, 0xffffffff, 0x100,
224 0x9500, 0xffffffff, 0x7,
225 0x9510, 0xffffffff, 0x100,
226 0x9500, 0xffffffff, 0x8,
227 0x9510, 0xffffffff, 0x100,
228 0x9500, 0xffffffff, 0x9,
229 0x9510, 0xffffffff, 0x100,
230 0x9500, 0xffffffff, 0x8000,
231 0x9490, 0xffffffff, 0x0,
232 0x949c, 0xffffffff, 0x100,
233 0x9490, 0xffffffff, 0x1,
234 0x949c, 0xffffffff, 0x100,
235 0x9490, 0xffffffff, 0x2,
236 0x949c, 0xffffffff, 0x100,
237 0x9490, 0xffffffff, 0x3,
238 0x949c, 0xffffffff, 0x100,
239 0x9490, 0xffffffff, 0x4,
240 0x949c, 0xffffffff, 0x100,
241 0x9490, 0xffffffff, 0x5,
242 0x949c, 0xffffffff, 0x100,
243 0x9490, 0xffffffff, 0x6,
244 0x949c, 0xffffffff, 0x100,
245 0x9490, 0xffffffff, 0x7,
246 0x949c, 0xffffffff, 0x100,
247 0x9490, 0xffffffff, 0x8,
248 0x949c, 0xffffffff, 0x100,
249 0x9490, 0xffffffff, 0x9,
250 0x949c, 0xffffffff, 0x100,
251 0x9490, 0xffffffff, 0x8000,
252 0x9604, 0xffffffff, 0x0,
253 0x9654, 0xffffffff, 0x100,
254 0x9604, 0xffffffff, 0x1,
255 0x9654, 0xffffffff, 0x100,
256 0x9604, 0xffffffff, 0x2,
257 0x9654, 0xffffffff, 0x100,
258 0x9604, 0xffffffff, 0x3,
259 0x9654, 0xffffffff, 0x100,
260 0x9604, 0xffffffff, 0x4,
261 0x9654, 0xffffffff, 0x100,
262 0x9604, 0xffffffff, 0x5,
263 0x9654, 0xffffffff, 0x100,
264 0x9604, 0xffffffff, 0x6,
265 0x9654, 0xffffffff, 0x100,
266 0x9604, 0xffffffff, 0x7,
267 0x9654, 0xffffffff, 0x100,
268 0x9604, 0xffffffff, 0x8,
269 0x9654, 0xffffffff, 0x100,
270 0x9604, 0xffffffff, 0x9,
271 0x9654, 0xffffffff, 0x100,
272 0x9604, 0xffffffff, 0x80000000,
273 0x9030, 0xffffffff, 0x100,
274 0x9034, 0xffffffff, 0x100,
275 0x9038, 0xffffffff, 0x100,
276 0x903c, 0xffffffff, 0x100,
277 0x9040, 0xffffffff, 0x100,
278 0xa200, 0xffffffff, 0x100,
279 0xa204, 0xffffffff, 0x100,
280 0xa208, 0xffffffff, 0x100,
281 0xa20c, 0xffffffff, 0x100,
282 0x971c, 0xffffffff, 0x100,
283 0x915c, 0xffffffff, 0x00020001,
284 0x9160, 0xffffffff, 0x00040003,
285 0x916c, 0xffffffff, 0x00060005,
286 0x9170, 0xffffffff, 0x00080007,
287 0x9174, 0xffffffff, 0x000a0009,
288 0x9178, 0xffffffff, 0x000c000b,
289 0x917c, 0xffffffff, 0x000e000d,
290 0x9180, 0xffffffff, 0x0010000f,
291 0x918c, 0xffffffff, 0x00120011,
292 0x9190, 0xffffffff, 0x00140013,
293 0x9194, 0xffffffff, 0x00020001,
294 0x9198, 0xffffffff, 0x00040003,
295 0x919c, 0xffffffff, 0x00060005,
296 0x91a8, 0xffffffff, 0x00080007,
297 0x91ac, 0xffffffff, 0x000a0009,
298 0x91b0, 0xffffffff, 0x000c000b,
299 0x91b4, 0xffffffff, 0x000e000d,
300 0x91b8, 0xffffffff, 0x0010000f,
301 0x91c4, 0xffffffff, 0x00120011,
302 0x91c8, 0xffffffff, 0x00140013,
303 0x91cc, 0xffffffff, 0x00020001,
304 0x91d0, 0xffffffff, 0x00040003,
305 0x91d4, 0xffffffff, 0x00060005,
306 0x91e0, 0xffffffff, 0x00080007,
307 0x91e4, 0xffffffff, 0x000a0009,
308 0x91e8, 0xffffffff, 0x000c000b,
309 0x91ec, 0xffffffff, 0x00020001,
310 0x91f0, 0xffffffff, 0x00040003,
311 0x91f4, 0xffffffff, 0x00060005,
312 0x9200, 0xffffffff, 0x00080007,
313 0x9204, 0xffffffff, 0x000a0009,
314 0x9208, 0xffffffff, 0x000c000b,
315 0x920c, 0xffffffff, 0x000e000d,
316 0x9210, 0xffffffff, 0x0010000f,
317 0x921c, 0xffffffff, 0x00120011,
318 0x9220, 0xffffffff, 0x00140013,
319 0x9224, 0xffffffff, 0x00020001,
320 0x9228, 0xffffffff, 0x00040003,
321 0x922c, 0xffffffff, 0x00060005,
322 0x9238, 0xffffffff, 0x00080007,
323 0x923c, 0xffffffff, 0x000a0009,
324 0x9240, 0xffffffff, 0x000c000b,
325 0x9244, 0xffffffff, 0x000e000d,
326 0x9248, 0xffffffff, 0x0010000f,
327 0x9254, 0xffffffff, 0x00120011,
328 0x9258, 0xffffffff, 0x00140013,
329 0x925c, 0xffffffff, 0x00020001,
330 0x9260, 0xffffffff, 0x00040003,
331 0x9264, 0xffffffff, 0x00060005,
332 0x9270, 0xffffffff, 0x00080007,
333 0x9274, 0xffffffff, 0x000a0009,
334 0x9278, 0xffffffff, 0x000c000b,
335 0x927c, 0xffffffff, 0x000e000d,
336 0x9280, 0xffffffff, 0x0010000f,
337 0x928c, 0xffffffff, 0x00120011,
338 0x9290, 0xffffffff, 0x00140013,
339 0x9294, 0xffffffff, 0x00020001,
340 0x929c, 0xffffffff, 0x00040003,
341 0x92a0, 0xffffffff, 0x00060005,
342 0x92a4, 0xffffffff, 0x00080007
343};
344
345static const u32 rv710_golden_registers[] =
346{
347 0x3f90, 0x00ff0000, 0x00fc0000,
348 0x9148, 0x00ff0000, 0x00fc0000,
349 0x3f94, 0x00ff0000, 0x00fc0000,
350 0x914c, 0x00ff0000, 0x00fc0000,
351 0xb4c, 0x00000020, 0x00000020,
352 0xa180, 0xffffffff, 0x00003f3f
353};
354
355static const u32 rv710_mgcg_init[] =
356{
357 0x8bcc, 0xffffffff, 0x13030040,
358 0x5448, 0xffffffff, 0x100,
359 0x55e4, 0xffffffff, 0x100,
360 0x160c, 0xffffffff, 0x100,
361 0x5644, 0xffffffff, 0x100,
362 0xc164, 0xffffffff, 0x100,
363 0x8a18, 0xffffffff, 0x100,
364 0x897c, 0xffffffff, 0x8000100,
365 0x8b28, 0xffffffff, 0x3c000100,
366 0x9144, 0xffffffff, 0x100,
367 0x9a1c, 0xffffffff, 0x10000,
368 0x9a50, 0xffffffff, 0x100,
369 0x9a1c, 0xffffffff, 0x0,
370 0x9870, 0xffffffff, 0x100,
371 0x8d58, 0xffffffff, 0x100,
372 0x9500, 0xffffffff, 0x0,
373 0x9510, 0xffffffff, 0x100,
374 0x9500, 0xffffffff, 0x1,
375 0x9510, 0xffffffff, 0x100,
376 0x9500, 0xffffffff, 0x8000,
377 0x9490, 0xffffffff, 0x0,
378 0x949c, 0xffffffff, 0x100,
379 0x9490, 0xffffffff, 0x1,
380 0x949c, 0xffffffff, 0x100,
381 0x9490, 0xffffffff, 0x8000,
382 0x9604, 0xffffffff, 0x0,
383 0x9654, 0xffffffff, 0x100,
384 0x9604, 0xffffffff, 0x1,
385 0x9654, 0xffffffff, 0x100,
386 0x9604, 0xffffffff, 0x80000000,
387 0x9030, 0xffffffff, 0x100,
388 0x9034, 0xffffffff, 0x100,
389 0x9038, 0xffffffff, 0x100,
390 0x903c, 0xffffffff, 0x100,
391 0x9040, 0xffffffff, 0x100,
392 0xa200, 0xffffffff, 0x100,
393 0xa204, 0xffffffff, 0x100,
394 0xa208, 0xffffffff, 0x100,
395 0xa20c, 0xffffffff, 0x100,
396 0x971c, 0xffffffff, 0x100,
397 0x915c, 0xffffffff, 0x00020001,
398 0x9174, 0xffffffff, 0x00000003,
399 0x9178, 0xffffffff, 0x00050001,
400 0x917c, 0xffffffff, 0x00030002,
401 0x918c, 0xffffffff, 0x00000004,
402 0x9190, 0xffffffff, 0x00070006,
403 0x9194, 0xffffffff, 0x00050001,
404 0x9198, 0xffffffff, 0x00030002,
405 0x91a8, 0xffffffff, 0x00000004,
406 0x91ac, 0xffffffff, 0x00070006,
407 0x91e8, 0xffffffff, 0x00000001,
408 0x9294, 0xffffffff, 0x00000001,
409 0x929c, 0xffffffff, 0x00000002,
410 0x92a0, 0xffffffff, 0x00040003,
411 0x9150, 0xffffffff, 0x4d940000
412};
413
414static const u32 rv730_golden_registers[] =
415{
416 0x3f90, 0x00ff0000, 0x00f00000,
417 0x9148, 0x00ff0000, 0x00f00000,
418 0x3f94, 0x00ff0000, 0x00f00000,
419 0x914c, 0x00ff0000, 0x00f00000,
420 0x900c, 0xffffffff, 0x003b033f,
421 0xb4c, 0x00000020, 0x00000020,
422 0xa180, 0xffffffff, 0x00003f3f
423};
424
425static const u32 rv730_mgcg_init[] =
426{
427 0x8bcc, 0xffffffff, 0x130300f9,
428 0x5448, 0xffffffff, 0x100,
429 0x55e4, 0xffffffff, 0x100,
430 0x160c, 0xffffffff, 0x100,
431 0x5644, 0xffffffff, 0x100,
432 0xc164, 0xffffffff, 0x100,
433 0x8a18, 0xffffffff, 0x100,
434 0x897c, 0xffffffff, 0x8000100,
435 0x8b28, 0xffffffff, 0x3c000100,
436 0x9144, 0xffffffff, 0x100,
437 0x9a1c, 0xffffffff, 0x10000,
438 0x9a50, 0xffffffff, 0x100,
439 0x9a1c, 0xffffffff, 0x10001,
440 0x9a50, 0xffffffff, 0x100,
441 0x9a1c, 0xffffffff, 0x0,
442 0x9870, 0xffffffff, 0x100,
443 0x8d58, 0xffffffff, 0x100,
444 0x9500, 0xffffffff, 0x0,
445 0x9510, 0xffffffff, 0x100,
446 0x9500, 0xffffffff, 0x1,
447 0x9510, 0xffffffff, 0x100,
448 0x9500, 0xffffffff, 0x2,
449 0x9510, 0xffffffff, 0x100,
450 0x9500, 0xffffffff, 0x3,
451 0x9510, 0xffffffff, 0x100,
452 0x9500, 0xffffffff, 0x4,
453 0x9510, 0xffffffff, 0x100,
454 0x9500, 0xffffffff, 0x5,
455 0x9510, 0xffffffff, 0x100,
456 0x9500, 0xffffffff, 0x6,
457 0x9510, 0xffffffff, 0x100,
458 0x9500, 0xffffffff, 0x7,
459 0x9510, 0xffffffff, 0x100,
460 0x9500, 0xffffffff, 0x8000,
461 0x9490, 0xffffffff, 0x0,
462 0x949c, 0xffffffff, 0x100,
463 0x9490, 0xffffffff, 0x1,
464 0x949c, 0xffffffff, 0x100,
465 0x9490, 0xffffffff, 0x2,
466 0x949c, 0xffffffff, 0x100,
467 0x9490, 0xffffffff, 0x3,
468 0x949c, 0xffffffff, 0x100,
469 0x9490, 0xffffffff, 0x4,
470 0x949c, 0xffffffff, 0x100,
471 0x9490, 0xffffffff, 0x5,
472 0x949c, 0xffffffff, 0x100,
473 0x9490, 0xffffffff, 0x6,
474 0x949c, 0xffffffff, 0x100,
475 0x9490, 0xffffffff, 0x7,
476 0x949c, 0xffffffff, 0x100,
477 0x9490, 0xffffffff, 0x8000,
478 0x9604, 0xffffffff, 0x0,
479 0x9654, 0xffffffff, 0x100,
480 0x9604, 0xffffffff, 0x1,
481 0x9654, 0xffffffff, 0x100,
482 0x9604, 0xffffffff, 0x2,
483 0x9654, 0xffffffff, 0x100,
484 0x9604, 0xffffffff, 0x3,
485 0x9654, 0xffffffff, 0x100,
486 0x9604, 0xffffffff, 0x4,
487 0x9654, 0xffffffff, 0x100,
488 0x9604, 0xffffffff, 0x5,
489 0x9654, 0xffffffff, 0x100,
490 0x9604, 0xffffffff, 0x6,
491 0x9654, 0xffffffff, 0x100,
492 0x9604, 0xffffffff, 0x7,
493 0x9654, 0xffffffff, 0x100,
494 0x9604, 0xffffffff, 0x80000000,
495 0x9030, 0xffffffff, 0x100,
496 0x9034, 0xffffffff, 0x100,
497 0x9038, 0xffffffff, 0x100,
498 0x903c, 0xffffffff, 0x100,
499 0x9040, 0xffffffff, 0x100,
500 0xa200, 0xffffffff, 0x100,
501 0xa204, 0xffffffff, 0x100,
502 0xa208, 0xffffffff, 0x100,
503 0xa20c, 0xffffffff, 0x100,
504 0x971c, 0xffffffff, 0x100,
505 0x915c, 0xffffffff, 0x00020001,
506 0x916c, 0xffffffff, 0x00040003,
507 0x9170, 0xffffffff, 0x00000005,
508 0x9178, 0xffffffff, 0x00050001,
509 0x917c, 0xffffffff, 0x00030002,
510 0x918c, 0xffffffff, 0x00000004,
511 0x9190, 0xffffffff, 0x00070006,
512 0x9194, 0xffffffff, 0x00050001,
513 0x9198, 0xffffffff, 0x00030002,
514 0x91a8, 0xffffffff, 0x00000004,
515 0x91ac, 0xffffffff, 0x00070006,
516 0x91b0, 0xffffffff, 0x00050001,
517 0x91b4, 0xffffffff, 0x00030002,
518 0x91c4, 0xffffffff, 0x00000004,
519 0x91c8, 0xffffffff, 0x00070006,
520 0x91cc, 0xffffffff, 0x00050001,
521 0x91d0, 0xffffffff, 0x00030002,
522 0x91e0, 0xffffffff, 0x00000004,
523 0x91e4, 0xffffffff, 0x00070006,
524 0x91e8, 0xffffffff, 0x00000001,
525 0x91ec, 0xffffffff, 0x00050001,
526 0x91f0, 0xffffffff, 0x00030002,
527 0x9200, 0xffffffff, 0x00000004,
528 0x9204, 0xffffffff, 0x00070006,
529 0x9208, 0xffffffff, 0x00050001,
530 0x920c, 0xffffffff, 0x00030002,
531 0x921c, 0xffffffff, 0x00000004,
532 0x9220, 0xffffffff, 0x00070006,
533 0x9224, 0xffffffff, 0x00050001,
534 0x9228, 0xffffffff, 0x00030002,
535 0x9238, 0xffffffff, 0x00000004,
536 0x923c, 0xffffffff, 0x00070006,
537 0x9240, 0xffffffff, 0x00050001,
538 0x9244, 0xffffffff, 0x00030002,
539 0x9254, 0xffffffff, 0x00000004,
540 0x9258, 0xffffffff, 0x00070006,
541 0x9294, 0xffffffff, 0x00000001,
542 0x929c, 0xffffffff, 0x00000002,
543 0x92a0, 0xffffffff, 0x00040003,
544 0x92a4, 0xffffffff, 0x00000005
545};
546
547static const u32 rv740_golden_registers[] =
548{
549 0x88c4, 0xffffffff, 0x00000082,
550 0x28a50, 0xfffffffc, 0x00000004,
551 0x2650, 0x00040000, 0,
552 0x20bc, 0x00040000, 0,
553 0x733c, 0xffffffff, 0x00000002,
554 0x7300, 0xffffffff, 0x001000f0,
555 0x3f90, 0x00ff0000, 0,
556 0x9148, 0x00ff0000, 0,
557 0x3f94, 0x00ff0000, 0,
558 0x914c, 0x00ff0000, 0,
559 0x240c, 0xffffffff, 0x00000380,
560 0x8a14, 0x00000007, 0x00000007,
561 0x8b24, 0xffffffff, 0x00ff0fff,
562 0x28a4c, 0xffffffff, 0x00004000,
563 0xa180, 0xffffffff, 0x00003f3f,
564 0x8d00, 0xffffffff, 0x0e0e003a,
565 0x8d04, 0xffffffff, 0x013a0e2a,
566 0x8c00, 0xffffffff, 0xe400000f,
567 0x8db0, 0xffffffff, 0x98989898,
568 0x8db4, 0xffffffff, 0x98989898,
569 0x8db8, 0xffffffff, 0x98989898,
570 0x8dbc, 0xffffffff, 0x98989898,
571 0x8dc0, 0xffffffff, 0x98989898,
572 0x8dc4, 0xffffffff, 0x98989898,
573 0x8dc8, 0xffffffff, 0x98989898,
574 0x8dcc, 0xffffffff, 0x98989898,
575 0x9058, 0xffffffff, 0x0fffc40f,
576 0x900c, 0xffffffff, 0x003b033f,
577 0x28350, 0xffffffff, 0,
578 0x8cf0, 0x1fffffff, 0x08e00420,
579 0x9508, 0xffffffff, 0x00000002,
580 0x88c4, 0xffffffff, 0x000000c2,
581 0x9698, 0x18000000, 0x18000000
582};
583
584static const u32 rv740_mgcg_init[] =
585{
586 0x8bcc, 0xffffffff, 0x13030100,
587 0x5448, 0xffffffff, 0x100,
588 0x55e4, 0xffffffff, 0x100,
589 0x160c, 0xffffffff, 0x100,
590 0x5644, 0xffffffff, 0x100,
591 0xc164, 0xffffffff, 0x100,
592 0x8a18, 0xffffffff, 0x100,
593 0x897c, 0xffffffff, 0x100,
594 0x8b28, 0xffffffff, 0x100,
595 0x9144, 0xffffffff, 0x100,
596 0x9a1c, 0xffffffff, 0x10000,
597 0x9a50, 0xffffffff, 0x100,
598 0x9a1c, 0xffffffff, 0x10001,
599 0x9a50, 0xffffffff, 0x100,
600 0x9a1c, 0xffffffff, 0x10002,
601 0x9a50, 0xffffffff, 0x100,
602 0x9a1c, 0xffffffff, 0x10003,
603 0x9a50, 0xffffffff, 0x100,
604 0x9a1c, 0xffffffff, 0x0,
605 0x9870, 0xffffffff, 0x100,
606 0x8d58, 0xffffffff, 0x100,
607 0x9500, 0xffffffff, 0x0,
608 0x9510, 0xffffffff, 0x100,
609 0x9500, 0xffffffff, 0x1,
610 0x9510, 0xffffffff, 0x100,
611 0x9500, 0xffffffff, 0x2,
612 0x9510, 0xffffffff, 0x100,
613 0x9500, 0xffffffff, 0x3,
614 0x9510, 0xffffffff, 0x100,
615 0x9500, 0xffffffff, 0x4,
616 0x9510, 0xffffffff, 0x100,
617 0x9500, 0xffffffff, 0x5,
618 0x9510, 0xffffffff, 0x100,
619 0x9500, 0xffffffff, 0x6,
620 0x9510, 0xffffffff, 0x100,
621 0x9500, 0xffffffff, 0x7,
622 0x9510, 0xffffffff, 0x100,
623 0x9500, 0xffffffff, 0x8000,
624 0x9490, 0xffffffff, 0x0,
625 0x949c, 0xffffffff, 0x100,
626 0x9490, 0xffffffff, 0x1,
627 0x949c, 0xffffffff, 0x100,
628 0x9490, 0xffffffff, 0x2,
629 0x949c, 0xffffffff, 0x100,
630 0x9490, 0xffffffff, 0x3,
631 0x949c, 0xffffffff, 0x100,
632 0x9490, 0xffffffff, 0x4,
633 0x949c, 0xffffffff, 0x100,
634 0x9490, 0xffffffff, 0x5,
635 0x949c, 0xffffffff, 0x100,
636 0x9490, 0xffffffff, 0x6,
637 0x949c, 0xffffffff, 0x100,
638 0x9490, 0xffffffff, 0x7,
639 0x949c, 0xffffffff, 0x100,
640 0x9490, 0xffffffff, 0x8000,
641 0x9604, 0xffffffff, 0x0,
642 0x9654, 0xffffffff, 0x100,
643 0x9604, 0xffffffff, 0x1,
644 0x9654, 0xffffffff, 0x100,
645 0x9604, 0xffffffff, 0x2,
646 0x9654, 0xffffffff, 0x100,
647 0x9604, 0xffffffff, 0x3,
648 0x9654, 0xffffffff, 0x100,
649 0x9604, 0xffffffff, 0x4,
650 0x9654, 0xffffffff, 0x100,
651 0x9604, 0xffffffff, 0x5,
652 0x9654, 0xffffffff, 0x100,
653 0x9604, 0xffffffff, 0x6,
654 0x9654, 0xffffffff, 0x100,
655 0x9604, 0xffffffff, 0x7,
656 0x9654, 0xffffffff, 0x100,
657 0x9604, 0xffffffff, 0x80000000,
658 0x9030, 0xffffffff, 0x100,
659 0x9034, 0xffffffff, 0x100,
660 0x9038, 0xffffffff, 0x100,
661 0x903c, 0xffffffff, 0x100,
662 0x9040, 0xffffffff, 0x100,
663 0xa200, 0xffffffff, 0x100,
664 0xa204, 0xffffffff, 0x100,
665 0xa208, 0xffffffff, 0x100,
666 0xa20c, 0xffffffff, 0x100,
667 0x971c, 0xffffffff, 0x100,
668 0x915c, 0xffffffff, 0x00020001,
669 0x9160, 0xffffffff, 0x00040003,
670 0x916c, 0xffffffff, 0x00060005,
671 0x9170, 0xffffffff, 0x00080007,
672 0x9174, 0xffffffff, 0x000a0009,
673 0x9178, 0xffffffff, 0x000c000b,
674 0x917c, 0xffffffff, 0x000e000d,
675 0x9180, 0xffffffff, 0x0010000f,
676 0x918c, 0xffffffff, 0x00120011,
677 0x9190, 0xffffffff, 0x00140013,
678 0x9194, 0xffffffff, 0x00020001,
679 0x9198, 0xffffffff, 0x00040003,
680 0x919c, 0xffffffff, 0x00060005,
681 0x91a8, 0xffffffff, 0x00080007,
682 0x91ac, 0xffffffff, 0x000a0009,
683 0x91b0, 0xffffffff, 0x000c000b,
684 0x91b4, 0xffffffff, 0x000e000d,
685 0x91b8, 0xffffffff, 0x0010000f,
686 0x91c4, 0xffffffff, 0x00120011,
687 0x91c8, 0xffffffff, 0x00140013,
688 0x91cc, 0xffffffff, 0x00020001,
689 0x91d0, 0xffffffff, 0x00040003,
690 0x91d4, 0xffffffff, 0x00060005,
691 0x91e0, 0xffffffff, 0x00080007,
692 0x91e4, 0xffffffff, 0x000a0009,
693 0x91e8, 0xffffffff, 0x000c000b,
694 0x91ec, 0xffffffff, 0x00020001,
695 0x91f0, 0xffffffff, 0x00040003,
696 0x91f4, 0xffffffff, 0x00060005,
697 0x9200, 0xffffffff, 0x00080007,
698 0x9204, 0xffffffff, 0x000a0009,
699 0x9208, 0xffffffff, 0x000c000b,
700 0x920c, 0xffffffff, 0x000e000d,
701 0x9210, 0xffffffff, 0x0010000f,
702 0x921c, 0xffffffff, 0x00120011,
703 0x9220, 0xffffffff, 0x00140013,
704 0x9224, 0xffffffff, 0x00020001,
705 0x9228, 0xffffffff, 0x00040003,
706 0x922c, 0xffffffff, 0x00060005,
707 0x9238, 0xffffffff, 0x00080007,
708 0x923c, 0xffffffff, 0x000a0009,
709 0x9240, 0xffffffff, 0x000c000b,
710 0x9244, 0xffffffff, 0x000e000d,
711 0x9248, 0xffffffff, 0x0010000f,
712 0x9254, 0xffffffff, 0x00120011,
713 0x9258, 0xffffffff, 0x00140013,
714 0x9294, 0xffffffff, 0x00020001,
715 0x929c, 0xffffffff, 0x00040003,
716 0x92a0, 0xffffffff, 0x00060005,
717 0x92a4, 0xffffffff, 0x00080007
718};
719
720static void rv770_init_golden_registers(struct radeon_device *rdev)
721{
722 switch (rdev->family) {
723 case CHIP_RV770:
724 radeon_program_register_sequence(rdev,
725 r7xx_golden_registers,
726 (const u32)ARRAY_SIZE(r7xx_golden_registers));
727 radeon_program_register_sequence(rdev,
728 r7xx_golden_dyn_gpr_registers,
729 (const u32)ARRAY_SIZE(r7xx_golden_dyn_gpr_registers));
730 if (rdev->pdev->device == 0x994e)
731 radeon_program_register_sequence(rdev,
732 rv770ce_golden_registers,
733 (const u32)ARRAY_SIZE(rv770ce_golden_registers));
734 else
735 radeon_program_register_sequence(rdev,
736 rv770_golden_registers,
737 (const u32)ARRAY_SIZE(rv770_golden_registers));
738 radeon_program_register_sequence(rdev,
739 rv770_mgcg_init,
740 (const u32)ARRAY_SIZE(rv770_mgcg_init));
741 break;
742 case CHIP_RV730:
743 radeon_program_register_sequence(rdev,
744 r7xx_golden_registers,
745 (const u32)ARRAY_SIZE(r7xx_golden_registers));
746 radeon_program_register_sequence(rdev,
747 r7xx_golden_dyn_gpr_registers,
748 (const u32)ARRAY_SIZE(r7xx_golden_dyn_gpr_registers));
749 radeon_program_register_sequence(rdev,
750 rv730_golden_registers,
751 (const u32)ARRAY_SIZE(rv730_golden_registers));
752 radeon_program_register_sequence(rdev,
753 rv730_mgcg_init,
754 (const u32)ARRAY_SIZE(rv730_mgcg_init));
755 break;
756 case CHIP_RV710:
757 radeon_program_register_sequence(rdev,
758 r7xx_golden_registers,
759 (const u32)ARRAY_SIZE(r7xx_golden_registers));
760 radeon_program_register_sequence(rdev,
761 r7xx_golden_dyn_gpr_registers,
762 (const u32)ARRAY_SIZE(r7xx_golden_dyn_gpr_registers));
763 radeon_program_register_sequence(rdev,
764 rv710_golden_registers,
765 (const u32)ARRAY_SIZE(rv710_golden_registers));
766 radeon_program_register_sequence(rdev,
767 rv710_mgcg_init,
768 (const u32)ARRAY_SIZE(rv710_mgcg_init));
769 break;
770 case CHIP_RV740:
771 radeon_program_register_sequence(rdev,
772 rv740_golden_registers,
773 (const u32)ARRAY_SIZE(rv740_golden_registers));
774 radeon_program_register_sequence(rdev,
775 rv740_mgcg_init,
776 (const u32)ARRAY_SIZE(rv740_mgcg_init));
777 break;
778 default:
779 break;
780 }
781}
782
783#define PCIE_BUS_CLK 10000
784#define TCLK (PCIE_BUS_CLK / 10)
785
786/**
787 * rv770_get_xclk - get the xclk
788 *
789 * @rdev: radeon_device pointer
790 *
791 * Returns the reference clock used by the gfx engine
792 * (r7xx-cayman).
793 */
794u32 rv770_get_xclk(struct radeon_device *rdev)
795{
796 u32 reference_clock = rdev->clock.spll.reference_freq;
797 u32 tmp = RREG32(CG_CLKPIN_CNTL);
798
799 if (tmp & MUX_TCLK_TO_XCLK)
800 return TCLK;
801
802 if (tmp & XTALIN_DIVIDE)
803 return reference_clock / 4;
804
805 return reference_clock;
806}
807
808void rv770_page_flip(struct radeon_device *rdev, int crtc_id, u64 crtc_base, bool async)
809{
810 struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
811 u32 tmp = RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset);
812 int i;
813
814 /* Lock the graphics update lock */
815 tmp |= AVIVO_D1GRPH_UPDATE_LOCK;
816 WREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset, tmp);
817
818 /* update the scanout addresses */
819 WREG32(AVIVO_D1GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset,
820 async ? AVIVO_D1GRPH_SURFACE_UPDATE_H_RETRACE_EN : 0);
821 if (radeon_crtc->crtc_id) {
822 WREG32(D2GRPH_SECONDARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
823 WREG32(D2GRPH_PRIMARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
824 } else {
825 WREG32(D1GRPH_SECONDARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
826 WREG32(D1GRPH_PRIMARY_SURFACE_ADDRESS_HIGH, upper_32_bits(crtc_base));
827 }
828 WREG32(D1GRPH_SECONDARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset,
829 (u32)crtc_base);
830 WREG32(D1GRPH_PRIMARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset,
831 (u32)crtc_base);
832
833 /* Wait for update_pending to go high. */
834 for (i = 0; i < rdev->usec_timeout; i++) {
835 if (RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset) & AVIVO_D1GRPH_SURFACE_UPDATE_PENDING)
836 break;
837 udelay(1);
838 }
839 DRM_DEBUG("Update pending now high. Unlocking vupdate_lock.\n");
840
841 /* Unlock the lock, so double-buffering can take place inside vblank */
842 tmp &= ~AVIVO_D1GRPH_UPDATE_LOCK;
843 WREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset, tmp);
844}
845
846bool rv770_page_flip_pending(struct radeon_device *rdev, int crtc_id)
847{
848 struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
849
850 /* Return current update_pending status: */
851 return !!(RREG32(AVIVO_D1GRPH_UPDATE + radeon_crtc->crtc_offset) &
852 AVIVO_D1GRPH_SURFACE_UPDATE_PENDING);
853}
854
855/* get temperature in millidegrees */
856int rv770_get_temp(struct radeon_device *rdev)
857{
858 u32 temp = (RREG32(CG_MULT_THERMAL_STATUS) & ASIC_T_MASK) >>
859 ASIC_T_SHIFT;
860 int actual_temp;
861
862 if (temp & 0x400)
863 actual_temp = -256;
864 else if (temp & 0x200)
865 actual_temp = 255;
866 else if (temp & 0x100) {
867 actual_temp = temp & 0x1ff;
868 actual_temp |= ~0x1ff;
869 } else
870 actual_temp = temp & 0xff;
871
872 return (actual_temp * 1000) / 2;
873}
874
875void rv770_pm_misc(struct radeon_device *rdev)
876{
877 int req_ps_idx = rdev->pm.requested_power_state_index;
878 int req_cm_idx = rdev->pm.requested_clock_mode_index;
879 struct radeon_power_state *ps = &rdev->pm.power_state[req_ps_idx];
880 struct radeon_voltage *voltage = &ps->clock_info[req_cm_idx].voltage;
881
882 if ((voltage->type == VOLTAGE_SW) && voltage->voltage) {
883 /* 0xff01 is a flag rather then an actual voltage */
884 if (voltage->voltage == 0xff01)
885 return;
886 if (voltage->voltage != rdev->pm.current_vddc) {
887 radeon_atom_set_voltage(rdev, voltage->voltage, SET_VOLTAGE_TYPE_ASIC_VDDC);
888 rdev->pm.current_vddc = voltage->voltage;
889 DRM_DEBUG("Setting: v: %d\n", voltage->voltage);
890 }
891 }
892}
893
894/*
895 * GART
896 */
897static int rv770_pcie_gart_enable(struct radeon_device *rdev)
898{
899 u32 tmp;
900 int r, i;
901
902 if (rdev->gart.robj == NULL) {
903 dev_err(rdev->dev, "No VRAM object for PCIE GART.\n");
904 return -EINVAL;
905 }
906 r = radeon_gart_table_vram_pin(rdev);
907 if (r)
908 return r;
909 /* Setup L2 cache */
910 WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
911 ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
912 EFFECTIVE_L2_QUEUE_SIZE(7));
913 WREG32(VM_L2_CNTL2, 0);
914 WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
915 /* Setup TLB control */
916 tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
917 SYSTEM_ACCESS_MODE_NOT_IN_SYS |
918 SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
919 EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
920 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
921 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
922 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
923 if (rdev->family == CHIP_RV740)
924 WREG32(MC_VM_MD_L1_TLB3_CNTL, tmp);
925 WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
926 WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
927 WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
928 WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
929 WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR, rdev->mc.gtt_start >> 12);
930 WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR, rdev->mc.gtt_end >> 12);
931 WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12);
932 WREG32(VM_CONTEXT0_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) |
933 RANGE_PROTECTION_FAULT_ENABLE_DEFAULT);
934 WREG32(VM_CONTEXT0_PROTECTION_FAULT_DEFAULT_ADDR,
935 (u32)(rdev->dummy_page.addr >> 12));
936 for (i = 1; i < 7; i++)
937 WREG32(VM_CONTEXT0_CNTL + (i * 4), 0);
938
939 r600_pcie_gart_tlb_flush(rdev);
940 DRM_INFO("PCIE GART of %uM enabled (table at 0x%016llX).\n",
941 (unsigned)(rdev->mc.gtt_size >> 20),
942 (unsigned long long)rdev->gart.table_addr);
943 rdev->gart.ready = true;
944 return 0;
945}
946
947static void rv770_pcie_gart_disable(struct radeon_device *rdev)
948{
949 u32 tmp;
950 int i;
951
952 /* Disable all tables */
953 for (i = 0; i < 7; i++)
954 WREG32(VM_CONTEXT0_CNTL + (i * 4), 0);
955
956 /* Setup L2 cache */
957 WREG32(VM_L2_CNTL, ENABLE_L2_FRAGMENT_PROCESSING |
958 EFFECTIVE_L2_QUEUE_SIZE(7));
959 WREG32(VM_L2_CNTL2, 0);
960 WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
961 /* Setup TLB control */
962 tmp = EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
963 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
964 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
965 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
966 WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
967 WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
968 WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
969 WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
970 radeon_gart_table_vram_unpin(rdev);
971}
972
973static void rv770_pcie_gart_fini(struct radeon_device *rdev)
974{
975 radeon_gart_fini(rdev);
976 rv770_pcie_gart_disable(rdev);
977 radeon_gart_table_vram_free(rdev);
978}
979
980
981static void rv770_agp_enable(struct radeon_device *rdev)
982{
983 u32 tmp;
984 int i;
985
986 /* Setup L2 cache */
987 WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
988 ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
989 EFFECTIVE_L2_QUEUE_SIZE(7));
990 WREG32(VM_L2_CNTL2, 0);
991 WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
992 /* Setup TLB control */
993 tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
994 SYSTEM_ACCESS_MODE_NOT_IN_SYS |
995 SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
996 EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
997 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
998 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
999 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
1000 WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
1001 WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
1002 WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
1003 WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
1004 for (i = 0; i < 7; i++)
1005 WREG32(VM_CONTEXT0_CNTL + (i * 4), 0);
1006}
1007
1008static void rv770_mc_program(struct radeon_device *rdev)
1009{
1010 struct rv515_mc_save save;
1011 u32 tmp;
1012 int i, j;
1013
1014 /* Initialize HDP */
1015 for (i = 0, j = 0; i < 32; i++, j += 0x18) {
1016 WREG32((0x2c14 + j), 0x00000000);
1017 WREG32((0x2c18 + j), 0x00000000);
1018 WREG32((0x2c1c + j), 0x00000000);
1019 WREG32((0x2c20 + j), 0x00000000);
1020 WREG32((0x2c24 + j), 0x00000000);
1021 }
1022 /* r7xx hw bug. Read from HDP_DEBUG1 rather
1023 * than writing to HDP_REG_COHERENCY_FLUSH_CNTL
1024 */
1025 tmp = RREG32(HDP_DEBUG1);
1026
1027 rv515_mc_stop(rdev, &save);
1028 if (r600_mc_wait_for_idle(rdev)) {
1029 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
1030 }
1031 /* Lockout access through VGA aperture*/
1032 WREG32(VGA_HDP_CONTROL, VGA_MEMORY_DISABLE);
1033 /* Update configuration */
1034 if (rdev->flags & RADEON_IS_AGP) {
1035 if (rdev->mc.vram_start < rdev->mc.gtt_start) {
1036 /* VRAM before AGP */
1037 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
1038 rdev->mc.vram_start >> 12);
1039 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
1040 rdev->mc.gtt_end >> 12);
1041 } else {
1042 /* VRAM after AGP */
1043 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
1044 rdev->mc.gtt_start >> 12);
1045 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
1046 rdev->mc.vram_end >> 12);
1047 }
1048 } else {
1049 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
1050 rdev->mc.vram_start >> 12);
1051 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
1052 rdev->mc.vram_end >> 12);
1053 }
1054 WREG32(MC_VM_SYSTEM_APERTURE_DEFAULT_ADDR, rdev->vram_scratch.gpu_addr >> 12);
1055 tmp = ((rdev->mc.vram_end >> 24) & 0xFFFF) << 16;
1056 tmp |= ((rdev->mc.vram_start >> 24) & 0xFFFF);
1057 WREG32(MC_VM_FB_LOCATION, tmp);
1058 WREG32(HDP_NONSURFACE_BASE, (rdev->mc.vram_start >> 8));
1059 WREG32(HDP_NONSURFACE_INFO, (2 << 7));
1060 WREG32(HDP_NONSURFACE_SIZE, 0x3FFFFFFF);
1061 if (rdev->flags & RADEON_IS_AGP) {
1062 WREG32(MC_VM_AGP_TOP, rdev->mc.gtt_end >> 16);
1063 WREG32(MC_VM_AGP_BOT, rdev->mc.gtt_start >> 16);
1064 WREG32(MC_VM_AGP_BASE, rdev->mc.agp_base >> 22);
1065 } else {
1066 WREG32(MC_VM_AGP_BASE, 0);
1067 WREG32(MC_VM_AGP_TOP, 0x0FFFFFFF);
1068 WREG32(MC_VM_AGP_BOT, 0x0FFFFFFF);
1069 }
1070 if (r600_mc_wait_for_idle(rdev)) {
1071 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
1072 }
1073 rv515_mc_resume(rdev, &save);
1074 /* we need to own VRAM, so turn off the VGA renderer here
1075 * to stop it overwriting our objects */
1076 rv515_vga_render_disable(rdev);
1077}
1078
1079
1080/*
1081 * CP.
1082 */
1083void r700_cp_stop(struct radeon_device *rdev)
1084{
1085 if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
1086 radeon_ttm_set_active_vram_size(rdev, rdev->mc.visible_vram_size);
1087 WREG32(CP_ME_CNTL, (CP_ME_HALT | CP_PFP_HALT));
1088 WREG32(SCRATCH_UMSK, 0);
1089 rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ready = false;
1090}
1091
1092static int rv770_cp_load_microcode(struct radeon_device *rdev)
1093{
1094 const __be32 *fw_data;
1095 int i;
1096
1097 if (!rdev->me_fw || !rdev->pfp_fw)
1098 return -EINVAL;
1099
1100 r700_cp_stop(rdev);
1101 WREG32(CP_RB_CNTL,
1102#ifdef __BIG_ENDIAN
1103 BUF_SWAP_32BIT |
1104#endif
1105 RB_NO_UPDATE | RB_BLKSZ(15) | RB_BUFSZ(3));
1106
1107 /* Reset cp */
1108 WREG32(GRBM_SOFT_RESET, SOFT_RESET_CP);
1109 RREG32(GRBM_SOFT_RESET);
1110 mdelay(15);
1111 WREG32(GRBM_SOFT_RESET, 0);
1112
1113 fw_data = (const __be32 *)rdev->pfp_fw->data;
1114 WREG32(CP_PFP_UCODE_ADDR, 0);
1115 for (i = 0; i < R700_PFP_UCODE_SIZE; i++)
1116 WREG32(CP_PFP_UCODE_DATA, be32_to_cpup(fw_data++));
1117 WREG32(CP_PFP_UCODE_ADDR, 0);
1118
1119 fw_data = (const __be32 *)rdev->me_fw->data;
1120 WREG32(CP_ME_RAM_WADDR, 0);
1121 for (i = 0; i < R700_PM4_UCODE_SIZE; i++)
1122 WREG32(CP_ME_RAM_DATA, be32_to_cpup(fw_data++));
1123
1124 WREG32(CP_PFP_UCODE_ADDR, 0);
1125 WREG32(CP_ME_RAM_WADDR, 0);
1126 WREG32(CP_ME_RAM_RADDR, 0);
1127 return 0;
1128}
1129
1130void r700_cp_fini(struct radeon_device *rdev)
1131{
1132 struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
1133 r700_cp_stop(rdev);
1134 radeon_ring_fini(rdev, ring);
1135 radeon_scratch_free(rdev, ring->rptr_save_reg);
1136}
1137
1138void rv770_set_clk_bypass_mode(struct radeon_device *rdev)
1139{
1140 u32 tmp, i;
1141
1142 if (rdev->flags & RADEON_IS_IGP)
1143 return;
1144
1145 tmp = RREG32(CG_SPLL_FUNC_CNTL_2);
1146 tmp &= SCLK_MUX_SEL_MASK;
1147 tmp |= SCLK_MUX_SEL(1) | SCLK_MUX_UPDATE;
1148 WREG32(CG_SPLL_FUNC_CNTL_2, tmp);
1149
1150 for (i = 0; i < rdev->usec_timeout; i++) {
1151 if (RREG32(CG_SPLL_STATUS) & SPLL_CHG_STATUS)
1152 break;
1153 udelay(1);
1154 }
1155
1156 tmp &= ~SCLK_MUX_UPDATE;
1157 WREG32(CG_SPLL_FUNC_CNTL_2, tmp);
1158
1159 tmp = RREG32(MPLL_CNTL_MODE);
1160 if ((rdev->family == CHIP_RV710) || (rdev->family == CHIP_RV730))
1161 tmp &= ~RV730_MPLL_MCLK_SEL;
1162 else
1163 tmp &= ~MPLL_MCLK_SEL;
1164 WREG32(MPLL_CNTL_MODE, tmp);
1165}
1166
1167/*
1168 * Core functions
1169 */
1170static void rv770_gpu_init(struct radeon_device *rdev)
1171{
1172 int i, j, num_qd_pipes;
1173 u32 ta_aux_cntl;
1174 u32 sx_debug_1;
1175 u32 smx_dc_ctl0;
1176 u32 db_debug3;
1177 u32 num_gs_verts_per_thread;
1178 u32 vgt_gs_per_es;
1179 u32 gs_prim_buffer_depth = 0;
1180 u32 sq_ms_fifo_sizes;
1181 u32 sq_config;
1182 u32 sq_thread_resource_mgmt;
1183 u32 hdp_host_path_cntl;
1184 u32 sq_dyn_gpr_size_simd_ab_0;
1185 u32 gb_tiling_config = 0;
1186 u32 cc_gc_shader_pipe_config = 0;
1187 u32 mc_arb_ramcfg;
1188 u32 db_debug4, tmp;
1189 u32 inactive_pipes, shader_pipe_config;
1190 u32 disabled_rb_mask;
1191 unsigned active_number;
1192
1193 /* setup chip specs */
1194 rdev->config.rv770.tiling_group_size = 256;
1195 switch (rdev->family) {
1196 case CHIP_RV770:
1197 rdev->config.rv770.max_pipes = 4;
1198 rdev->config.rv770.max_tile_pipes = 8;
1199 rdev->config.rv770.max_simds = 10;
1200 rdev->config.rv770.max_backends = 4;
1201 rdev->config.rv770.max_gprs = 256;
1202 rdev->config.rv770.max_threads = 248;
1203 rdev->config.rv770.max_stack_entries = 512;
1204 rdev->config.rv770.max_hw_contexts = 8;
1205 rdev->config.rv770.max_gs_threads = 16 * 2;
1206 rdev->config.rv770.sx_max_export_size = 128;
1207 rdev->config.rv770.sx_max_export_pos_size = 16;
1208 rdev->config.rv770.sx_max_export_smx_size = 112;
1209 rdev->config.rv770.sq_num_cf_insts = 2;
1210
1211 rdev->config.rv770.sx_num_of_sets = 7;
1212 rdev->config.rv770.sc_prim_fifo_size = 0xF9;
1213 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
1214 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
1215 break;
1216 case CHIP_RV730:
1217 rdev->config.rv770.max_pipes = 2;
1218 rdev->config.rv770.max_tile_pipes = 4;
1219 rdev->config.rv770.max_simds = 8;
1220 rdev->config.rv770.max_backends = 2;
1221 rdev->config.rv770.max_gprs = 128;
1222 rdev->config.rv770.max_threads = 248;
1223 rdev->config.rv770.max_stack_entries = 256;
1224 rdev->config.rv770.max_hw_contexts = 8;
1225 rdev->config.rv770.max_gs_threads = 16 * 2;
1226 rdev->config.rv770.sx_max_export_size = 256;
1227 rdev->config.rv770.sx_max_export_pos_size = 32;
1228 rdev->config.rv770.sx_max_export_smx_size = 224;
1229 rdev->config.rv770.sq_num_cf_insts = 2;
1230
1231 rdev->config.rv770.sx_num_of_sets = 7;
1232 rdev->config.rv770.sc_prim_fifo_size = 0xf9;
1233 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
1234 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
1235 if (rdev->config.rv770.sx_max_export_pos_size > 16) {
1236 rdev->config.rv770.sx_max_export_pos_size -= 16;
1237 rdev->config.rv770.sx_max_export_smx_size += 16;
1238 }
1239 break;
1240 case CHIP_RV710:
1241 rdev->config.rv770.max_pipes = 2;
1242 rdev->config.rv770.max_tile_pipes = 2;
1243 rdev->config.rv770.max_simds = 2;
1244 rdev->config.rv770.max_backends = 1;
1245 rdev->config.rv770.max_gprs = 256;
1246 rdev->config.rv770.max_threads = 192;
1247 rdev->config.rv770.max_stack_entries = 256;
1248 rdev->config.rv770.max_hw_contexts = 4;
1249 rdev->config.rv770.max_gs_threads = 8 * 2;
1250 rdev->config.rv770.sx_max_export_size = 128;
1251 rdev->config.rv770.sx_max_export_pos_size = 16;
1252 rdev->config.rv770.sx_max_export_smx_size = 112;
1253 rdev->config.rv770.sq_num_cf_insts = 1;
1254
1255 rdev->config.rv770.sx_num_of_sets = 7;
1256 rdev->config.rv770.sc_prim_fifo_size = 0x40;
1257 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
1258 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
1259 break;
1260 case CHIP_RV740:
1261 rdev->config.rv770.max_pipes = 4;
1262 rdev->config.rv770.max_tile_pipes = 4;
1263 rdev->config.rv770.max_simds = 8;
1264 rdev->config.rv770.max_backends = 4;
1265 rdev->config.rv770.max_gprs = 256;
1266 rdev->config.rv770.max_threads = 248;
1267 rdev->config.rv770.max_stack_entries = 512;
1268 rdev->config.rv770.max_hw_contexts = 8;
1269 rdev->config.rv770.max_gs_threads = 16 * 2;
1270 rdev->config.rv770.sx_max_export_size = 256;
1271 rdev->config.rv770.sx_max_export_pos_size = 32;
1272 rdev->config.rv770.sx_max_export_smx_size = 224;
1273 rdev->config.rv770.sq_num_cf_insts = 2;
1274
1275 rdev->config.rv770.sx_num_of_sets = 7;
1276 rdev->config.rv770.sc_prim_fifo_size = 0x100;
1277 rdev->config.rv770.sc_hiz_tile_fifo_size = 0x30;
1278 rdev->config.rv770.sc_earlyz_tile_fifo_fize = 0x130;
1279
1280 if (rdev->config.rv770.sx_max_export_pos_size > 16) {
1281 rdev->config.rv770.sx_max_export_pos_size -= 16;
1282 rdev->config.rv770.sx_max_export_smx_size += 16;
1283 }
1284 break;
1285 default:
1286 break;
1287 }
1288
1289 /* Initialize HDP */
1290 j = 0;
1291 for (i = 0; i < 32; i++) {
1292 WREG32((0x2c14 + j), 0x00000000);
1293 WREG32((0x2c18 + j), 0x00000000);
1294 WREG32((0x2c1c + j), 0x00000000);
1295 WREG32((0x2c20 + j), 0x00000000);
1296 WREG32((0x2c24 + j), 0x00000000);
1297 j += 0x18;
1298 }
1299
1300 WREG32(GRBM_CNTL, GRBM_READ_TIMEOUT(0xff));
1301
1302 /* setup tiling, simd, pipe config */
1303 mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG);
1304
1305 shader_pipe_config = RREG32(CC_GC_SHADER_PIPE_CONFIG);
1306 inactive_pipes = (shader_pipe_config & INACTIVE_QD_PIPES_MASK) >> INACTIVE_QD_PIPES_SHIFT;
1307 for (i = 0, tmp = 1, active_number = 0; i < R7XX_MAX_PIPES; i++) {
1308 if (!(inactive_pipes & tmp)) {
1309 active_number++;
1310 }
1311 tmp <<= 1;
1312 }
1313 if (active_number == 1) {
1314 WREG32(SPI_CONFIG_CNTL, DISABLE_INTERP_1);
1315 } else {
1316 WREG32(SPI_CONFIG_CNTL, 0);
1317 }
1318
1319 cc_gc_shader_pipe_config = RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffffff00;
1320 tmp = rdev->config.rv770.max_simds -
1321 r600_count_pipe_bits((cc_gc_shader_pipe_config >> 16) & R7XX_MAX_SIMDS_MASK);
1322 rdev->config.rv770.active_simds = tmp;
1323
1324 switch (rdev->config.rv770.max_tile_pipes) {
1325 case 1:
1326 default:
1327 gb_tiling_config = PIPE_TILING(0);
1328 break;
1329 case 2:
1330 gb_tiling_config = PIPE_TILING(1);
1331 break;
1332 case 4:
1333 gb_tiling_config = PIPE_TILING(2);
1334 break;
1335 case 8:
1336 gb_tiling_config = PIPE_TILING(3);
1337 break;
1338 }
1339 rdev->config.rv770.tiling_npipes = rdev->config.rv770.max_tile_pipes;
1340
1341 disabled_rb_mask = (RREG32(CC_RB_BACKEND_DISABLE) >> 16) & R7XX_MAX_BACKENDS_MASK;
1342 tmp = 0;
1343 for (i = 0; i < rdev->config.rv770.max_backends; i++)
1344 tmp |= (1 << i);
1345 /* if all the backends are disabled, fix it up here */
1346 if ((disabled_rb_mask & tmp) == tmp) {
1347 for (i = 0; i < rdev->config.rv770.max_backends; i++)
1348 disabled_rb_mask &= ~(1 << i);
1349 }
1350 tmp = (gb_tiling_config & PIPE_TILING__MASK) >> PIPE_TILING__SHIFT;
1351 tmp = r6xx_remap_render_backend(rdev, tmp, rdev->config.rv770.max_backends,
1352 R7XX_MAX_BACKENDS, disabled_rb_mask);
1353 gb_tiling_config |= tmp << 16;
1354 rdev->config.rv770.backend_map = tmp;
1355
1356 if (rdev->family == CHIP_RV770)
1357 gb_tiling_config |= BANK_TILING(1);
1358 else {
1359 if ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT)
1360 gb_tiling_config |= BANK_TILING(1);
1361 else
1362 gb_tiling_config |= BANK_TILING(0);
1363 }
1364 rdev->config.rv770.tiling_nbanks = 4 << ((gb_tiling_config >> 4) & 0x3);
1365 gb_tiling_config |= GROUP_SIZE((mc_arb_ramcfg & BURSTLENGTH_MASK) >> BURSTLENGTH_SHIFT);
1366 if (((mc_arb_ramcfg & NOOFROWS_MASK) >> NOOFROWS_SHIFT) > 3) {
1367 gb_tiling_config |= ROW_TILING(3);
1368 gb_tiling_config |= SAMPLE_SPLIT(3);
1369 } else {
1370 gb_tiling_config |=
1371 ROW_TILING(((mc_arb_ramcfg & NOOFROWS_MASK) >> NOOFROWS_SHIFT));
1372 gb_tiling_config |=
1373 SAMPLE_SPLIT(((mc_arb_ramcfg & NOOFROWS_MASK) >> NOOFROWS_SHIFT));
1374 }
1375
1376 gb_tiling_config |= BANK_SWAPS(1);
1377 rdev->config.rv770.tile_config = gb_tiling_config;
1378
1379 WREG32(GB_TILING_CONFIG, gb_tiling_config);
1380 WREG32(DCP_TILING_CONFIG, (gb_tiling_config & 0xffff));
1381 WREG32(HDP_TILING_CONFIG, (gb_tiling_config & 0xffff));
1382 WREG32(DMA_TILING_CONFIG, (gb_tiling_config & 0xffff));
1383 WREG32(DMA_TILING_CONFIG2, (gb_tiling_config & 0xffff));
1384 if (rdev->family == CHIP_RV730) {
1385 WREG32(UVD_UDEC_DB_TILING_CONFIG, (gb_tiling_config & 0xffff));
1386 WREG32(UVD_UDEC_DBW_TILING_CONFIG, (gb_tiling_config & 0xffff));
1387 WREG32(UVD_UDEC_TILING_CONFIG, (gb_tiling_config & 0xffff));
1388 }
1389
1390 WREG32(CGTS_SYS_TCC_DISABLE, 0);
1391 WREG32(CGTS_TCC_DISABLE, 0);
1392 WREG32(CGTS_USER_SYS_TCC_DISABLE, 0);
1393 WREG32(CGTS_USER_TCC_DISABLE, 0);
1394
1395
1396 num_qd_pipes = R7XX_MAX_PIPES - r600_count_pipe_bits((cc_gc_shader_pipe_config & INACTIVE_QD_PIPES_MASK) >> 8);
1397 WREG32(VGT_OUT_DEALLOC_CNTL, (num_qd_pipes * 4) & DEALLOC_DIST_MASK);
1398 WREG32(VGT_VERTEX_REUSE_BLOCK_CNTL, ((num_qd_pipes * 4) - 2) & VTX_REUSE_DEPTH_MASK);
1399
1400 /* set HW defaults for 3D engine */
1401 WREG32(CP_QUEUE_THRESHOLDS, (ROQ_IB1_START(0x16) |
1402 ROQ_IB2_START(0x2b)));
1403
1404 WREG32(CP_MEQ_THRESHOLDS, STQ_SPLIT(0x30));
1405
1406 ta_aux_cntl = RREG32(TA_CNTL_AUX);
1407 WREG32(TA_CNTL_AUX, ta_aux_cntl | DISABLE_CUBE_ANISO);
1408
1409 sx_debug_1 = RREG32(SX_DEBUG_1);
1410 sx_debug_1 |= ENABLE_NEW_SMX_ADDRESS;
1411 WREG32(SX_DEBUG_1, sx_debug_1);
1412
1413 smx_dc_ctl0 = RREG32(SMX_DC_CTL0);
1414 smx_dc_ctl0 &= ~CACHE_DEPTH(0x1ff);
1415 smx_dc_ctl0 |= CACHE_DEPTH((rdev->config.rv770.sx_num_of_sets * 64) - 1);
1416 WREG32(SMX_DC_CTL0, smx_dc_ctl0);
1417
1418 if (rdev->family != CHIP_RV740)
1419 WREG32(SMX_EVENT_CTL, (ES_FLUSH_CTL(4) |
1420 GS_FLUSH_CTL(4) |
1421 ACK_FLUSH_CTL(3) |
1422 SYNC_FLUSH_CTL));
1423
1424 if (rdev->family != CHIP_RV770)
1425 WREG32(SMX_SAR_CTL0, 0x00003f3f);
1426
1427 db_debug3 = RREG32(DB_DEBUG3);
1428 db_debug3 &= ~DB_CLK_OFF_DELAY(0x1f);
1429 switch (rdev->family) {
1430 case CHIP_RV770:
1431 case CHIP_RV740:
1432 db_debug3 |= DB_CLK_OFF_DELAY(0x1f);
1433 break;
1434 case CHIP_RV710:
1435 case CHIP_RV730:
1436 default:
1437 db_debug3 |= DB_CLK_OFF_DELAY(2);
1438 break;
1439 }
1440 WREG32(DB_DEBUG3, db_debug3);
1441
1442 if (rdev->family != CHIP_RV770) {
1443 db_debug4 = RREG32(DB_DEBUG4);
1444 db_debug4 |= DISABLE_TILE_COVERED_FOR_PS_ITER;
1445 WREG32(DB_DEBUG4, db_debug4);
1446 }
1447
1448 WREG32(SX_EXPORT_BUFFER_SIZES, (COLOR_BUFFER_SIZE((rdev->config.rv770.sx_max_export_size / 4) - 1) |
1449 POSITION_BUFFER_SIZE((rdev->config.rv770.sx_max_export_pos_size / 4) - 1) |
1450 SMX_BUFFER_SIZE((rdev->config.rv770.sx_max_export_smx_size / 4) - 1)));
1451
1452 WREG32(PA_SC_FIFO_SIZE, (SC_PRIM_FIFO_SIZE(rdev->config.rv770.sc_prim_fifo_size) |
1453 SC_HIZ_TILE_FIFO_SIZE(rdev->config.rv770.sc_hiz_tile_fifo_size) |
1454 SC_EARLYZ_TILE_FIFO_SIZE(rdev->config.rv770.sc_earlyz_tile_fifo_fize)));
1455
1456 WREG32(PA_SC_MULTI_CHIP_CNTL, 0);
1457
1458 WREG32(VGT_NUM_INSTANCES, 1);
1459
1460 WREG32(SPI_CONFIG_CNTL_1, VTX_DONE_DELAY(4));
1461
1462 WREG32(CP_PERFMON_CNTL, 0);
1463
1464 sq_ms_fifo_sizes = (CACHE_FIFO_SIZE(16 * rdev->config.rv770.sq_num_cf_insts) |
1465 DONE_FIFO_HIWATER(0xe0) |
1466 ALU_UPDATE_FIFO_HIWATER(0x8));
1467 switch (rdev->family) {
1468 case CHIP_RV770:
1469 case CHIP_RV730:
1470 case CHIP_RV710:
1471 sq_ms_fifo_sizes |= FETCH_FIFO_HIWATER(0x1);
1472 break;
1473 case CHIP_RV740:
1474 default:
1475 sq_ms_fifo_sizes |= FETCH_FIFO_HIWATER(0x4);
1476 break;
1477 }
1478 WREG32(SQ_MS_FIFO_SIZES, sq_ms_fifo_sizes);
1479
1480 /* SQ_CONFIG, SQ_GPR_RESOURCE_MGMT, SQ_THREAD_RESOURCE_MGMT, SQ_STACK_RESOURCE_MGMT
1481 * should be adjusted as needed by the 2D/3D drivers. This just sets default values
1482 */
1483 sq_config = RREG32(SQ_CONFIG);
1484 sq_config &= ~(PS_PRIO(3) |
1485 VS_PRIO(3) |
1486 GS_PRIO(3) |
1487 ES_PRIO(3));
1488 sq_config |= (DX9_CONSTS |
1489 VC_ENABLE |
1490 EXPORT_SRC_C |
1491 PS_PRIO(0) |
1492 VS_PRIO(1) |
1493 GS_PRIO(2) |
1494 ES_PRIO(3));
1495 if (rdev->family == CHIP_RV710)
1496 /* no vertex cache */
1497 sq_config &= ~VC_ENABLE;
1498
1499 WREG32(SQ_CONFIG, sq_config);
1500
1501 WREG32(SQ_GPR_RESOURCE_MGMT_1, (NUM_PS_GPRS((rdev->config.rv770.max_gprs * 24)/64) |
1502 NUM_VS_GPRS((rdev->config.rv770.max_gprs * 24)/64) |
1503 NUM_CLAUSE_TEMP_GPRS(((rdev->config.rv770.max_gprs * 24)/64)/2)));
1504
1505 WREG32(SQ_GPR_RESOURCE_MGMT_2, (NUM_GS_GPRS((rdev->config.rv770.max_gprs * 7)/64) |
1506 NUM_ES_GPRS((rdev->config.rv770.max_gprs * 7)/64)));
1507
1508 sq_thread_resource_mgmt = (NUM_PS_THREADS((rdev->config.rv770.max_threads * 4)/8) |
1509 NUM_VS_THREADS((rdev->config.rv770.max_threads * 2)/8) |
1510 NUM_ES_THREADS((rdev->config.rv770.max_threads * 1)/8));
1511 if (((rdev->config.rv770.max_threads * 1) / 8) > rdev->config.rv770.max_gs_threads)
1512 sq_thread_resource_mgmt |= NUM_GS_THREADS(rdev->config.rv770.max_gs_threads);
1513 else
1514 sq_thread_resource_mgmt |= NUM_GS_THREADS((rdev->config.rv770.max_gs_threads * 1)/8);
1515 WREG32(SQ_THREAD_RESOURCE_MGMT, sq_thread_resource_mgmt);
1516
1517 WREG32(SQ_STACK_RESOURCE_MGMT_1, (NUM_PS_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4) |
1518 NUM_VS_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4)));
1519
1520 WREG32(SQ_STACK_RESOURCE_MGMT_2, (NUM_GS_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4) |
1521 NUM_ES_STACK_ENTRIES((rdev->config.rv770.max_stack_entries * 1)/4)));
1522
1523 sq_dyn_gpr_size_simd_ab_0 = (SIMDA_RING0((rdev->config.rv770.max_gprs * 38)/64) |
1524 SIMDA_RING1((rdev->config.rv770.max_gprs * 38)/64) |
1525 SIMDB_RING0((rdev->config.rv770.max_gprs * 38)/64) |
1526 SIMDB_RING1((rdev->config.rv770.max_gprs * 38)/64));
1527
1528 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_0, sq_dyn_gpr_size_simd_ab_0);
1529 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_1, sq_dyn_gpr_size_simd_ab_0);
1530 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_2, sq_dyn_gpr_size_simd_ab_0);
1531 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_3, sq_dyn_gpr_size_simd_ab_0);
1532 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_4, sq_dyn_gpr_size_simd_ab_0);
1533 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_5, sq_dyn_gpr_size_simd_ab_0);
1534 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_6, sq_dyn_gpr_size_simd_ab_0);
1535 WREG32(SQ_DYN_GPR_SIZE_SIMD_AB_7, sq_dyn_gpr_size_simd_ab_0);
1536
1537 WREG32(PA_SC_FORCE_EOV_MAX_CNTS, (FORCE_EOV_MAX_CLK_CNT(4095) |
1538 FORCE_EOV_MAX_REZ_CNT(255)));
1539
1540 if (rdev->family == CHIP_RV710)
1541 WREG32(VGT_CACHE_INVALIDATION, (CACHE_INVALIDATION(TC_ONLY) |
1542 AUTO_INVLD_EN(ES_AND_GS_AUTO)));
1543 else
1544 WREG32(VGT_CACHE_INVALIDATION, (CACHE_INVALIDATION(VC_AND_TC) |
1545 AUTO_INVLD_EN(ES_AND_GS_AUTO)));
1546
1547 switch (rdev->family) {
1548 case CHIP_RV770:
1549 case CHIP_RV730:
1550 case CHIP_RV740:
1551 gs_prim_buffer_depth = 384;
1552 break;
1553 case CHIP_RV710:
1554 gs_prim_buffer_depth = 128;
1555 break;
1556 default:
1557 break;
1558 }
1559
1560 num_gs_verts_per_thread = rdev->config.rv770.max_pipes * 16;
1561 vgt_gs_per_es = gs_prim_buffer_depth + num_gs_verts_per_thread;
1562 /* Max value for this is 256 */
1563 if (vgt_gs_per_es > 256)
1564 vgt_gs_per_es = 256;
1565
1566 WREG32(VGT_ES_PER_GS, 128);
1567 WREG32(VGT_GS_PER_ES, vgt_gs_per_es);
1568 WREG32(VGT_GS_PER_VS, 2);
1569
1570 /* more default values. 2D/3D driver should adjust as needed */
1571 WREG32(VGT_GS_VERTEX_REUSE, 16);
1572 WREG32(PA_SC_LINE_STIPPLE_STATE, 0);
1573 WREG32(VGT_STRMOUT_EN, 0);
1574 WREG32(SX_MISC, 0);
1575 WREG32(PA_SC_MODE_CNTL, 0);
1576 WREG32(PA_SC_EDGERULE, 0xaaaaaaaa);
1577 WREG32(PA_SC_AA_CONFIG, 0);
1578 WREG32(PA_SC_CLIPRECT_RULE, 0xffff);
1579 WREG32(PA_SC_LINE_STIPPLE, 0);
1580 WREG32(SPI_INPUT_Z, 0);
1581 WREG32(SPI_PS_IN_CONTROL_0, NUM_INTERP(2));
1582 WREG32(CB_COLOR7_FRAG, 0);
1583
1584 /* clear render buffer base addresses */
1585 WREG32(CB_COLOR0_BASE, 0);
1586 WREG32(CB_COLOR1_BASE, 0);
1587 WREG32(CB_COLOR2_BASE, 0);
1588 WREG32(CB_COLOR3_BASE, 0);
1589 WREG32(CB_COLOR4_BASE, 0);
1590 WREG32(CB_COLOR5_BASE, 0);
1591 WREG32(CB_COLOR6_BASE, 0);
1592 WREG32(CB_COLOR7_BASE, 0);
1593
1594 WREG32(TCP_CNTL, 0);
1595
1596 hdp_host_path_cntl = RREG32(HDP_HOST_PATH_CNTL);
1597 WREG32(HDP_HOST_PATH_CNTL, hdp_host_path_cntl);
1598
1599 WREG32(PA_SC_MULTI_CHIP_CNTL, 0);
1600
1601 WREG32(PA_CL_ENHANCE, (CLIP_VTX_REORDER_ENA |
1602 NUM_CLIP_SEQ(3)));
1603 WREG32(VC_ENHANCE, 0);
1604}
1605
1606void r700_vram_gtt_location(struct radeon_device *rdev, struct radeon_mc *mc)
1607{
1608 u64 size_bf, size_af;
1609
1610 if (mc->mc_vram_size > 0xE0000000) {
1611 /* leave room for at least 512M GTT */
1612 dev_warn(rdev->dev, "limiting VRAM\n");
1613 mc->real_vram_size = 0xE0000000;
1614 mc->mc_vram_size = 0xE0000000;
1615 }
1616 if (rdev->flags & RADEON_IS_AGP) {
1617 size_bf = mc->gtt_start;
1618 size_af = mc->mc_mask - mc->gtt_end;
1619 if (size_bf > size_af) {
1620 if (mc->mc_vram_size > size_bf) {
1621 dev_warn(rdev->dev, "limiting VRAM\n");
1622 mc->real_vram_size = size_bf;
1623 mc->mc_vram_size = size_bf;
1624 }
1625 mc->vram_start = mc->gtt_start - mc->mc_vram_size;
1626 } else {
1627 if (mc->mc_vram_size > size_af) {
1628 dev_warn(rdev->dev, "limiting VRAM\n");
1629 mc->real_vram_size = size_af;
1630 mc->mc_vram_size = size_af;
1631 }
1632 mc->vram_start = mc->gtt_end + 1;
1633 }
1634 mc->vram_end = mc->vram_start + mc->mc_vram_size - 1;
1635 dev_info(rdev->dev, "VRAM: %lluM 0x%08llX - 0x%08llX (%lluM used)\n",
1636 mc->mc_vram_size >> 20, mc->vram_start,
1637 mc->vram_end, mc->real_vram_size >> 20);
1638 } else {
1639 radeon_vram_location(rdev, &rdev->mc, 0);
1640 rdev->mc.gtt_base_align = 0;
1641 radeon_gtt_location(rdev, mc);
1642 }
1643}
1644
1645static int rv770_mc_init(struct radeon_device *rdev)
1646{
1647 u32 tmp;
1648 int chansize, numchan;
1649
1650 /* Get VRAM informations */
1651 rdev->mc.vram_is_ddr = true;
1652 tmp = RREG32(MC_ARB_RAMCFG);
1653 if (tmp & CHANSIZE_OVERRIDE) {
1654 chansize = 16;
1655 } else if (tmp & CHANSIZE_MASK) {
1656 chansize = 64;
1657 } else {
1658 chansize = 32;
1659 }
1660 tmp = RREG32(MC_SHARED_CHMAP);
1661 switch ((tmp & NOOFCHAN_MASK) >> NOOFCHAN_SHIFT) {
1662 case 0:
1663 default:
1664 numchan = 1;
1665 break;
1666 case 1:
1667 numchan = 2;
1668 break;
1669 case 2:
1670 numchan = 4;
1671 break;
1672 case 3:
1673 numchan = 8;
1674 break;
1675 }
1676 rdev->mc.vram_width = numchan * chansize;
1677 /* Could aper size report 0 ? */
1678 rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0);
1679 rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0);
1680 /* Setup GPU memory space */
1681 rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE);
1682 rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE);
1683 rdev->mc.visible_vram_size = rdev->mc.aper_size;
1684 r700_vram_gtt_location(rdev, &rdev->mc);
1685 radeon_update_bandwidth_info(rdev);
1686
1687 return 0;
1688}
1689
1690static void rv770_uvd_init(struct radeon_device *rdev)
1691{
1692 int r;
1693
1694 if (!rdev->has_uvd)
1695 return;
1696
1697 r = radeon_uvd_init(rdev);
1698 if (r) {
1699 dev_err(rdev->dev, "failed UVD (%d) init.\n", r);
1700 /*
1701 * At this point rdev->uvd.vcpu_bo is NULL which trickles down
1702 * to early fails uvd_v2_2_resume() and thus nothing happens
1703 * there. So it is pointless to try to go through that code
1704 * hence why we disable uvd here.
1705 */
1706 rdev->has_uvd = false;
1707 return;
1708 }
1709 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL;
1710 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096);
1711}
1712
1713static void rv770_uvd_start(struct radeon_device *rdev)
1714{
1715 int r;
1716
1717 if (!rdev->has_uvd)
1718 return;
1719
1720 r = uvd_v2_2_resume(rdev);
1721 if (r) {
1722 dev_err(rdev->dev, "failed UVD resume (%d).\n", r);
1723 goto error;
1724 }
1725 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX);
1726 if (r) {
1727 dev_err(rdev->dev, "failed initializing UVD fences (%d).\n", r);
1728 goto error;
1729 }
1730 return;
1731
1732error:
1733 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0;
1734}
1735
1736static void rv770_uvd_resume(struct radeon_device *rdev)
1737{
1738 struct radeon_ring *ring;
1739 int r;
1740
1741 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size)
1742 return;
1743
1744 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
1745 r = radeon_ring_init(rdev, ring, ring->ring_size, 0, PACKET0(UVD_NO_OP, 0));
1746 if (r) {
1747 dev_err(rdev->dev, "failed initializing UVD ring (%d).\n", r);
1748 return;
1749 }
1750 r = uvd_v1_0_init(rdev);
1751 if (r) {
1752 dev_err(rdev->dev, "failed initializing UVD (%d).\n", r);
1753 return;
1754 }
1755}
1756
1757static int rv770_startup(struct radeon_device *rdev)
1758{
1759 struct radeon_ring *ring;
1760 int r;
1761
1762 /* enable pcie gen2 link */
1763 rv770_pcie_gen2_enable(rdev);
1764
1765 /* scratch needs to be initialized before MC */
1766 r = r600_vram_scratch_init(rdev);
1767 if (r)
1768 return r;
1769
1770 rv770_mc_program(rdev);
1771
1772 if (rdev->flags & RADEON_IS_AGP) {
1773 rv770_agp_enable(rdev);
1774 } else {
1775 r = rv770_pcie_gart_enable(rdev);
1776 if (r)
1777 return r;
1778 }
1779
1780 rv770_gpu_init(rdev);
1781
1782 /* allocate wb buffer */
1783 r = radeon_wb_init(rdev);
1784 if (r)
1785 return r;
1786
1787 r = radeon_fence_driver_start_ring(rdev, RADEON_RING_TYPE_GFX_INDEX);
1788 if (r) {
1789 dev_err(rdev->dev, "failed initializing CP fences (%d).\n", r);
1790 return r;
1791 }
1792
1793 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX);
1794 if (r) {
1795 dev_err(rdev->dev, "failed initializing DMA fences (%d).\n", r);
1796 return r;
1797 }
1798
1799 rv770_uvd_start(rdev);
1800
1801 /* Enable IRQ */
1802 if (!rdev->irq.installed) {
1803 r = radeon_irq_kms_init(rdev);
1804 if (r)
1805 return r;
1806 }
1807
1808 r = r600_irq_init(rdev);
1809 if (r) {
1810 DRM_ERROR("radeon: IH init failed (%d).\n", r);
1811 radeon_irq_kms_fini(rdev);
1812 return r;
1813 }
1814 r600_irq_set(rdev);
1815
1816 ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
1817 r = radeon_ring_init(rdev, ring, ring->ring_size, RADEON_WB_CP_RPTR_OFFSET,
1818 RADEON_CP_PACKET2);
1819 if (r)
1820 return r;
1821
1822 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX];
1823 r = radeon_ring_init(rdev, ring, ring->ring_size, R600_WB_DMA_RPTR_OFFSET,
1824 DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0));
1825 if (r)
1826 return r;
1827
1828 r = rv770_cp_load_microcode(rdev);
1829 if (r)
1830 return r;
1831 r = r600_cp_resume(rdev);
1832 if (r)
1833 return r;
1834
1835 r = r600_dma_resume(rdev);
1836 if (r)
1837 return r;
1838
1839 rv770_uvd_resume(rdev);
1840
1841 r = radeon_ib_pool_init(rdev);
1842 if (r) {
1843 dev_err(rdev->dev, "IB initialization failed (%d).\n", r);
1844 return r;
1845 }
1846
1847 r = radeon_audio_init(rdev);
1848 if (r) {
1849 DRM_ERROR("radeon: audio init failed\n");
1850 return r;
1851 }
1852
1853 return 0;
1854}
1855
1856int rv770_resume(struct radeon_device *rdev)
1857{
1858 int r;
1859
1860 /* Do not reset GPU before posting, on rv770 hw unlike on r500 hw,
1861 * posting will perform necessary task to bring back GPU into good
1862 * shape.
1863 */
1864 /* post card */
1865 atom_asic_init(rdev->mode_info.atom_context);
1866
1867 /* init golden registers */
1868 rv770_init_golden_registers(rdev);
1869
1870 if (rdev->pm.pm_method == PM_METHOD_DPM)
1871 radeon_pm_resume(rdev);
1872
1873 rdev->accel_working = true;
1874 r = rv770_startup(rdev);
1875 if (r) {
1876 DRM_ERROR("r600 startup failed on resume\n");
1877 rdev->accel_working = false;
1878 return r;
1879 }
1880
1881 return r;
1882
1883}
1884
1885int rv770_suspend(struct radeon_device *rdev)
1886{
1887 radeon_pm_suspend(rdev);
1888 radeon_audio_fini(rdev);
1889 if (rdev->has_uvd) {
1890 uvd_v1_0_fini(rdev);
1891 radeon_uvd_suspend(rdev);
1892 }
1893 r700_cp_stop(rdev);
1894 r600_dma_stop(rdev);
1895 r600_irq_suspend(rdev);
1896 radeon_wb_disable(rdev);
1897 rv770_pcie_gart_disable(rdev);
1898
1899 return 0;
1900}
1901
1902/* Plan is to move initialization in that function and use
1903 * helper function so that radeon_device_init pretty much
1904 * do nothing more than calling asic specific function. This
1905 * should also allow to remove a bunch of callback function
1906 * like vram_info.
1907 */
1908int rv770_init(struct radeon_device *rdev)
1909{
1910 int r;
1911
1912 /* Read BIOS */
1913 if (!radeon_get_bios(rdev)) {
1914 if (ASIC_IS_AVIVO(rdev))
1915 return -EINVAL;
1916 }
1917 /* Must be an ATOMBIOS */
1918 if (!rdev->is_atom_bios) {
1919 dev_err(rdev->dev, "Expecting atombios for R600 GPU\n");
1920 return -EINVAL;
1921 }
1922 r = radeon_atombios_init(rdev);
1923 if (r)
1924 return r;
1925 /* Post card if necessary */
1926 if (!radeon_card_posted(rdev)) {
1927 if (!rdev->bios) {
1928 dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n");
1929 return -EINVAL;
1930 }
1931 DRM_INFO("GPU not posted. posting now...\n");
1932 atom_asic_init(rdev->mode_info.atom_context);
1933 }
1934 /* init golden registers */
1935 rv770_init_golden_registers(rdev);
1936 /* Initialize scratch registers */
1937 r600_scratch_init(rdev);
1938 /* Initialize surface registers */
1939 radeon_surface_init(rdev);
1940 /* Initialize clocks */
1941 radeon_get_clock_info(rdev->ddev);
1942 /* Fence driver */
1943 r = radeon_fence_driver_init(rdev);
1944 if (r)
1945 return r;
1946 /* initialize AGP */
1947 if (rdev->flags & RADEON_IS_AGP) {
1948 r = radeon_agp_init(rdev);
1949 if (r)
1950 radeon_agp_disable(rdev);
1951 }
1952 r = rv770_mc_init(rdev);
1953 if (r)
1954 return r;
1955 /* Memory manager */
1956 r = radeon_bo_init(rdev);
1957 if (r)
1958 return r;
1959
1960 if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) {
1961 r = r600_init_microcode(rdev);
1962 if (r) {
1963 DRM_ERROR("Failed to load firmware!\n");
1964 return r;
1965 }
1966 }
1967
1968 /* Initialize power management */
1969 radeon_pm_init(rdev);
1970
1971 rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ring_obj = NULL;
1972 r600_ring_init(rdev, &rdev->ring[RADEON_RING_TYPE_GFX_INDEX], 1024 * 1024);
1973
1974 rdev->ring[R600_RING_TYPE_DMA_INDEX].ring_obj = NULL;
1975 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX], 64 * 1024);
1976
1977 rv770_uvd_init(rdev);
1978
1979 rdev->ih.ring_obj = NULL;
1980 r600_ih_ring_init(rdev, 64 * 1024);
1981
1982 r = r600_pcie_gart_init(rdev);
1983 if (r)
1984 return r;
1985
1986 rdev->accel_working = true;
1987 r = rv770_startup(rdev);
1988 if (r) {
1989 dev_err(rdev->dev, "disabling GPU acceleration\n");
1990 r700_cp_fini(rdev);
1991 r600_dma_fini(rdev);
1992 r600_irq_fini(rdev);
1993 radeon_wb_fini(rdev);
1994 radeon_ib_pool_fini(rdev);
1995 radeon_irq_kms_fini(rdev);
1996 rv770_pcie_gart_fini(rdev);
1997 rdev->accel_working = false;
1998 }
1999
2000 return 0;
2001}
2002
2003void rv770_fini(struct radeon_device *rdev)
2004{
2005 radeon_pm_fini(rdev);
2006 r700_cp_fini(rdev);
2007 r600_dma_fini(rdev);
2008 r600_irq_fini(rdev);
2009 radeon_wb_fini(rdev);
2010 radeon_ib_pool_fini(rdev);
2011 radeon_irq_kms_fini(rdev);
2012 uvd_v1_0_fini(rdev);
2013 radeon_uvd_fini(rdev);
2014 rv770_pcie_gart_fini(rdev);
2015 r600_vram_scratch_fini(rdev);
2016 radeon_gem_fini(rdev);
2017 radeon_fence_driver_fini(rdev);
2018 radeon_agp_fini(rdev);
2019 radeon_bo_fini(rdev);
2020 radeon_atombios_fini(rdev);
2021 kfree(rdev->bios);
2022 rdev->bios = NULL;
2023}
2024
2025static void rv770_pcie_gen2_enable(struct radeon_device *rdev)
2026{
2027 u32 link_width_cntl, lanes, speed_cntl, tmp;
2028 u16 link_cntl2;
2029
2030 if (radeon_pcie_gen2 == 0)
2031 return;
2032
2033 if (rdev->flags & RADEON_IS_IGP)
2034 return;
2035
2036 if (!(rdev->flags & RADEON_IS_PCIE))
2037 return;
2038
2039 /* x2 cards have a special sequence */
2040 if (ASIC_IS_X2(rdev))
2041 return;
2042
2043 if ((rdev->pdev->bus->max_bus_speed != PCIE_SPEED_5_0GT) &&
2044 (rdev->pdev->bus->max_bus_speed != PCIE_SPEED_8_0GT))
2045 return;
2046
2047 DRM_INFO("enabling PCIE gen 2 link speeds, disable with radeon.pcie_gen2=0\n");
2048
2049 /* advertise upconfig capability */
2050 link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
2051 link_width_cntl &= ~LC_UPCONFIGURE_DIS;
2052 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
2053 link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
2054 if (link_width_cntl & LC_RENEGOTIATION_SUPPORT) {
2055 lanes = (link_width_cntl & LC_LINK_WIDTH_RD_MASK) >> LC_LINK_WIDTH_RD_SHIFT;
2056 link_width_cntl &= ~(LC_LINK_WIDTH_MASK |
2057 LC_RECONFIG_ARC_MISSING_ESCAPE);
2058 link_width_cntl |= lanes | LC_RECONFIG_NOW |
2059 LC_RENEGOTIATE_EN | LC_UPCONFIGURE_SUPPORT;
2060 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
2061 } else {
2062 link_width_cntl |= LC_UPCONFIGURE_DIS;
2063 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
2064 }
2065
2066 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
2067 if ((speed_cntl & LC_OTHER_SIDE_EVER_SENT_GEN2) &&
2068 (speed_cntl & LC_OTHER_SIDE_SUPPORTS_GEN2)) {
2069
2070 tmp = RREG32(0x541c);
2071 WREG32(0x541c, tmp | 0x8);
2072 WREG32(MM_CFGREGS_CNTL, MM_WR_TO_CFG_EN);
2073 link_cntl2 = RREG16(0x4088);
2074 link_cntl2 &= ~TARGET_LINK_SPEED_MASK;
2075 link_cntl2 |= 0x2;
2076 WREG16(0x4088, link_cntl2);
2077 WREG32(MM_CFGREGS_CNTL, 0);
2078
2079 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
2080 speed_cntl &= ~LC_TARGET_LINK_SPEED_OVERRIDE_EN;
2081 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
2082
2083 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
2084 speed_cntl |= LC_CLR_FAILED_SPD_CHANGE_CNT;
2085 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
2086
2087 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
2088 speed_cntl &= ~LC_CLR_FAILED_SPD_CHANGE_CNT;
2089 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
2090
2091 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
2092 speed_cntl |= LC_GEN2_EN_STRAP;
2093 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
2094
2095 } else {
2096 link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
2097 /* XXX: only disable it if gen1 bridge vendor == 0x111d or 0x1106 */
2098 if (1)
2099 link_width_cntl |= LC_UPCONFIGURE_DIS;
2100 else
2101 link_width_cntl &= ~LC_UPCONFIGURE_DIS;
2102 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
2103 }
2104}