Linux Audio

Check our new training course

Loading...
v3.1
   1/*
   2 * Copyright 2010 Advanced Micro Devices, Inc.
   3 *
   4 * Permission is hereby granted, free of charge, to any person obtaining a
   5 * copy of this software and associated documentation files (the "Software"),
   6 * to deal in the Software without restriction, including without limitation
   7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
   8 * and/or sell copies of the Software, and to permit persons to whom the
   9 * Software is furnished to do so, subject to the following conditions:
  10 *
  11 * The above copyright notice and this permission notice shall be included in
  12 * all copies or substantial portions of the Software.
  13 *
  14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
  17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
  18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
  19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
  20 * OTHER DEALINGS IN THE SOFTWARE.
  21 *
  22 * Authors: Alex Deucher
  23 */
 
  24#include <linux/firmware.h>
  25#include <linux/platform_device.h>
 
  26#include <linux/slab.h>
  27#include "drmP.h"
  28#include "radeon.h"
  29#include "radeon_asic.h"
  30#include "radeon_drm.h"
  31#include "nid.h"
  32#include "atom.h"
  33#include "ni_reg.h"
  34#include "cayman_blit_shaders.h"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  35
  36extern void evergreen_mc_stop(struct radeon_device *rdev, struct evergreen_mc_save *save);
  37extern void evergreen_mc_resume(struct radeon_device *rdev, struct evergreen_mc_save *save);
  38extern int evergreen_mc_wait_for_idle(struct radeon_device *rdev);
  39extern void evergreen_mc_program(struct radeon_device *rdev);
  40extern void evergreen_irq_suspend(struct radeon_device *rdev);
  41extern int evergreen_mc_init(struct radeon_device *rdev);
  42extern void evergreen_fix_pci_max_read_req_size(struct radeon_device *rdev);
  43
  44#define EVERGREEN_PFP_UCODE_SIZE 1120
  45#define EVERGREEN_PM4_UCODE_SIZE 1376
  46#define EVERGREEN_RLC_UCODE_SIZE 768
  47#define BTC_MC_UCODE_SIZE 6024
  48
  49#define CAYMAN_PFP_UCODE_SIZE 2176
  50#define CAYMAN_PM4_UCODE_SIZE 2176
  51#define CAYMAN_RLC_UCODE_SIZE 1024
  52#define CAYMAN_MC_UCODE_SIZE 6037
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  53
  54/* Firmware Names */
  55MODULE_FIRMWARE("radeon/BARTS_pfp.bin");
  56MODULE_FIRMWARE("radeon/BARTS_me.bin");
  57MODULE_FIRMWARE("radeon/BARTS_mc.bin");
 
  58MODULE_FIRMWARE("radeon/BTC_rlc.bin");
  59MODULE_FIRMWARE("radeon/TURKS_pfp.bin");
  60MODULE_FIRMWARE("radeon/TURKS_me.bin");
  61MODULE_FIRMWARE("radeon/TURKS_mc.bin");
 
  62MODULE_FIRMWARE("radeon/CAICOS_pfp.bin");
  63MODULE_FIRMWARE("radeon/CAICOS_me.bin");
  64MODULE_FIRMWARE("radeon/CAICOS_mc.bin");
 
  65MODULE_FIRMWARE("radeon/CAYMAN_pfp.bin");
  66MODULE_FIRMWARE("radeon/CAYMAN_me.bin");
  67MODULE_FIRMWARE("radeon/CAYMAN_mc.bin");
  68MODULE_FIRMWARE("radeon/CAYMAN_rlc.bin");
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
  69
  70#define BTC_IO_MC_REGS_SIZE 29
  71
  72static const u32 barts_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
  73	{0x00000077, 0xff010100},
  74	{0x00000078, 0x00000000},
  75	{0x00000079, 0x00001434},
  76	{0x0000007a, 0xcc08ec08},
  77	{0x0000007b, 0x00040000},
  78	{0x0000007c, 0x000080c0},
  79	{0x0000007d, 0x09000000},
  80	{0x0000007e, 0x00210404},
  81	{0x00000081, 0x08a8e800},
  82	{0x00000082, 0x00030444},
  83	{0x00000083, 0x00000000},
  84	{0x00000085, 0x00000001},
  85	{0x00000086, 0x00000002},
  86	{0x00000087, 0x48490000},
  87	{0x00000088, 0x20244647},
  88	{0x00000089, 0x00000005},
  89	{0x0000008b, 0x66030000},
  90	{0x0000008c, 0x00006603},
  91	{0x0000008d, 0x00000100},
  92	{0x0000008f, 0x00001c0a},
  93	{0x00000090, 0xff000001},
  94	{0x00000094, 0x00101101},
  95	{0x00000095, 0x00000fff},
  96	{0x00000096, 0x00116fff},
  97	{0x00000097, 0x60010000},
  98	{0x00000098, 0x10010000},
  99	{0x00000099, 0x00006000},
 100	{0x0000009a, 0x00001000},
 101	{0x0000009f, 0x00946a00}
 102};
 103
 104static const u32 turks_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 105	{0x00000077, 0xff010100},
 106	{0x00000078, 0x00000000},
 107	{0x00000079, 0x00001434},
 108	{0x0000007a, 0xcc08ec08},
 109	{0x0000007b, 0x00040000},
 110	{0x0000007c, 0x000080c0},
 111	{0x0000007d, 0x09000000},
 112	{0x0000007e, 0x00210404},
 113	{0x00000081, 0x08a8e800},
 114	{0x00000082, 0x00030444},
 115	{0x00000083, 0x00000000},
 116	{0x00000085, 0x00000001},
 117	{0x00000086, 0x00000002},
 118	{0x00000087, 0x48490000},
 119	{0x00000088, 0x20244647},
 120	{0x00000089, 0x00000005},
 121	{0x0000008b, 0x66030000},
 122	{0x0000008c, 0x00006603},
 123	{0x0000008d, 0x00000100},
 124	{0x0000008f, 0x00001c0a},
 125	{0x00000090, 0xff000001},
 126	{0x00000094, 0x00101101},
 127	{0x00000095, 0x00000fff},
 128	{0x00000096, 0x00116fff},
 129	{0x00000097, 0x60010000},
 130	{0x00000098, 0x10010000},
 131	{0x00000099, 0x00006000},
 132	{0x0000009a, 0x00001000},
 133	{0x0000009f, 0x00936a00}
 134};
 135
 136static const u32 caicos_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 137	{0x00000077, 0xff010100},
 138	{0x00000078, 0x00000000},
 139	{0x00000079, 0x00001434},
 140	{0x0000007a, 0xcc08ec08},
 141	{0x0000007b, 0x00040000},
 142	{0x0000007c, 0x000080c0},
 143	{0x0000007d, 0x09000000},
 144	{0x0000007e, 0x00210404},
 145	{0x00000081, 0x08a8e800},
 146	{0x00000082, 0x00030444},
 147	{0x00000083, 0x00000000},
 148	{0x00000085, 0x00000001},
 149	{0x00000086, 0x00000002},
 150	{0x00000087, 0x48490000},
 151	{0x00000088, 0x20244647},
 152	{0x00000089, 0x00000005},
 153	{0x0000008b, 0x66030000},
 154	{0x0000008c, 0x00006603},
 155	{0x0000008d, 0x00000100},
 156	{0x0000008f, 0x00001c0a},
 157	{0x00000090, 0xff000001},
 158	{0x00000094, 0x00101101},
 159	{0x00000095, 0x00000fff},
 160	{0x00000096, 0x00116fff},
 161	{0x00000097, 0x60010000},
 162	{0x00000098, 0x10010000},
 163	{0x00000099, 0x00006000},
 164	{0x0000009a, 0x00001000},
 165	{0x0000009f, 0x00916a00}
 166};
 167
 168static const u32 cayman_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 169	{0x00000077, 0xff010100},
 170	{0x00000078, 0x00000000},
 171	{0x00000079, 0x00001434},
 172	{0x0000007a, 0xcc08ec08},
 173	{0x0000007b, 0x00040000},
 174	{0x0000007c, 0x000080c0},
 175	{0x0000007d, 0x09000000},
 176	{0x0000007e, 0x00210404},
 177	{0x00000081, 0x08a8e800},
 178	{0x00000082, 0x00030444},
 179	{0x00000083, 0x00000000},
 180	{0x00000085, 0x00000001},
 181	{0x00000086, 0x00000002},
 182	{0x00000087, 0x48490000},
 183	{0x00000088, 0x20244647},
 184	{0x00000089, 0x00000005},
 185	{0x0000008b, 0x66030000},
 186	{0x0000008c, 0x00006603},
 187	{0x0000008d, 0x00000100},
 188	{0x0000008f, 0x00001c0a},
 189	{0x00000090, 0xff000001},
 190	{0x00000094, 0x00101101},
 191	{0x00000095, 0x00000fff},
 192	{0x00000096, 0x00116fff},
 193	{0x00000097, 0x60010000},
 194	{0x00000098, 0x10010000},
 195	{0x00000099, 0x00006000},
 196	{0x0000009a, 0x00001000},
 197	{0x0000009f, 0x00976b00}
 198};
 199
 200int ni_mc_load_microcode(struct radeon_device *rdev)
 201{
 202	const __be32 *fw_data;
 203	u32 mem_type, running, blackout = 0;
 204	u32 *io_mc_regs;
 205	int i, ucode_size, regs_size;
 206
 207	if (!rdev->mc_fw)
 208		return -EINVAL;
 209
 210	switch (rdev->family) {
 211	case CHIP_BARTS:
 212		io_mc_regs = (u32 *)&barts_io_mc_regs;
 213		ucode_size = BTC_MC_UCODE_SIZE;
 214		regs_size = BTC_IO_MC_REGS_SIZE;
 215		break;
 216	case CHIP_TURKS:
 217		io_mc_regs = (u32 *)&turks_io_mc_regs;
 218		ucode_size = BTC_MC_UCODE_SIZE;
 219		regs_size = BTC_IO_MC_REGS_SIZE;
 220		break;
 221	case CHIP_CAICOS:
 222	default:
 223		io_mc_regs = (u32 *)&caicos_io_mc_regs;
 224		ucode_size = BTC_MC_UCODE_SIZE;
 225		regs_size = BTC_IO_MC_REGS_SIZE;
 226		break;
 227	case CHIP_CAYMAN:
 228		io_mc_regs = (u32 *)&cayman_io_mc_regs;
 229		ucode_size = CAYMAN_MC_UCODE_SIZE;
 230		regs_size = BTC_IO_MC_REGS_SIZE;
 231		break;
 232	}
 233
 234	mem_type = (RREG32(MC_SEQ_MISC0) & MC_SEQ_MISC0_GDDR5_MASK) >> MC_SEQ_MISC0_GDDR5_SHIFT;
 235	running = RREG32(MC_SEQ_SUP_CNTL) & RUN_MASK;
 236
 237	if ((mem_type == MC_SEQ_MISC0_GDDR5_VALUE) && (running == 0)) {
 238		if (running) {
 239			blackout = RREG32(MC_SHARED_BLACKOUT_CNTL);
 240			WREG32(MC_SHARED_BLACKOUT_CNTL, 1);
 241		}
 242
 243		/* reset the engine and set to writable */
 244		WREG32(MC_SEQ_SUP_CNTL, 0x00000008);
 245		WREG32(MC_SEQ_SUP_CNTL, 0x00000010);
 246
 247		/* load mc io regs */
 248		for (i = 0; i < regs_size; i++) {
 249			WREG32(MC_SEQ_IO_DEBUG_INDEX, io_mc_regs[(i << 1)]);
 250			WREG32(MC_SEQ_IO_DEBUG_DATA, io_mc_regs[(i << 1) + 1]);
 251		}
 252		/* load the MC ucode */
 253		fw_data = (const __be32 *)rdev->mc_fw->data;
 254		for (i = 0; i < ucode_size; i++)
 255			WREG32(MC_SEQ_SUP_PGM, be32_to_cpup(fw_data++));
 256
 257		/* put the engine back into the active state */
 258		WREG32(MC_SEQ_SUP_CNTL, 0x00000008);
 259		WREG32(MC_SEQ_SUP_CNTL, 0x00000004);
 260		WREG32(MC_SEQ_SUP_CNTL, 0x00000001);
 261
 262		/* wait for training to complete */
 263		while (!(RREG32(MC_IO_PAD_CNTL_D0) & MEM_FALL_OUT_CMD))
 264			udelay(10);
 265
 266		if (running)
 267			WREG32(MC_SHARED_BLACKOUT_CNTL, blackout);
 268	}
 269
 270	return 0;
 271}
 272
 273int ni_init_microcode(struct radeon_device *rdev)
 274{
 275	struct platform_device *pdev;
 276	const char *chip_name;
 277	const char *rlc_chip_name;
 278	size_t pfp_req_size, me_req_size, rlc_req_size, mc_req_size;
 
 279	char fw_name[30];
 280	int err;
 281
 282	DRM_DEBUG("\n");
 283
 284	pdev = platform_device_register_simple("radeon_cp", 0, NULL, 0);
 285	err = IS_ERR(pdev);
 286	if (err) {
 287		printk(KERN_ERR "radeon_cp: Failed to register firmware\n");
 288		return -EINVAL;
 289	}
 290
 291	switch (rdev->family) {
 292	case CHIP_BARTS:
 293		chip_name = "BARTS";
 294		rlc_chip_name = "BTC";
 295		pfp_req_size = EVERGREEN_PFP_UCODE_SIZE * 4;
 296		me_req_size = EVERGREEN_PM4_UCODE_SIZE * 4;
 297		rlc_req_size = EVERGREEN_RLC_UCODE_SIZE * 4;
 298		mc_req_size = BTC_MC_UCODE_SIZE * 4;
 
 299		break;
 300	case CHIP_TURKS:
 301		chip_name = "TURKS";
 302		rlc_chip_name = "BTC";
 303		pfp_req_size = EVERGREEN_PFP_UCODE_SIZE * 4;
 304		me_req_size = EVERGREEN_PM4_UCODE_SIZE * 4;
 305		rlc_req_size = EVERGREEN_RLC_UCODE_SIZE * 4;
 306		mc_req_size = BTC_MC_UCODE_SIZE * 4;
 
 307		break;
 308	case CHIP_CAICOS:
 309		chip_name = "CAICOS";
 310		rlc_chip_name = "BTC";
 311		pfp_req_size = EVERGREEN_PFP_UCODE_SIZE * 4;
 312		me_req_size = EVERGREEN_PM4_UCODE_SIZE * 4;
 313		rlc_req_size = EVERGREEN_RLC_UCODE_SIZE * 4;
 314		mc_req_size = BTC_MC_UCODE_SIZE * 4;
 
 315		break;
 316	case CHIP_CAYMAN:
 317		chip_name = "CAYMAN";
 318		rlc_chip_name = "CAYMAN";
 319		pfp_req_size = CAYMAN_PFP_UCODE_SIZE * 4;
 320		me_req_size = CAYMAN_PM4_UCODE_SIZE * 4;
 321		rlc_req_size = CAYMAN_RLC_UCODE_SIZE * 4;
 322		mc_req_size = CAYMAN_MC_UCODE_SIZE * 4;
 
 323		break;
 324	default: BUG();
 
 
 
 
 
 
 
 
 
 
 325	}
 326
 327	DRM_INFO("Loading %s Microcode\n", chip_name);
 328
 329	snprintf(fw_name, sizeof(fw_name), "radeon/%s_pfp.bin", chip_name);
 330	err = request_firmware(&rdev->pfp_fw, fw_name, &pdev->dev);
 331	if (err)
 332		goto out;
 333	if (rdev->pfp_fw->size != pfp_req_size) {
 334		printk(KERN_ERR
 335		       "ni_cp: Bogus length %zu in firmware \"%s\"\n",
 336		       rdev->pfp_fw->size, fw_name);
 337		err = -EINVAL;
 338		goto out;
 339	}
 340
 341	snprintf(fw_name, sizeof(fw_name), "radeon/%s_me.bin", chip_name);
 342	err = request_firmware(&rdev->me_fw, fw_name, &pdev->dev);
 343	if (err)
 344		goto out;
 345	if (rdev->me_fw->size != me_req_size) {
 346		printk(KERN_ERR
 347		       "ni_cp: Bogus length %zu in firmware \"%s\"\n",
 348		       rdev->me_fw->size, fw_name);
 349		err = -EINVAL;
 350	}
 351
 352	snprintf(fw_name, sizeof(fw_name), "radeon/%s_rlc.bin", rlc_chip_name);
 353	err = request_firmware(&rdev->rlc_fw, fw_name, &pdev->dev);
 354	if (err)
 355		goto out;
 356	if (rdev->rlc_fw->size != rlc_req_size) {
 357		printk(KERN_ERR
 358		       "ni_rlc: Bogus length %zu in firmware \"%s\"\n",
 359		       rdev->rlc_fw->size, fw_name);
 360		err = -EINVAL;
 361	}
 362
 363	snprintf(fw_name, sizeof(fw_name), "radeon/%s_mc.bin", chip_name);
 364	err = request_firmware(&rdev->mc_fw, fw_name, &pdev->dev);
 365	if (err)
 366		goto out;
 367	if (rdev->mc_fw->size != mc_req_size) {
 368		printk(KERN_ERR
 369		       "ni_mc: Bogus length %zu in firmware \"%s\"\n",
 370		       rdev->mc_fw->size, fw_name);
 371		err = -EINVAL;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 372	}
 373out:
 374	platform_device_unregister(pdev);
 375
 
 376	if (err) {
 377		if (err != -EINVAL)
 378			printk(KERN_ERR
 379			       "ni_cp: Failed to load firmware \"%s\"\n",
 380			       fw_name);
 381		release_firmware(rdev->pfp_fw);
 382		rdev->pfp_fw = NULL;
 383		release_firmware(rdev->me_fw);
 384		rdev->me_fw = NULL;
 385		release_firmware(rdev->rlc_fw);
 386		rdev->rlc_fw = NULL;
 387		release_firmware(rdev->mc_fw);
 388		rdev->mc_fw = NULL;
 389	}
 390	return err;
 391}
 392
 393/*
 394 * Core functions
 
 
 
 
 
 
 
 395 */
 396static u32 cayman_get_tile_pipe_to_backend_map(struct radeon_device *rdev,
 397					       u32 num_tile_pipes,
 398					       u32 num_backends_per_asic,
 399					       u32 *backend_disable_mask_per_asic,
 400					       u32 num_shader_engines)
 401{
 402	u32 backend_map = 0;
 403	u32 enabled_backends_mask = 0;
 404	u32 enabled_backends_count = 0;
 405	u32 num_backends_per_se;
 406	u32 cur_pipe;
 407	u32 swizzle_pipe[CAYMAN_MAX_PIPES];
 408	u32 cur_backend = 0;
 409	u32 i;
 410	bool force_no_swizzle;
 411
 412	/* force legal values */
 413	if (num_tile_pipes < 1)
 414		num_tile_pipes = 1;
 415	if (num_tile_pipes > rdev->config.cayman.max_tile_pipes)
 416		num_tile_pipes = rdev->config.cayman.max_tile_pipes;
 417	if (num_shader_engines < 1)
 418		num_shader_engines = 1;
 419	if (num_shader_engines > rdev->config.cayman.max_shader_engines)
 420		num_shader_engines = rdev->config.cayman.max_shader_engines;
 421	if (num_backends_per_asic < num_shader_engines)
 422		num_backends_per_asic = num_shader_engines;
 423	if (num_backends_per_asic > (rdev->config.cayman.max_backends_per_se * num_shader_engines))
 424		num_backends_per_asic = rdev->config.cayman.max_backends_per_se * num_shader_engines;
 425
 426	/* make sure we have the same number of backends per se */
 427	num_backends_per_asic = ALIGN(num_backends_per_asic, num_shader_engines);
 428	/* set up the number of backends per se */
 429	num_backends_per_se = num_backends_per_asic / num_shader_engines;
 430	if (num_backends_per_se > rdev->config.cayman.max_backends_per_se) {
 431		num_backends_per_se = rdev->config.cayman.max_backends_per_se;
 432		num_backends_per_asic = num_backends_per_se * num_shader_engines;
 433	}
 434
 435	/* create enable mask and count for enabled backends */
 436	for (i = 0; i < CAYMAN_MAX_BACKENDS; ++i) {
 437		if (((*backend_disable_mask_per_asic >> i) & 1) == 0) {
 438			enabled_backends_mask |= (1 << i);
 439			++enabled_backends_count;
 440		}
 441		if (enabled_backends_count == num_backends_per_asic)
 442			break;
 443	}
 444
 445	/* force the backends mask to match the current number of backends */
 446	if (enabled_backends_count != num_backends_per_asic) {
 447		u32 this_backend_enabled;
 448		u32 shader_engine;
 449		u32 backend_per_se;
 450
 451		enabled_backends_mask = 0;
 452		enabled_backends_count = 0;
 453		*backend_disable_mask_per_asic = CAYMAN_MAX_BACKENDS_MASK;
 454		for (i = 0; i < CAYMAN_MAX_BACKENDS; ++i) {
 455			/* calc the current se */
 456			shader_engine = i / rdev->config.cayman.max_backends_per_se;
 457			/* calc the backend per se */
 458			backend_per_se = i % rdev->config.cayman.max_backends_per_se;
 459			/* default to not enabled */
 460			this_backend_enabled = 0;
 461			if ((shader_engine < num_shader_engines) &&
 462			    (backend_per_se < num_backends_per_se))
 463				this_backend_enabled = 1;
 464			if (this_backend_enabled) {
 465				enabled_backends_mask |= (1 << i);
 466				*backend_disable_mask_per_asic &= ~(1 << i);
 467				++enabled_backends_count;
 468			}
 469		}
 470	}
 471
 472
 473	memset((uint8_t *)&swizzle_pipe[0], 0, sizeof(u32) * CAYMAN_MAX_PIPES);
 474	switch (rdev->family) {
 475	case CHIP_CAYMAN:
 476		force_no_swizzle = true;
 477		break;
 478	default:
 479		force_no_swizzle = false;
 480		break;
 481	}
 482	if (force_no_swizzle) {
 483		bool last_backend_enabled = false;
 484
 485		force_no_swizzle = false;
 486		for (i = 0; i < CAYMAN_MAX_BACKENDS; ++i) {
 487			if (((enabled_backends_mask >> i) & 1) == 1) {
 488				if (last_backend_enabled)
 489					force_no_swizzle = true;
 490				last_backend_enabled = true;
 491			} else
 492				last_backend_enabled = false;
 493		}
 494	}
 495
 496	switch (num_tile_pipes) {
 497	case 1:
 498	case 3:
 499	case 5:
 500	case 7:
 501		DRM_ERROR("odd number of pipes!\n");
 502		break;
 503	case 2:
 504		swizzle_pipe[0] = 0;
 505		swizzle_pipe[1] = 1;
 506		break;
 507	case 4:
 508		if (force_no_swizzle) {
 509			swizzle_pipe[0] = 0;
 510			swizzle_pipe[1] = 1;
 511			swizzle_pipe[2] = 2;
 512			swizzle_pipe[3] = 3;
 513		} else {
 514			swizzle_pipe[0] = 0;
 515			swizzle_pipe[1] = 2;
 516			swizzle_pipe[2] = 1;
 517			swizzle_pipe[3] = 3;
 518		}
 519		break;
 520	case 6:
 521		if (force_no_swizzle) {
 522			swizzle_pipe[0] = 0;
 523			swizzle_pipe[1] = 1;
 524			swizzle_pipe[2] = 2;
 525			swizzle_pipe[3] = 3;
 526			swizzle_pipe[4] = 4;
 527			swizzle_pipe[5] = 5;
 528		} else {
 529			swizzle_pipe[0] = 0;
 530			swizzle_pipe[1] = 2;
 531			swizzle_pipe[2] = 4;
 532			swizzle_pipe[3] = 1;
 533			swizzle_pipe[4] = 3;
 534			swizzle_pipe[5] = 5;
 535		}
 536		break;
 537	case 8:
 538		if (force_no_swizzle) {
 539			swizzle_pipe[0] = 0;
 540			swizzle_pipe[1] = 1;
 541			swizzle_pipe[2] = 2;
 542			swizzle_pipe[3] = 3;
 543			swizzle_pipe[4] = 4;
 544			swizzle_pipe[5] = 5;
 545			swizzle_pipe[6] = 6;
 546			swizzle_pipe[7] = 7;
 547		} else {
 548			swizzle_pipe[0] = 0;
 549			swizzle_pipe[1] = 2;
 550			swizzle_pipe[2] = 4;
 551			swizzle_pipe[3] = 6;
 552			swizzle_pipe[4] = 1;
 553			swizzle_pipe[5] = 3;
 554			swizzle_pipe[6] = 5;
 555			swizzle_pipe[7] = 7;
 556		}
 557		break;
 558	}
 559
 560	for (cur_pipe = 0; cur_pipe < num_tile_pipes; ++cur_pipe) {
 561		while (((1 << cur_backend) & enabled_backends_mask) == 0)
 562			cur_backend = (cur_backend + 1) % CAYMAN_MAX_BACKENDS;
 563
 564		backend_map |= (((cur_backend & 0xf) << (swizzle_pipe[cur_pipe] * 4)));
 565
 566		cur_backend = (cur_backend + 1) % CAYMAN_MAX_BACKENDS;
 567	}
 568
 569	return backend_map;
 570}
 571
 572static u32 cayman_get_disable_mask_per_asic(struct radeon_device *rdev,
 573					    u32 disable_mask_per_se,
 574					    u32 max_disable_mask_per_se,
 575					    u32 num_shader_engines)
 576{
 577	u32 disable_field_width_per_se = r600_count_pipe_bits(disable_mask_per_se);
 578	u32 disable_mask_per_asic = disable_mask_per_se & max_disable_mask_per_se;
 579
 580	if (num_shader_engines == 1)
 581		return disable_mask_per_asic;
 582	else if (num_shader_engines == 2)
 583		return disable_mask_per_asic | (disable_mask_per_asic << disable_field_width_per_se);
 584	else
 585		return 0xffffffff;
 586}
 587
 
 
 
 588static void cayman_gpu_init(struct radeon_device *rdev)
 589{
 590	u32 cc_rb_backend_disable = 0;
 591	u32 cc_gc_shader_pipe_config;
 592	u32 gb_addr_config = 0;
 593	u32 mc_shared_chmap, mc_arb_ramcfg;
 594	u32 gb_backend_map;
 595	u32 cgts_tcc_disable;
 596	u32 sx_debug_1;
 597	u32 smx_dc_ctl0;
 598	u32 gc_user_shader_pipe_config;
 599	u32 gc_user_rb_backend_disable;
 600	u32 cgts_user_tcc_disable;
 601	u32 cgts_sm_ctrl_reg;
 602	u32 hdp_host_path_cntl;
 603	u32 tmp;
 
 604	int i, j;
 605
 606	switch (rdev->family) {
 607	case CHIP_CAYMAN:
 608	default:
 609		rdev->config.cayman.max_shader_engines = 2;
 610		rdev->config.cayman.max_pipes_per_simd = 4;
 611		rdev->config.cayman.max_tile_pipes = 8;
 612		rdev->config.cayman.max_simds_per_se = 12;
 613		rdev->config.cayman.max_backends_per_se = 4;
 614		rdev->config.cayman.max_texture_channel_caches = 8;
 615		rdev->config.cayman.max_gprs = 256;
 616		rdev->config.cayman.max_threads = 256;
 617		rdev->config.cayman.max_gs_threads = 32;
 618		rdev->config.cayman.max_stack_entries = 512;
 619		rdev->config.cayman.sx_num_of_sets = 8;
 620		rdev->config.cayman.sx_max_export_size = 256;
 621		rdev->config.cayman.sx_max_export_pos_size = 64;
 622		rdev->config.cayman.sx_max_export_smx_size = 192;
 623		rdev->config.cayman.max_hw_contexts = 8;
 624		rdev->config.cayman.sq_num_cf_insts = 2;
 625
 626		rdev->config.cayman.sc_prim_fifo_size = 0x100;
 627		rdev->config.cayman.sc_hiz_tile_fifo_size = 0x30;
 628		rdev->config.cayman.sc_earlyz_tile_fifo_size = 0x130;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 629		break;
 630	}
 631
 632	/* Initialize HDP */
 633	for (i = 0, j = 0; i < 32; i++, j += 0x18) {
 634		WREG32((0x2c14 + j), 0x00000000);
 635		WREG32((0x2c18 + j), 0x00000000);
 636		WREG32((0x2c1c + j), 0x00000000);
 637		WREG32((0x2c20 + j), 0x00000000);
 638		WREG32((0x2c24 + j), 0x00000000);
 639	}
 640
 641	WREG32(GRBM_CNTL, GRBM_READ_TIMEOUT(0xff));
 
 
 642
 643	evergreen_fix_pci_max_read_req_size(rdev);
 644
 645	mc_shared_chmap = RREG32(MC_SHARED_CHMAP);
 646	mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG);
 647
 648	cc_rb_backend_disable = RREG32(CC_RB_BACKEND_DISABLE);
 649	cc_gc_shader_pipe_config = RREG32(CC_GC_SHADER_PIPE_CONFIG);
 650	cgts_tcc_disable = 0xff000000;
 651	gc_user_rb_backend_disable = RREG32(GC_USER_RB_BACKEND_DISABLE);
 652	gc_user_shader_pipe_config = RREG32(GC_USER_SHADER_PIPE_CONFIG);
 653	cgts_user_tcc_disable = RREG32(CGTS_USER_TCC_DISABLE);
 654
 655	rdev->config.cayman.num_shader_engines = rdev->config.cayman.max_shader_engines;
 656	tmp = ((~gc_user_shader_pipe_config) & INACTIVE_QD_PIPES_MASK) >> INACTIVE_QD_PIPES_SHIFT;
 657	rdev->config.cayman.num_shader_pipes_per_simd = r600_count_pipe_bits(tmp);
 658	rdev->config.cayman.num_tile_pipes = rdev->config.cayman.max_tile_pipes;
 659	tmp = ((~gc_user_shader_pipe_config) & INACTIVE_SIMDS_MASK) >> INACTIVE_SIMDS_SHIFT;
 660	rdev->config.cayman.num_simds_per_se = r600_count_pipe_bits(tmp);
 661	tmp = ((~gc_user_rb_backend_disable) & BACKEND_DISABLE_MASK) >> BACKEND_DISABLE_SHIFT;
 662	rdev->config.cayman.num_backends_per_se = r600_count_pipe_bits(tmp);
 663	tmp = (gc_user_rb_backend_disable & BACKEND_DISABLE_MASK) >> BACKEND_DISABLE_SHIFT;
 664	rdev->config.cayman.backend_disable_mask_per_asic =
 665		cayman_get_disable_mask_per_asic(rdev, tmp, CAYMAN_MAX_BACKENDS_PER_SE_MASK,
 666						 rdev->config.cayman.num_shader_engines);
 667	rdev->config.cayman.backend_map =
 668		cayman_get_tile_pipe_to_backend_map(rdev, rdev->config.cayman.num_tile_pipes,
 669						    rdev->config.cayman.num_backends_per_se *
 670						    rdev->config.cayman.num_shader_engines,
 671						    &rdev->config.cayman.backend_disable_mask_per_asic,
 672						    rdev->config.cayman.num_shader_engines);
 673	tmp = ((~cgts_user_tcc_disable) & TCC_DISABLE_MASK) >> TCC_DISABLE_SHIFT;
 674	rdev->config.cayman.num_texture_channel_caches = r600_count_pipe_bits(tmp);
 675	tmp = (mc_arb_ramcfg & BURSTLENGTH_MASK) >> BURSTLENGTH_SHIFT;
 676	rdev->config.cayman.mem_max_burst_length_bytes = (tmp + 1) * 256;
 677	if (rdev->config.cayman.mem_max_burst_length_bytes > 512)
 678		rdev->config.cayman.mem_max_burst_length_bytes = 512;
 679	tmp = (mc_arb_ramcfg & NOOFCOLS_MASK) >> NOOFCOLS_SHIFT;
 680	rdev->config.cayman.mem_row_size_in_kb = (4 * (1 << (8 + tmp))) / 1024;
 681	if (rdev->config.cayman.mem_row_size_in_kb > 4)
 682		rdev->config.cayman.mem_row_size_in_kb = 4;
 683	/* XXX use MC settings? */
 684	rdev->config.cayman.shader_engine_tile_size = 32;
 685	rdev->config.cayman.num_gpus = 1;
 686	rdev->config.cayman.multi_gpu_tile_size = 64;
 687
 688	//gb_addr_config = 0x02011003
 689#if 0
 690	gb_addr_config = RREG32(GB_ADDR_CONFIG);
 691#else
 692	gb_addr_config = 0;
 693	switch (rdev->config.cayman.num_tile_pipes) {
 694	case 1:
 695	default:
 696		gb_addr_config |= NUM_PIPES(0);
 697		break;
 698	case 2:
 699		gb_addr_config |= NUM_PIPES(1);
 700		break;
 701	case 4:
 702		gb_addr_config |= NUM_PIPES(2);
 703		break;
 704	case 8:
 705		gb_addr_config |= NUM_PIPES(3);
 706		break;
 707	}
 708
 709	tmp = (rdev->config.cayman.mem_max_burst_length_bytes / 256) - 1;
 710	gb_addr_config |= PIPE_INTERLEAVE_SIZE(tmp);
 711	gb_addr_config |= NUM_SHADER_ENGINES(rdev->config.cayman.num_shader_engines - 1);
 712	tmp = (rdev->config.cayman.shader_engine_tile_size / 16) - 1;
 713	gb_addr_config |= SHADER_ENGINE_TILE_SIZE(tmp);
 714	switch (rdev->config.cayman.num_gpus) {
 715	case 1:
 716	default:
 717		gb_addr_config |= NUM_GPUS(0);
 718		break;
 719	case 2:
 720		gb_addr_config |= NUM_GPUS(1);
 721		break;
 722	case 4:
 723		gb_addr_config |= NUM_GPUS(2);
 724		break;
 725	}
 726	switch (rdev->config.cayman.multi_gpu_tile_size) {
 727	case 16:
 728		gb_addr_config |= MULTI_GPU_TILE_SIZE(0);
 729		break;
 730	case 32:
 731	default:
 732		gb_addr_config |= MULTI_GPU_TILE_SIZE(1);
 733		break;
 734	case 64:
 735		gb_addr_config |= MULTI_GPU_TILE_SIZE(2);
 736		break;
 737	case 128:
 738		gb_addr_config |= MULTI_GPU_TILE_SIZE(3);
 739		break;
 740	}
 741	switch (rdev->config.cayman.mem_row_size_in_kb) {
 742	case 1:
 743	default:
 744		gb_addr_config |= ROW_SIZE(0);
 745		break;
 746	case 2:
 747		gb_addr_config |= ROW_SIZE(1);
 748		break;
 749	case 4:
 750		gb_addr_config |= ROW_SIZE(2);
 751		break;
 752	}
 753#endif
 754
 755	tmp = (gb_addr_config & NUM_PIPES_MASK) >> NUM_PIPES_SHIFT;
 756	rdev->config.cayman.num_tile_pipes = (1 << tmp);
 757	tmp = (gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT;
 758	rdev->config.cayman.mem_max_burst_length_bytes = (tmp + 1) * 256;
 759	tmp = (gb_addr_config & NUM_SHADER_ENGINES_MASK) >> NUM_SHADER_ENGINES_SHIFT;
 760	rdev->config.cayman.num_shader_engines = tmp + 1;
 761	tmp = (gb_addr_config & NUM_GPUS_MASK) >> NUM_GPUS_SHIFT;
 762	rdev->config.cayman.num_gpus = tmp + 1;
 763	tmp = (gb_addr_config & MULTI_GPU_TILE_SIZE_MASK) >> MULTI_GPU_TILE_SIZE_SHIFT;
 764	rdev->config.cayman.multi_gpu_tile_size = 1 << tmp;
 765	tmp = (gb_addr_config & ROW_SIZE_MASK) >> ROW_SIZE_SHIFT;
 766	rdev->config.cayman.mem_row_size_in_kb = 1 << tmp;
 767
 768	//gb_backend_map = 0x76541032;
 769#if 0
 770	gb_backend_map = RREG32(GB_BACKEND_MAP);
 771#else
 772	gb_backend_map =
 773		cayman_get_tile_pipe_to_backend_map(rdev, rdev->config.cayman.num_tile_pipes,
 774						    rdev->config.cayman.num_backends_per_se *
 775						    rdev->config.cayman.num_shader_engines,
 776						    &rdev->config.cayman.backend_disable_mask_per_asic,
 777						    rdev->config.cayman.num_shader_engines);
 778#endif
 779	/* setup tiling info dword.  gb_addr_config is not adequate since it does
 780	 * not have bank info, so create a custom tiling dword.
 781	 * bits 3:0   num_pipes
 782	 * bits 7:4   num_banks
 783	 * bits 11:8  group_size
 784	 * bits 15:12 row_size
 785	 */
 786	rdev->config.cayman.tile_config = 0;
 787	switch (rdev->config.cayman.num_tile_pipes) {
 788	case 1:
 789	default:
 790		rdev->config.cayman.tile_config |= (0 << 0);
 791		break;
 792	case 2:
 793		rdev->config.cayman.tile_config |= (1 << 0);
 794		break;
 795	case 4:
 796		rdev->config.cayman.tile_config |= (2 << 0);
 797		break;
 798	case 8:
 799		rdev->config.cayman.tile_config |= (3 << 0);
 800		break;
 801	}
 802	rdev->config.cayman.tile_config |=
 803		((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) << 4;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 804	rdev->config.cayman.tile_config |=
 805		((gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT) << 8;
 806	rdev->config.cayman.tile_config |=
 807		((gb_addr_config & ROW_SIZE_MASK) >> ROW_SIZE_SHIFT) << 12;
 808
 809	rdev->config.cayman.backend_map = gb_backend_map;
 810	WREG32(GB_BACKEND_MAP, gb_backend_map);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 811	WREG32(GB_ADDR_CONFIG, gb_addr_config);
 812	WREG32(DMIF_ADDR_CONFIG, gb_addr_config);
 
 
 813	WREG32(HDP_ADDR_CONFIG, gb_addr_config);
 814
 815	/* primary versions */
 816	WREG32(CC_RB_BACKEND_DISABLE, cc_rb_backend_disable);
 817	WREG32(CC_SYS_RB_BACKEND_DISABLE, cc_rb_backend_disable);
 818	WREG32(CC_GC_SHADER_PIPE_CONFIG, cc_gc_shader_pipe_config);
 819
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 820	WREG32(CGTS_TCC_DISABLE, cgts_tcc_disable);
 821	WREG32(CGTS_SYS_TCC_DISABLE, cgts_tcc_disable);
 822
 823	/* user versions */
 824	WREG32(GC_USER_RB_BACKEND_DISABLE, cc_rb_backend_disable);
 825	WREG32(GC_USER_SYS_RB_BACKEND_DISABLE, cc_rb_backend_disable);
 826	WREG32(GC_USER_SHADER_PIPE_CONFIG, cc_gc_shader_pipe_config);
 827
 828	WREG32(CGTS_USER_SYS_TCC_DISABLE, cgts_tcc_disable);
 829	WREG32(CGTS_USER_TCC_DISABLE, cgts_tcc_disable);
 830
 831	/* reprogram the shader complex */
 832	cgts_sm_ctrl_reg = RREG32(CGTS_SM_CTRL_REG);
 833	for (i = 0; i < 16; i++)
 834		WREG32(CGTS_SM_CTRL_REG, OVERRIDE);
 835	WREG32(CGTS_SM_CTRL_REG, cgts_sm_ctrl_reg);
 836
 837	/* set HW defaults for 3D engine */
 838	WREG32(CP_MEQ_THRESHOLDS, MEQ1_START(0x30) | MEQ2_START(0x60));
 839
 840	sx_debug_1 = RREG32(SX_DEBUG_1);
 841	sx_debug_1 |= ENABLE_NEW_SMX_ADDRESS;
 842	WREG32(SX_DEBUG_1, sx_debug_1);
 843
 844	smx_dc_ctl0 = RREG32(SMX_DC_CTL0);
 845	smx_dc_ctl0 &= ~NUMBER_OF_SETS(0x1ff);
 846	smx_dc_ctl0 |= NUMBER_OF_SETS(rdev->config.cayman.sx_num_of_sets);
 847	WREG32(SMX_DC_CTL0, smx_dc_ctl0);
 848
 849	WREG32(SPI_CONFIG_CNTL_1, VTX_DONE_DELAY(4) | CRC_SIMD_ID_WADDR_DISABLE);
 850
 851	/* need to be explicitly zero-ed */
 852	WREG32(VGT_OFFCHIP_LDS_BASE, 0);
 853	WREG32(SQ_LSTMP_RING_BASE, 0);
 854	WREG32(SQ_HSTMP_RING_BASE, 0);
 855	WREG32(SQ_ESTMP_RING_BASE, 0);
 856	WREG32(SQ_GSTMP_RING_BASE, 0);
 857	WREG32(SQ_VSTMP_RING_BASE, 0);
 858	WREG32(SQ_PSTMP_RING_BASE, 0);
 859
 860	WREG32(TA_CNTL_AUX, DISABLE_CUBE_ANISO);
 861
 862	WREG32(SX_EXPORT_BUFFER_SIZES, (COLOR_BUFFER_SIZE((rdev->config.cayman.sx_max_export_size / 4) - 1) |
 863					POSITION_BUFFER_SIZE((rdev->config.cayman.sx_max_export_pos_size / 4) - 1) |
 864					SMX_BUFFER_SIZE((rdev->config.cayman.sx_max_export_smx_size / 4) - 1)));
 865
 866	WREG32(PA_SC_FIFO_SIZE, (SC_PRIM_FIFO_SIZE(rdev->config.cayman.sc_prim_fifo_size) |
 867				 SC_HIZ_TILE_FIFO_SIZE(rdev->config.cayman.sc_hiz_tile_fifo_size) |
 868				 SC_EARLYZ_TILE_FIFO_SIZE(rdev->config.cayman.sc_earlyz_tile_fifo_size)));
 869
 870
 871	WREG32(VGT_NUM_INSTANCES, 1);
 872
 873	WREG32(CP_PERFMON_CNTL, 0);
 874
 875	WREG32(SQ_MS_FIFO_SIZES, (CACHE_FIFO_SIZE(16 * rdev->config.cayman.sq_num_cf_insts) |
 876				  FETCH_FIFO_HIWATER(0x4) |
 877				  DONE_FIFO_HIWATER(0xe0) |
 878				  ALU_UPDATE_FIFO_HIWATER(0x8)));
 879
 880	WREG32(SQ_GPR_RESOURCE_MGMT_1, NUM_CLAUSE_TEMP_GPRS(4));
 881	WREG32(SQ_CONFIG, (VC_ENABLE |
 882			   EXPORT_SRC_C |
 883			   GFX_PRIO(0) |
 884			   CS1_PRIO(0) |
 885			   CS2_PRIO(1)));
 886	WREG32(SQ_DYN_GPR_CNTL_PS_FLUSH_REQ, DYN_GPR_ENABLE);
 887
 888	WREG32(PA_SC_FORCE_EOV_MAX_CNTS, (FORCE_EOV_MAX_CLK_CNT(4095) |
 889					  FORCE_EOV_MAX_REZ_CNT(255)));
 890
 891	WREG32(VGT_CACHE_INVALIDATION, CACHE_INVALIDATION(VC_AND_TC) |
 892	       AUTO_INVLD_EN(ES_AND_GS_AUTO));
 893
 894	WREG32(VGT_GS_VERTEX_REUSE, 16);
 895	WREG32(PA_SC_LINE_STIPPLE_STATE, 0);
 896
 897	WREG32(CB_PERF_CTR0_SEL_0, 0);
 898	WREG32(CB_PERF_CTR0_SEL_1, 0);
 899	WREG32(CB_PERF_CTR1_SEL_0, 0);
 900	WREG32(CB_PERF_CTR1_SEL_1, 0);
 901	WREG32(CB_PERF_CTR2_SEL_0, 0);
 902	WREG32(CB_PERF_CTR2_SEL_1, 0);
 903	WREG32(CB_PERF_CTR3_SEL_0, 0);
 904	WREG32(CB_PERF_CTR3_SEL_1, 0);
 905
 906	tmp = RREG32(HDP_MISC_CNTL);
 907	tmp |= HDP_FLUSH_INVALIDATE_CACHE;
 908	WREG32(HDP_MISC_CNTL, tmp);
 909
 910	hdp_host_path_cntl = RREG32(HDP_HOST_PATH_CNTL);
 911	WREG32(HDP_HOST_PATH_CNTL, hdp_host_path_cntl);
 912
 913	WREG32(PA_CL_ENHANCE, CLIP_VTX_REORDER_ENA | NUM_CLIP_SEQ(3));
 914
 915	udelay(50);
 
 
 
 
 
 
 
 
 
 
 916}
 917
 918/*
 919 * GART
 920 */
 921void cayman_pcie_gart_tlb_flush(struct radeon_device *rdev)
 922{
 923	/* flush hdp cache */
 924	WREG32(HDP_MEM_COHERENCY_FLUSH_CNTL, 0x1);
 925
 926	/* bits 0-7 are the VM contexts0-7 */
 927	WREG32(VM_INVALIDATE_REQUEST, 1);
 928}
 929
 930int cayman_pcie_gart_enable(struct radeon_device *rdev)
 931{
 932	int r;
 933
 934	if (rdev->gart.table.vram.robj == NULL) {
 935		dev_err(rdev->dev, "No VRAM object for PCIE GART.\n");
 936		return -EINVAL;
 937	}
 938	r = radeon_gart_table_vram_pin(rdev);
 939	if (r)
 940		return r;
 941	radeon_gart_restore(rdev);
 942	/* Setup TLB control */
 943	WREG32(MC_VM_MX_L1_TLB_CNTL, ENABLE_L1_TLB |
 
 
 944	       ENABLE_L1_FRAGMENT_PROCESSING |
 945	       SYSTEM_ACCESS_MODE_NOT_IN_SYS |
 
 946	       SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU);
 947	/* Setup L2 cache */
 948	WREG32(VM_L2_CNTL, ENABLE_L2_CACHE |
 
 949	       ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
 950	       ENABLE_L2_PDE0_CACHE_LRU_UPDATE_BY_WRITE |
 951	       EFFECTIVE_L2_QUEUE_SIZE(7) |
 952	       CONTEXT1_IDENTITY_ACCESS_MODE(1));
 953	WREG32(VM_L2_CNTL2, INVALIDATE_ALL_L1_TLBS | INVALIDATE_L2_CACHE);
 954	WREG32(VM_L2_CNTL3, L2_CACHE_BIGK_ASSOCIATIVITY |
 
 955	       L2_CACHE_BIGK_FRAGMENT_SIZE(6));
 956	/* setup context0 */
 957	WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR, rdev->mc.gtt_start >> 12);
 958	WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR, rdev->mc.gtt_end >> 12);
 959	WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12);
 960	WREG32(VM_CONTEXT0_PROTECTION_FAULT_DEFAULT_ADDR,
 961			(u32)(rdev->dummy_page.addr >> 12));
 962	WREG32(VM_CONTEXT0_CNTL2, 0);
 963	WREG32(VM_CONTEXT0_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) |
 964				RANGE_PROTECTION_FAULT_ENABLE_DEFAULT);
 965	/* disable context1-7 */
 966	WREG32(VM_CONTEXT1_CNTL2, 0);
 967	WREG32(VM_CONTEXT1_CNTL, 0);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 968
 969	cayman_pcie_gart_tlb_flush(rdev);
 
 
 
 970	rdev->gart.ready = true;
 971	return 0;
 972}
 973
 974void cayman_pcie_gart_disable(struct radeon_device *rdev)
 975{
 976	int r;
 
 
 
 
 
 977
 978	/* Disable all tables */
 979	WREG32(VM_CONTEXT0_CNTL, 0);
 980	WREG32(VM_CONTEXT1_CNTL, 0);
 981	/* Setup TLB control */
 982	WREG32(MC_VM_MX_L1_TLB_CNTL, ENABLE_L1_FRAGMENT_PROCESSING |
 983	       SYSTEM_ACCESS_MODE_NOT_IN_SYS |
 984	       SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU);
 985	/* Setup L2 cache */
 986	WREG32(VM_L2_CNTL, ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
 987	       ENABLE_L2_PDE0_CACHE_LRU_UPDATE_BY_WRITE |
 988	       EFFECTIVE_L2_QUEUE_SIZE(7) |
 989	       CONTEXT1_IDENTITY_ACCESS_MODE(1));
 990	WREG32(VM_L2_CNTL2, 0);
 991	WREG32(VM_L2_CNTL3, L2_CACHE_BIGK_ASSOCIATIVITY |
 992	       L2_CACHE_BIGK_FRAGMENT_SIZE(6));
 993	if (rdev->gart.table.vram.robj) {
 994		r = radeon_bo_reserve(rdev->gart.table.vram.robj, false);
 995		if (likely(r == 0)) {
 996			radeon_bo_kunmap(rdev->gart.table.vram.robj);
 997			radeon_bo_unpin(rdev->gart.table.vram.robj);
 998			radeon_bo_unreserve(rdev->gart.table.vram.robj);
 999		}
1000	}
1001}
1002
1003void cayman_pcie_gart_fini(struct radeon_device *rdev)
1004{
1005	cayman_pcie_gart_disable(rdev);
1006	radeon_gart_table_vram_free(rdev);
1007	radeon_gart_fini(rdev);
1008}
1009
 
 
 
 
 
 
 
1010/*
1011 * CP.
1012 */
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1013static void cayman_cp_enable(struct radeon_device *rdev, bool enable)
1014{
1015	if (enable)
1016		WREG32(CP_ME_CNTL, 0);
1017	else {
1018		radeon_ttm_set_active_vram_size(rdev, rdev->mc.visible_vram_size);
 
1019		WREG32(CP_ME_CNTL, (CP_ME_HALT | CP_PFP_HALT));
1020		WREG32(SCRATCH_UMSK, 0);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1021	}
1022}
1023
1024static int cayman_cp_load_microcode(struct radeon_device *rdev)
1025{
1026	const __be32 *fw_data;
1027	int i;
1028
1029	if (!rdev->me_fw || !rdev->pfp_fw)
1030		return -EINVAL;
1031
1032	cayman_cp_enable(rdev, false);
1033
1034	fw_data = (const __be32 *)rdev->pfp_fw->data;
1035	WREG32(CP_PFP_UCODE_ADDR, 0);
1036	for (i = 0; i < CAYMAN_PFP_UCODE_SIZE; i++)
1037		WREG32(CP_PFP_UCODE_DATA, be32_to_cpup(fw_data++));
1038	WREG32(CP_PFP_UCODE_ADDR, 0);
1039
1040	fw_data = (const __be32 *)rdev->me_fw->data;
1041	WREG32(CP_ME_RAM_WADDR, 0);
1042	for (i = 0; i < CAYMAN_PM4_UCODE_SIZE; i++)
1043		WREG32(CP_ME_RAM_DATA, be32_to_cpup(fw_data++));
1044
1045	WREG32(CP_PFP_UCODE_ADDR, 0);
1046	WREG32(CP_ME_RAM_WADDR, 0);
1047	WREG32(CP_ME_RAM_RADDR, 0);
1048	return 0;
1049}
1050
1051static int cayman_cp_start(struct radeon_device *rdev)
1052{
 
1053	int r, i;
1054
1055	r = radeon_ring_lock(rdev, 7);
1056	if (r) {
1057		DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
1058		return r;
1059	}
1060	radeon_ring_write(rdev, PACKET3(PACKET3_ME_INITIALIZE, 5));
1061	radeon_ring_write(rdev, 0x1);
1062	radeon_ring_write(rdev, 0x0);
1063	radeon_ring_write(rdev, rdev->config.cayman.max_hw_contexts - 1);
1064	radeon_ring_write(rdev, PACKET3_ME_INITIALIZE_DEVICE_ID(1));
1065	radeon_ring_write(rdev, 0);
1066	radeon_ring_write(rdev, 0);
1067	radeon_ring_unlock_commit(rdev);
1068
1069	cayman_cp_enable(rdev, true);
1070
1071	r = radeon_ring_lock(rdev, cayman_default_size + 19);
1072	if (r) {
1073		DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
1074		return r;
1075	}
1076
1077	/* setup clear context state */
1078	radeon_ring_write(rdev, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
1079	radeon_ring_write(rdev, PACKET3_PREAMBLE_BEGIN_CLEAR_STATE);
1080
1081	for (i = 0; i < cayman_default_size; i++)
1082		radeon_ring_write(rdev, cayman_default_state[i]);
1083
1084	radeon_ring_write(rdev, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
1085	radeon_ring_write(rdev, PACKET3_PREAMBLE_END_CLEAR_STATE);
1086
1087	/* set clear context state */
1088	radeon_ring_write(rdev, PACKET3(PACKET3_CLEAR_STATE, 0));
1089	radeon_ring_write(rdev, 0);
1090
1091	/* SQ_VTX_BASE_VTX_LOC */
1092	radeon_ring_write(rdev, 0xc0026f00);
1093	radeon_ring_write(rdev, 0x00000000);
1094	radeon_ring_write(rdev, 0x00000000);
1095	radeon_ring_write(rdev, 0x00000000);
1096
1097	/* Clear consts */
1098	radeon_ring_write(rdev, 0xc0036f00);
1099	radeon_ring_write(rdev, 0x00000bc4);
1100	radeon_ring_write(rdev, 0xffffffff);
1101	radeon_ring_write(rdev, 0xffffffff);
1102	radeon_ring_write(rdev, 0xffffffff);
1103
1104	radeon_ring_write(rdev, 0xc0026900);
1105	radeon_ring_write(rdev, 0x00000316);
1106	radeon_ring_write(rdev, 0x0000000e); /* VGT_VERTEX_REUSE_BLOCK_CNTL */
1107	radeon_ring_write(rdev, 0x00000010); /*  */
1108
1109	radeon_ring_unlock_commit(rdev);
1110
1111	/* XXX init other rings */
1112
1113	return 0;
1114}
1115
1116static void cayman_cp_fini(struct radeon_device *rdev)
1117{
 
1118	cayman_cp_enable(rdev, false);
1119	radeon_ring_fini(rdev);
 
1120}
1121
1122int cayman_cp_resume(struct radeon_device *rdev)
1123{
1124	u32 tmp;
1125	u32 rb_bufsz;
1126	int r;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1127
1128	/* Reset cp; if cp is reset, then PA, SH, VGT also need to be reset */
1129	WREG32(GRBM_SOFT_RESET, (SOFT_RESET_CP |
1130				 SOFT_RESET_PA |
1131				 SOFT_RESET_SH |
1132				 SOFT_RESET_VGT |
1133				 SOFT_RESET_SPI |
1134				 SOFT_RESET_SX));
1135	RREG32(GRBM_SOFT_RESET);
1136	mdelay(15);
1137	WREG32(GRBM_SOFT_RESET, 0);
1138	RREG32(GRBM_SOFT_RESET);
1139
1140	WREG32(CP_SEM_WAIT_TIMER, 0x4);
 
1141
1142	/* Set the write pointer delay */
1143	WREG32(CP_RB_WPTR_DELAY, 0);
1144
1145	WREG32(CP_DEBUG, (1 << 27));
1146
1147	/* ring 0 - compute and gfx */
1148	/* Set ring buffer size */
1149	rb_bufsz = drm_order(rdev->cp.ring_size / 8);
1150	tmp = (drm_order(RADEON_GPU_PAGE_SIZE/8) << 8) | rb_bufsz;
 
 
 
 
 
 
 
 
1151#ifdef __BIG_ENDIAN
1152	tmp |= BUF_SWAP_32BIT;
1153#endif
1154	WREG32(CP_RB0_CNTL, tmp);
1155
1156	/* Initialize the ring buffer's read and write pointers */
1157	WREG32(CP_RB0_CNTL, tmp | RB_RPTR_WR_ENA);
1158	rdev->cp.wptr = 0;
1159	WREG32(CP_RB0_WPTR, rdev->cp.wptr);
1160
1161	/* set the wb address wether it's enabled or not */
1162	WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC);
1163	WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
1164	WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
1165
1166	if (rdev->wb.enabled)
1167		WREG32(SCRATCH_UMSK, 0xff);
1168	else {
1169		tmp |= RB_NO_UPDATE;
1170		WREG32(SCRATCH_UMSK, 0);
1171	}
1172
1173	mdelay(1);
1174	WREG32(CP_RB0_CNTL, tmp);
 
 
1175
1176	WREG32(CP_RB0_BASE, rdev->cp.gpu_addr >> 8);
 
 
1177
1178	rdev->cp.rptr = RREG32(CP_RB0_RPTR);
 
 
1179
1180	/* ring1  - compute only */
1181	/* Set ring buffer size */
1182	rb_bufsz = drm_order(rdev->cp1.ring_size / 8);
1183	tmp = (drm_order(RADEON_GPU_PAGE_SIZE/8) << 8) | rb_bufsz;
1184#ifdef __BIG_ENDIAN
1185	tmp |= BUF_SWAP_32BIT;
1186#endif
1187	WREG32(CP_RB1_CNTL, tmp);
 
 
 
 
 
1188
1189	/* Initialize the ring buffer's read and write pointers */
1190	WREG32(CP_RB1_CNTL, tmp | RB_RPTR_WR_ENA);
1191	rdev->cp1.wptr = 0;
1192	WREG32(CP_RB1_WPTR, rdev->cp1.wptr);
1193
1194	/* set the wb address wether it's enabled or not */
1195	WREG32(CP_RB1_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFFFFFFFC);
1196	WREG32(CP_RB1_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFF);
1197
1198	mdelay(1);
1199	WREG32(CP_RB1_CNTL, tmp);
 
 
1200
1201	WREG32(CP_RB1_BASE, rdev->cp1.gpu_addr >> 8);
 
 
 
 
 
 
 
 
1202
1203	rdev->cp1.rptr = RREG32(CP_RB1_RPTR);
 
 
1204
1205	/* ring2 - compute only */
1206	/* Set ring buffer size */
1207	rb_bufsz = drm_order(rdev->cp2.ring_size / 8);
1208	tmp = (drm_order(RADEON_GPU_PAGE_SIZE/8) << 8) | rb_bufsz;
1209#ifdef __BIG_ENDIAN
1210	tmp |= BUF_SWAP_32BIT;
1211#endif
1212	WREG32(CP_RB2_CNTL, tmp);
1213
1214	/* Initialize the ring buffer's read and write pointers */
1215	WREG32(CP_RB2_CNTL, tmp | RB_RPTR_WR_ENA);
1216	rdev->cp2.wptr = 0;
1217	WREG32(CP_RB2_WPTR, rdev->cp2.wptr);
1218
1219	/* set the wb address wether it's enabled or not */
1220	WREG32(CP_RB2_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFFFFFFFC);
1221	WREG32(CP_RB2_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFF);
 
1222
1223	mdelay(1);
1224	WREG32(CP_RB2_CNTL, tmp);
 
 
1225
1226	WREG32(CP_RB2_BASE, rdev->cp2.gpu_addr >> 8);
 
1227
1228	rdev->cp2.rptr = RREG32(CP_RB2_RPTR);
 
 
 
1229
1230	/* start the rings */
1231	cayman_cp_start(rdev);
1232	rdev->cp.ready = true;
1233	rdev->cp1.ready = true;
1234	rdev->cp2.ready = true;
1235	/* this only test cp0 */
1236	r = radeon_ring_test(rdev);
1237	if (r) {
1238		rdev->cp.ready = false;
1239		rdev->cp1.ready = false;
1240		rdev->cp2.ready = false;
1241		return r;
1242	}
1243
1244	return 0;
1245}
1246
1247bool cayman_gpu_is_lockup(struct radeon_device *rdev)
1248{
1249	u32 srbm_status;
1250	u32 grbm_status;
1251	u32 grbm_status_se0, grbm_status_se1;
1252	struct r100_gpu_lockup *lockup = &rdev->config.cayman.lockup;
1253	int r;
1254
1255	srbm_status = RREG32(SRBM_STATUS);
1256	grbm_status = RREG32(GRBM_STATUS);
1257	grbm_status_se0 = RREG32(GRBM_STATUS_SE0);
1258	grbm_status_se1 = RREG32(GRBM_STATUS_SE1);
1259	if (!(grbm_status & GUI_ACTIVE)) {
1260		r100_gpu_lockup_update(lockup, &rdev->cp);
1261		return false;
 
 
 
 
 
 
 
 
 
 
 
 
1262	}
1263	/* force CP activities */
1264	r = radeon_ring_lock(rdev, 2);
1265	if (!r) {
1266		/* PACKET2 NOP */
1267		radeon_ring_write(rdev, 0x80000000);
1268		radeon_ring_write(rdev, 0x80000000);
1269		radeon_ring_unlock_commit(rdev);
1270	}
1271	/* XXX deal with CP0,1,2 */
1272	rdev->cp.rptr = RREG32(CP_RB0_RPTR);
1273	return r100_gpu_cp_is_lockup(rdev, lockup, &rdev->cp);
1274}
1275
1276static int cayman_gpu_soft_reset(struct radeon_device *rdev)
1277{
1278	struct evergreen_mc_save save;
1279	u32 grbm_reset = 0;
 
1280
1281	if (!(RREG32(GRBM_STATUS) & GUI_ACTIVE))
1282		return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1283
1284	dev_info(rdev->dev, "GPU softreset \n");
1285	dev_info(rdev->dev, "  GRBM_STATUS=0x%08X\n",
1286		RREG32(GRBM_STATUS));
1287	dev_info(rdev->dev, "  GRBM_STATUS_SE0=0x%08X\n",
1288		RREG32(GRBM_STATUS_SE0));
1289	dev_info(rdev->dev, "  GRBM_STATUS_SE1=0x%08X\n",
1290		RREG32(GRBM_STATUS_SE1));
1291	dev_info(rdev->dev, "  SRBM_STATUS=0x%08X\n",
1292		RREG32(SRBM_STATUS));
1293	evergreen_mc_stop(rdev, &save);
1294	if (evergreen_mc_wait_for_idle(rdev)) {
1295		dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
1296	}
1297	/* Disable CP parsing/prefetching */
1298	WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT);
1299
1300	/* reset all the gfx blocks */
1301	grbm_reset = (SOFT_RESET_CP |
1302		      SOFT_RESET_CB |
1303		      SOFT_RESET_DB |
1304		      SOFT_RESET_GDS |
1305		      SOFT_RESET_PA |
1306		      SOFT_RESET_SC |
1307		      SOFT_RESET_SPI |
1308		      SOFT_RESET_SH |
1309		      SOFT_RESET_SX |
1310		      SOFT_RESET_TC |
1311		      SOFT_RESET_TA |
1312		      SOFT_RESET_VGT |
1313		      SOFT_RESET_IA);
1314
1315	dev_info(rdev->dev, "  GRBM_SOFT_RESET=0x%08X\n", grbm_reset);
1316	WREG32(GRBM_SOFT_RESET, grbm_reset);
1317	(void)RREG32(GRBM_SOFT_RESET);
1318	udelay(50);
1319	WREG32(GRBM_SOFT_RESET, 0);
1320	(void)RREG32(GRBM_SOFT_RESET);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1321	/* Wait a little for things to settle down */
1322	udelay(50);
1323	dev_info(rdev->dev, "  GRBM_STATUS=0x%08X\n",
1324		RREG32(GRBM_STATUS));
1325	dev_info(rdev->dev, "  GRBM_STATUS_SE0=0x%08X\n",
1326		RREG32(GRBM_STATUS_SE0));
1327	dev_info(rdev->dev, "  GRBM_STATUS_SE1=0x%08X\n",
1328		RREG32(GRBM_STATUS_SE1));
1329	dev_info(rdev->dev, "  SRBM_STATUS=0x%08X\n",
1330		RREG32(SRBM_STATUS));
1331	evergreen_mc_resume(rdev, &save);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1332	return 0;
1333}
1334
1335int cayman_asic_reset(struct radeon_device *rdev)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1336{
1337	return cayman_gpu_soft_reset(rdev);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1338}
1339
1340static int cayman_startup(struct radeon_device *rdev)
1341{
 
1342	int r;
1343
1344	if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw || !rdev->mc_fw) {
1345		r = ni_init_microcode(rdev);
 
 
 
 
 
 
 
 
 
 
 
 
1346		if (r) {
1347			DRM_ERROR("Failed to load firmware!\n");
1348			return r;
1349		}
1350	}
1351	r = ni_mc_load_microcode(rdev);
1352	if (r) {
1353		DRM_ERROR("Failed to load MC firmware!\n");
1354		return r;
1355	}
1356
1357	evergreen_mc_program(rdev);
1358	r = cayman_pcie_gart_enable(rdev);
1359	if (r)
1360		return r;
1361	cayman_gpu_init(rdev);
1362
1363	r = evergreen_blit_init(rdev);
1364	if (r) {
1365		evergreen_blit_fini(rdev);
1366		rdev->asic->copy = NULL;
1367		dev_warn(rdev->dev, "failed blitter (%d) falling back to memcpy\n", r);
 
 
 
 
 
 
1368	}
1369
1370	/* allocate wb buffer */
1371	r = radeon_wb_init(rdev);
1372	if (r)
1373		return r;
1374
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1375	/* Enable IRQ */
 
 
 
 
 
 
1376	r = r600_irq_init(rdev);
1377	if (r) {
1378		DRM_ERROR("radeon: IH init failed (%d).\n", r);
1379		radeon_irq_kms_fini(rdev);
1380		return r;
1381	}
1382	evergreen_irq_set(rdev);
1383
1384	r = radeon_ring_init(rdev, rdev->cp.ring_size);
 
 
 
 
 
 
 
1385	if (r)
1386		return r;
 
 
 
 
 
 
 
1387	r = cayman_cp_load_microcode(rdev);
1388	if (r)
1389		return r;
1390	r = cayman_cp_resume(rdev);
1391	if (r)
1392		return r;
1393
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1394	return 0;
1395}
1396
1397int cayman_resume(struct radeon_device *rdev)
1398{
1399	int r;
1400
1401	/* Do not reset GPU before posting, on rv770 hw unlike on r500 hw,
1402	 * posting will perform necessary task to bring back GPU into good
1403	 * shape.
1404	 */
1405	/* post card */
1406	atom_asic_init(rdev->mode_info.atom_context);
1407
 
 
 
 
 
 
 
1408	r = cayman_startup(rdev);
1409	if (r) {
1410		DRM_ERROR("cayman startup failed on resume\n");
 
1411		return r;
1412	}
1413
1414	r = r600_ib_test(rdev);
1415	if (r) {
1416		DRM_ERROR("radeon: failled testing IB (%d).\n", r);
1417		return r;
1418	}
1419
1420	return r;
1421
1422}
1423
1424int cayman_suspend(struct radeon_device *rdev)
1425{
1426	int r;
1427
1428	/* FIXME: we should wait for ring to be empty */
1429	cayman_cp_enable(rdev, false);
1430	rdev->cp.ready = false;
 
 
 
 
1431	evergreen_irq_suspend(rdev);
1432	radeon_wb_disable(rdev);
1433	cayman_pcie_gart_disable(rdev);
1434
1435	/* unpin shaders bo */
1436	r = radeon_bo_reserve(rdev->r600_blit.shader_obj, false);
1437	if (likely(r == 0)) {
1438		radeon_bo_unpin(rdev->r600_blit.shader_obj);
1439		radeon_bo_unreserve(rdev->r600_blit.shader_obj);
1440	}
1441
1442	return 0;
1443}
1444
1445/* Plan is to move initialization in that function and use
1446 * helper function so that radeon_device_init pretty much
1447 * do nothing more than calling asic specific function. This
1448 * should also allow to remove a bunch of callback function
1449 * like vram_info.
1450 */
1451int cayman_init(struct radeon_device *rdev)
1452{
 
1453	int r;
1454
1455	/* This don't do much */
1456	r = radeon_gem_init(rdev);
1457	if (r)
1458		return r;
1459	/* Read BIOS */
1460	if (!radeon_get_bios(rdev)) {
1461		if (ASIC_IS_AVIVO(rdev))
1462			return -EINVAL;
1463	}
1464	/* Must be an ATOMBIOS */
1465	if (!rdev->is_atom_bios) {
1466		dev_err(rdev->dev, "Expecting atombios for cayman GPU\n");
1467		return -EINVAL;
1468	}
1469	r = radeon_atombios_init(rdev);
1470	if (r)
1471		return r;
1472
1473	/* Post card if necessary */
1474	if (!radeon_card_posted(rdev)) {
1475		if (!rdev->bios) {
1476			dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n");
1477			return -EINVAL;
1478		}
1479		DRM_INFO("GPU not posted. posting now...\n");
1480		atom_asic_init(rdev->mode_info.atom_context);
1481	}
 
 
1482	/* Initialize scratch registers */
1483	r600_scratch_init(rdev);
1484	/* Initialize surface registers */
1485	radeon_surface_init(rdev);
1486	/* Initialize clocks */
1487	radeon_get_clock_info(rdev->ddev);
1488	/* Fence driver */
1489	r = radeon_fence_driver_init(rdev);
1490	if (r)
1491		return r;
1492	/* initialize memory controller */
1493	r = evergreen_mc_init(rdev);
1494	if (r)
1495		return r;
1496	/* Memory manager */
1497	r = radeon_bo_init(rdev);
1498	if (r)
1499		return r;
1500
1501	r = radeon_irq_kms_init(rdev);
1502	if (r)
1503		return r;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1504
1505	rdev->cp.ring_obj = NULL;
1506	r600_ring_init(rdev, 1024 * 1024);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1507
1508	rdev->ih.ring_obj = NULL;
1509	r600_ih_ring_init(rdev, 64 * 1024);
1510
1511	r = r600_pcie_gart_init(rdev);
1512	if (r)
1513		return r;
1514
1515	rdev->accel_working = true;
1516	r = cayman_startup(rdev);
1517	if (r) {
1518		dev_err(rdev->dev, "disabling GPU acceleration\n");
1519		cayman_cp_fini(rdev);
 
1520		r600_irq_fini(rdev);
 
 
1521		radeon_wb_fini(rdev);
 
 
1522		radeon_irq_kms_fini(rdev);
1523		cayman_pcie_gart_fini(rdev);
1524		rdev->accel_working = false;
1525	}
1526	if (rdev->accel_working) {
1527		r = radeon_ib_pool_init(rdev);
1528		if (r) {
1529			DRM_ERROR("radeon: failed initializing IB pool (%d).\n", r);
1530			rdev->accel_working = false;
1531		}
1532		r = r600_ib_test(rdev);
1533		if (r) {
1534			DRM_ERROR("radeon: failed testing IB (%d).\n", r);
1535			rdev->accel_working = false;
1536		}
1537	}
1538
1539	/* Don't start up if the MC ucode is missing.
1540	 * The default clocks and voltages before the MC ucode
1541	 * is loaded are not suffient for advanced operations.
 
 
 
1542	 */
1543	if (!rdev->mc_fw) {
1544		DRM_ERROR("radeon: MC ucode required for NI+.\n");
1545		return -EINVAL;
1546	}
1547
1548	return 0;
1549}
1550
1551void cayman_fini(struct radeon_device *rdev)
1552{
1553	evergreen_blit_fini(rdev);
1554	cayman_cp_fini(rdev);
 
1555	r600_irq_fini(rdev);
 
 
1556	radeon_wb_fini(rdev);
 
1557	radeon_ib_pool_fini(rdev);
1558	radeon_irq_kms_fini(rdev);
 
 
 
 
1559	cayman_pcie_gart_fini(rdev);
 
1560	radeon_gem_fini(rdev);
1561	radeon_fence_driver_fini(rdev);
1562	radeon_bo_fini(rdev);
1563	radeon_atombios_fini(rdev);
1564	kfree(rdev->bios);
1565	rdev->bios = NULL;
1566}
1567
v6.13.7
   1/*
   2 * Copyright 2010 Advanced Micro Devices, Inc.
   3 *
   4 * Permission is hereby granted, free of charge, to any person obtaining a
   5 * copy of this software and associated documentation files (the "Software"),
   6 * to deal in the Software without restriction, including without limitation
   7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
   8 * and/or sell copies of the Software, and to permit persons to whom the
   9 * Software is furnished to do so, subject to the following conditions:
  10 *
  11 * The above copyright notice and this permission notice shall be included in
  12 * all copies or substantial portions of the Software.
  13 *
  14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
  17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
  18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
  19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
  20 * OTHER DEALINGS IN THE SOFTWARE.
  21 *
  22 * Authors: Alex Deucher
  23 */
  24
  25#include <linux/firmware.h>
  26#include <linux/module.h>
  27#include <linux/pci.h>
  28#include <linux/slab.h>
  29
  30#include <drm/radeon_drm.h>
  31
 
 
  32#include "atom.h"
 
  33#include "cayman_blit_shaders.h"
  34#include "clearstate_cayman.h"
  35#include "evergreen.h"
  36#include "ni.h"
  37#include "ni_reg.h"
  38#include "nid.h"
  39#include "radeon.h"
  40#include "radeon_asic.h"
  41#include "radeon_audio.h"
  42#include "radeon_ucode.h"
  43
  44/*
  45 * Indirect registers accessor
  46 */
  47u32 tn_smc_rreg(struct radeon_device *rdev, u32 reg)
  48{
  49	unsigned long flags;
  50	u32 r;
  51
  52	spin_lock_irqsave(&rdev->smc_idx_lock, flags);
  53	WREG32(TN_SMC_IND_INDEX_0, (reg));
  54	r = RREG32(TN_SMC_IND_DATA_0);
  55	spin_unlock_irqrestore(&rdev->smc_idx_lock, flags);
  56	return r;
  57}
  58
  59void tn_smc_wreg(struct radeon_device *rdev, u32 reg, u32 v)
  60{
  61	unsigned long flags;
  62
  63	spin_lock_irqsave(&rdev->smc_idx_lock, flags);
  64	WREG32(TN_SMC_IND_INDEX_0, (reg));
  65	WREG32(TN_SMC_IND_DATA_0, (v));
  66	spin_unlock_irqrestore(&rdev->smc_idx_lock, flags);
  67}
  68
  69static const u32 tn_rlc_save_restore_register_list[] = {
  70	0x98fc,
  71	0x98f0,
  72	0x9834,
  73	0x9838,
  74	0x9870,
  75	0x9874,
  76	0x8a14,
  77	0x8b24,
  78	0x8bcc,
  79	0x8b10,
  80	0x8c30,
  81	0x8d00,
  82	0x8d04,
  83	0x8c00,
  84	0x8c04,
  85	0x8c10,
  86	0x8c14,
  87	0x8d8c,
  88	0x8cf0,
  89	0x8e38,
  90	0x9508,
  91	0x9688,
  92	0x9608,
  93	0x960c,
  94	0x9610,
  95	0x9614,
  96	0x88c4,
  97	0x8978,
  98	0x88d4,
  99	0x900c,
 100	0x9100,
 101	0x913c,
 102	0x90e8,
 103	0x9354,
 104	0xa008,
 105	0x98f8,
 106	0x9148,
 107	0x914c,
 108	0x3f94,
 109	0x98f4,
 110	0x9b7c,
 111	0x3f8c,
 112	0x8950,
 113	0x8954,
 114	0x8a18,
 115	0x8b28,
 116	0x9144,
 117	0x3f90,
 118	0x915c,
 119	0x9160,
 120	0x9178,
 121	0x917c,
 122	0x9180,
 123	0x918c,
 124	0x9190,
 125	0x9194,
 126	0x9198,
 127	0x919c,
 128	0x91a8,
 129	0x91ac,
 130	0x91b0,
 131	0x91b4,
 132	0x91b8,
 133	0x91c4,
 134	0x91c8,
 135	0x91cc,
 136	0x91d0,
 137	0x91d4,
 138	0x91e0,
 139	0x91e4,
 140	0x91ec,
 141	0x91f0,
 142	0x91f4,
 143	0x9200,
 144	0x9204,
 145	0x929c,
 146	0x8030,
 147	0x9150,
 148	0x9a60,
 149	0x920c,
 150	0x9210,
 151	0x9228,
 152	0x922c,
 153	0x9244,
 154	0x9248,
 155	0x91e8,
 156	0x9294,
 157	0x9208,
 158	0x9224,
 159	0x9240,
 160	0x9220,
 161	0x923c,
 162	0x9258,
 163	0x9744,
 164	0xa200,
 165	0xa204,
 166	0xa208,
 167	0xa20c,
 168	0x8d58,
 169	0x9030,
 170	0x9034,
 171	0x9038,
 172	0x903c,
 173	0x9040,
 174	0x9654,
 175	0x897c,
 176	0xa210,
 177	0xa214,
 178	0x9868,
 179	0xa02c,
 180	0x9664,
 181	0x9698,
 182	0x949c,
 183	0x8e10,
 184	0x8e18,
 185	0x8c50,
 186	0x8c58,
 187	0x8c60,
 188	0x8c68,
 189	0x89b4,
 190	0x9830,
 191	0x802c,
 192};
 193
 194/* Firmware Names */
 195MODULE_FIRMWARE("radeon/BARTS_pfp.bin");
 196MODULE_FIRMWARE("radeon/BARTS_me.bin");
 197MODULE_FIRMWARE("radeon/BARTS_mc.bin");
 198MODULE_FIRMWARE("radeon/BARTS_smc.bin");
 199MODULE_FIRMWARE("radeon/BTC_rlc.bin");
 200MODULE_FIRMWARE("radeon/TURKS_pfp.bin");
 201MODULE_FIRMWARE("radeon/TURKS_me.bin");
 202MODULE_FIRMWARE("radeon/TURKS_mc.bin");
 203MODULE_FIRMWARE("radeon/TURKS_smc.bin");
 204MODULE_FIRMWARE("radeon/CAICOS_pfp.bin");
 205MODULE_FIRMWARE("radeon/CAICOS_me.bin");
 206MODULE_FIRMWARE("radeon/CAICOS_mc.bin");
 207MODULE_FIRMWARE("radeon/CAICOS_smc.bin");
 208MODULE_FIRMWARE("radeon/CAYMAN_pfp.bin");
 209MODULE_FIRMWARE("radeon/CAYMAN_me.bin");
 210MODULE_FIRMWARE("radeon/CAYMAN_mc.bin");
 211MODULE_FIRMWARE("radeon/CAYMAN_rlc.bin");
 212MODULE_FIRMWARE("radeon/CAYMAN_smc.bin");
 213MODULE_FIRMWARE("radeon/ARUBA_pfp.bin");
 214MODULE_FIRMWARE("radeon/ARUBA_me.bin");
 215MODULE_FIRMWARE("radeon/ARUBA_rlc.bin");
 216
 217
 218static const u32 cayman_golden_registers2[] = {
 219	0x3e5c, 0xffffffff, 0x00000000,
 220	0x3e48, 0xffffffff, 0x00000000,
 221	0x3e4c, 0xffffffff, 0x00000000,
 222	0x3e64, 0xffffffff, 0x00000000,
 223	0x3e50, 0xffffffff, 0x00000000,
 224	0x3e60, 0xffffffff, 0x00000000
 225};
 226
 227static const u32 cayman_golden_registers[] = {
 228	0x5eb4, 0xffffffff, 0x00000002,
 229	0x5e78, 0x8f311ff1, 0x001000f0,
 230	0x3f90, 0xffff0000, 0xff000000,
 231	0x9148, 0xffff0000, 0xff000000,
 232	0x3f94, 0xffff0000, 0xff000000,
 233	0x914c, 0xffff0000, 0xff000000,
 234	0xc78, 0x00000080, 0x00000080,
 235	0xbd4, 0x70073777, 0x00011003,
 236	0xd02c, 0xbfffff1f, 0x08421000,
 237	0xd0b8, 0x73773777, 0x02011003,
 238	0x5bc0, 0x00200000, 0x50100000,
 239	0x98f8, 0x33773777, 0x02011003,
 240	0x98fc, 0xffffffff, 0x76541032,
 241	0x7030, 0x31000311, 0x00000011,
 242	0x2f48, 0x33773777, 0x42010001,
 243	0x6b28, 0x00000010, 0x00000012,
 244	0x7728, 0x00000010, 0x00000012,
 245	0x10328, 0x00000010, 0x00000012,
 246	0x10f28, 0x00000010, 0x00000012,
 247	0x11b28, 0x00000010, 0x00000012,
 248	0x12728, 0x00000010, 0x00000012,
 249	0x240c, 0x000007ff, 0x00000000,
 250	0x8a14, 0xf000001f, 0x00000007,
 251	0x8b24, 0x3fff3fff, 0x00ff0fff,
 252	0x8b10, 0x0000ff0f, 0x00000000,
 253	0x28a4c, 0x07ffffff, 0x06000000,
 254	0x10c, 0x00000001, 0x00010003,
 255	0xa02c, 0xffffffff, 0x0000009b,
 256	0x913c, 0x0000010f, 0x01000100,
 257	0x8c04, 0xf8ff00ff, 0x40600060,
 258	0x28350, 0x00000f01, 0x00000000,
 259	0x9508, 0x3700001f, 0x00000002,
 260	0x960c, 0xffffffff, 0x54763210,
 261	0x88c4, 0x001f3ae3, 0x00000082,
 262	0x88d0, 0xffffffff, 0x0f40df40,
 263	0x88d4, 0x0000001f, 0x00000010,
 264	0x8974, 0xffffffff, 0x00000000
 265};
 266
 267static const u32 dvst_golden_registers2[] = {
 268	0x8f8, 0xffffffff, 0,
 269	0x8fc, 0x00380000, 0,
 270	0x8f8, 0xffffffff, 1,
 271	0x8fc, 0x0e000000, 0
 272};
 273
 274static const u32 dvst_golden_registers[] = {
 275	0x690, 0x3fff3fff, 0x20c00033,
 276	0x918c, 0x0fff0fff, 0x00010006,
 277	0x91a8, 0x0fff0fff, 0x00010006,
 278	0x9150, 0xffffdfff, 0x6e944040,
 279	0x917c, 0x0fff0fff, 0x00030002,
 280	0x9198, 0x0fff0fff, 0x00030002,
 281	0x915c, 0x0fff0fff, 0x00010000,
 282	0x3f90, 0xffff0001, 0xff000000,
 283	0x9178, 0x0fff0fff, 0x00070000,
 284	0x9194, 0x0fff0fff, 0x00070000,
 285	0x9148, 0xffff0001, 0xff000000,
 286	0x9190, 0x0fff0fff, 0x00090008,
 287	0x91ac, 0x0fff0fff, 0x00090008,
 288	0x3f94, 0xffff0000, 0xff000000,
 289	0x914c, 0xffff0000, 0xff000000,
 290	0x929c, 0x00000fff, 0x00000001,
 291	0x55e4, 0xff607fff, 0xfc000100,
 292	0x8a18, 0xff000fff, 0x00000100,
 293	0x8b28, 0xff000fff, 0x00000100,
 294	0x9144, 0xfffc0fff, 0x00000100,
 295	0x6ed8, 0x00010101, 0x00010000,
 296	0x9830, 0xffffffff, 0x00000000,
 297	0x9834, 0xf00fffff, 0x00000400,
 298	0x9838, 0xfffffffe, 0x00000000,
 299	0xd0c0, 0xff000fff, 0x00000100,
 300	0xd02c, 0xbfffff1f, 0x08421000,
 301	0xd0b8, 0x73773777, 0x12010001,
 302	0x5bb0, 0x000000f0, 0x00000070,
 303	0x98f8, 0x73773777, 0x12010001,
 304	0x98fc, 0xffffffff, 0x00000010,
 305	0x9b7c, 0x00ff0000, 0x00fc0000,
 306	0x8030, 0x00001f0f, 0x0000100a,
 307	0x2f48, 0x73773777, 0x12010001,
 308	0x2408, 0x00030000, 0x000c007f,
 309	0x8a14, 0xf000003f, 0x00000007,
 310	0x8b24, 0x3fff3fff, 0x00ff0fff,
 311	0x8b10, 0x0000ff0f, 0x00000000,
 312	0x28a4c, 0x07ffffff, 0x06000000,
 313	0x4d8, 0x00000fff, 0x00000100,
 314	0xa008, 0xffffffff, 0x00010000,
 315	0x913c, 0xffff03ff, 0x01000100,
 316	0x8c00, 0x000000ff, 0x00000003,
 317	0x8c04, 0xf8ff00ff, 0x40600060,
 318	0x8cf0, 0x1fff1fff, 0x08e00410,
 319	0x28350, 0x00000f01, 0x00000000,
 320	0x9508, 0xf700071f, 0x00000002,
 321	0x960c, 0xffffffff, 0x54763210,
 322	0x20ef8, 0x01ff01ff, 0x00000002,
 323	0x20e98, 0xfffffbff, 0x00200000,
 324	0x2015c, 0xffffffff, 0x00000f40,
 325	0x88c4, 0x001f3ae3, 0x00000082,
 326	0x8978, 0x3fffffff, 0x04050140,
 327	0x88d4, 0x0000001f, 0x00000010,
 328	0x8974, 0xffffffff, 0x00000000
 329};
 330
 331static const u32 scrapper_golden_registers[] = {
 332	0x690, 0x3fff3fff, 0x20c00033,
 333	0x918c, 0x0fff0fff, 0x00010006,
 334	0x918c, 0x0fff0fff, 0x00010006,
 335	0x91a8, 0x0fff0fff, 0x00010006,
 336	0x91a8, 0x0fff0fff, 0x00010006,
 337	0x9150, 0xffffdfff, 0x6e944040,
 338	0x9150, 0xffffdfff, 0x6e944040,
 339	0x917c, 0x0fff0fff, 0x00030002,
 340	0x917c, 0x0fff0fff, 0x00030002,
 341	0x9198, 0x0fff0fff, 0x00030002,
 342	0x9198, 0x0fff0fff, 0x00030002,
 343	0x915c, 0x0fff0fff, 0x00010000,
 344	0x915c, 0x0fff0fff, 0x00010000,
 345	0x3f90, 0xffff0001, 0xff000000,
 346	0x3f90, 0xffff0001, 0xff000000,
 347	0x9178, 0x0fff0fff, 0x00070000,
 348	0x9178, 0x0fff0fff, 0x00070000,
 349	0x9194, 0x0fff0fff, 0x00070000,
 350	0x9194, 0x0fff0fff, 0x00070000,
 351	0x9148, 0xffff0001, 0xff000000,
 352	0x9148, 0xffff0001, 0xff000000,
 353	0x9190, 0x0fff0fff, 0x00090008,
 354	0x9190, 0x0fff0fff, 0x00090008,
 355	0x91ac, 0x0fff0fff, 0x00090008,
 356	0x91ac, 0x0fff0fff, 0x00090008,
 357	0x3f94, 0xffff0000, 0xff000000,
 358	0x3f94, 0xffff0000, 0xff000000,
 359	0x914c, 0xffff0000, 0xff000000,
 360	0x914c, 0xffff0000, 0xff000000,
 361	0x929c, 0x00000fff, 0x00000001,
 362	0x929c, 0x00000fff, 0x00000001,
 363	0x55e4, 0xff607fff, 0xfc000100,
 364	0x8a18, 0xff000fff, 0x00000100,
 365	0x8a18, 0xff000fff, 0x00000100,
 366	0x8b28, 0xff000fff, 0x00000100,
 367	0x8b28, 0xff000fff, 0x00000100,
 368	0x9144, 0xfffc0fff, 0x00000100,
 369	0x9144, 0xfffc0fff, 0x00000100,
 370	0x6ed8, 0x00010101, 0x00010000,
 371	0x9830, 0xffffffff, 0x00000000,
 372	0x9830, 0xffffffff, 0x00000000,
 373	0x9834, 0xf00fffff, 0x00000400,
 374	0x9834, 0xf00fffff, 0x00000400,
 375	0x9838, 0xfffffffe, 0x00000000,
 376	0x9838, 0xfffffffe, 0x00000000,
 377	0xd0c0, 0xff000fff, 0x00000100,
 378	0xd02c, 0xbfffff1f, 0x08421000,
 379	0xd02c, 0xbfffff1f, 0x08421000,
 380	0xd0b8, 0x73773777, 0x12010001,
 381	0xd0b8, 0x73773777, 0x12010001,
 382	0x5bb0, 0x000000f0, 0x00000070,
 383	0x98f8, 0x73773777, 0x12010001,
 384	0x98f8, 0x73773777, 0x12010001,
 385	0x98fc, 0xffffffff, 0x00000010,
 386	0x98fc, 0xffffffff, 0x00000010,
 387	0x9b7c, 0x00ff0000, 0x00fc0000,
 388	0x9b7c, 0x00ff0000, 0x00fc0000,
 389	0x8030, 0x00001f0f, 0x0000100a,
 390	0x8030, 0x00001f0f, 0x0000100a,
 391	0x2f48, 0x73773777, 0x12010001,
 392	0x2f48, 0x73773777, 0x12010001,
 393	0x2408, 0x00030000, 0x000c007f,
 394	0x8a14, 0xf000003f, 0x00000007,
 395	0x8a14, 0xf000003f, 0x00000007,
 396	0x8b24, 0x3fff3fff, 0x00ff0fff,
 397	0x8b24, 0x3fff3fff, 0x00ff0fff,
 398	0x8b10, 0x0000ff0f, 0x00000000,
 399	0x8b10, 0x0000ff0f, 0x00000000,
 400	0x28a4c, 0x07ffffff, 0x06000000,
 401	0x28a4c, 0x07ffffff, 0x06000000,
 402	0x4d8, 0x00000fff, 0x00000100,
 403	0x4d8, 0x00000fff, 0x00000100,
 404	0xa008, 0xffffffff, 0x00010000,
 405	0xa008, 0xffffffff, 0x00010000,
 406	0x913c, 0xffff03ff, 0x01000100,
 407	0x913c, 0xffff03ff, 0x01000100,
 408	0x90e8, 0x001fffff, 0x010400c0,
 409	0x8c00, 0x000000ff, 0x00000003,
 410	0x8c00, 0x000000ff, 0x00000003,
 411	0x8c04, 0xf8ff00ff, 0x40600060,
 412	0x8c04, 0xf8ff00ff, 0x40600060,
 413	0x8c30, 0x0000000f, 0x00040005,
 414	0x8cf0, 0x1fff1fff, 0x08e00410,
 415	0x8cf0, 0x1fff1fff, 0x08e00410,
 416	0x900c, 0x00ffffff, 0x0017071f,
 417	0x28350, 0x00000f01, 0x00000000,
 418	0x28350, 0x00000f01, 0x00000000,
 419	0x9508, 0xf700071f, 0x00000002,
 420	0x9508, 0xf700071f, 0x00000002,
 421	0x9688, 0x00300000, 0x0017000f,
 422	0x960c, 0xffffffff, 0x54763210,
 423	0x960c, 0xffffffff, 0x54763210,
 424	0x20ef8, 0x01ff01ff, 0x00000002,
 425	0x20e98, 0xfffffbff, 0x00200000,
 426	0x2015c, 0xffffffff, 0x00000f40,
 427	0x88c4, 0x001f3ae3, 0x00000082,
 428	0x88c4, 0x001f3ae3, 0x00000082,
 429	0x8978, 0x3fffffff, 0x04050140,
 430	0x8978, 0x3fffffff, 0x04050140,
 431	0x88d4, 0x0000001f, 0x00000010,
 432	0x88d4, 0x0000001f, 0x00000010,
 433	0x8974, 0xffffffff, 0x00000000,
 434	0x8974, 0xffffffff, 0x00000000
 435};
 436
 437static void ni_init_golden_registers(struct radeon_device *rdev)
 438{
 439	switch (rdev->family) {
 440	case CHIP_CAYMAN:
 441		radeon_program_register_sequence(rdev,
 442						 cayman_golden_registers,
 443						 (const u32)ARRAY_SIZE(cayman_golden_registers));
 444		radeon_program_register_sequence(rdev,
 445						 cayman_golden_registers2,
 446						 (const u32)ARRAY_SIZE(cayman_golden_registers2));
 447		break;
 448	case CHIP_ARUBA:
 449		if ((rdev->pdev->device == 0x9900) ||
 450		    (rdev->pdev->device == 0x9901) ||
 451		    (rdev->pdev->device == 0x9903) ||
 452		    (rdev->pdev->device == 0x9904) ||
 453		    (rdev->pdev->device == 0x9905) ||
 454		    (rdev->pdev->device == 0x9906) ||
 455		    (rdev->pdev->device == 0x9907) ||
 456		    (rdev->pdev->device == 0x9908) ||
 457		    (rdev->pdev->device == 0x9909) ||
 458		    (rdev->pdev->device == 0x990A) ||
 459		    (rdev->pdev->device == 0x990B) ||
 460		    (rdev->pdev->device == 0x990C) ||
 461		    (rdev->pdev->device == 0x990D) ||
 462		    (rdev->pdev->device == 0x990E) ||
 463		    (rdev->pdev->device == 0x990F) ||
 464		    (rdev->pdev->device == 0x9910) ||
 465		    (rdev->pdev->device == 0x9913) ||
 466		    (rdev->pdev->device == 0x9917) ||
 467		    (rdev->pdev->device == 0x9918)) {
 468			radeon_program_register_sequence(rdev,
 469							 dvst_golden_registers,
 470							 (const u32)ARRAY_SIZE(dvst_golden_registers));
 471			radeon_program_register_sequence(rdev,
 472							 dvst_golden_registers2,
 473							 (const u32)ARRAY_SIZE(dvst_golden_registers2));
 474		} else {
 475			radeon_program_register_sequence(rdev,
 476							 scrapper_golden_registers,
 477							 (const u32)ARRAY_SIZE(scrapper_golden_registers));
 478			radeon_program_register_sequence(rdev,
 479							 dvst_golden_registers2,
 480							 (const u32)ARRAY_SIZE(dvst_golden_registers2));
 481		}
 482		break;
 483	default:
 484		break;
 485	}
 486}
 487
 488#define BTC_IO_MC_REGS_SIZE 29
 489
 490static const u32 barts_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 491	{0x00000077, 0xff010100},
 492	{0x00000078, 0x00000000},
 493	{0x00000079, 0x00001434},
 494	{0x0000007a, 0xcc08ec08},
 495	{0x0000007b, 0x00040000},
 496	{0x0000007c, 0x000080c0},
 497	{0x0000007d, 0x09000000},
 498	{0x0000007e, 0x00210404},
 499	{0x00000081, 0x08a8e800},
 500	{0x00000082, 0x00030444},
 501	{0x00000083, 0x00000000},
 502	{0x00000085, 0x00000001},
 503	{0x00000086, 0x00000002},
 504	{0x00000087, 0x48490000},
 505	{0x00000088, 0x20244647},
 506	{0x00000089, 0x00000005},
 507	{0x0000008b, 0x66030000},
 508	{0x0000008c, 0x00006603},
 509	{0x0000008d, 0x00000100},
 510	{0x0000008f, 0x00001c0a},
 511	{0x00000090, 0xff000001},
 512	{0x00000094, 0x00101101},
 513	{0x00000095, 0x00000fff},
 514	{0x00000096, 0x00116fff},
 515	{0x00000097, 0x60010000},
 516	{0x00000098, 0x10010000},
 517	{0x00000099, 0x00006000},
 518	{0x0000009a, 0x00001000},
 519	{0x0000009f, 0x00946a00}
 520};
 521
 522static const u32 turks_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 523	{0x00000077, 0xff010100},
 524	{0x00000078, 0x00000000},
 525	{0x00000079, 0x00001434},
 526	{0x0000007a, 0xcc08ec08},
 527	{0x0000007b, 0x00040000},
 528	{0x0000007c, 0x000080c0},
 529	{0x0000007d, 0x09000000},
 530	{0x0000007e, 0x00210404},
 531	{0x00000081, 0x08a8e800},
 532	{0x00000082, 0x00030444},
 533	{0x00000083, 0x00000000},
 534	{0x00000085, 0x00000001},
 535	{0x00000086, 0x00000002},
 536	{0x00000087, 0x48490000},
 537	{0x00000088, 0x20244647},
 538	{0x00000089, 0x00000005},
 539	{0x0000008b, 0x66030000},
 540	{0x0000008c, 0x00006603},
 541	{0x0000008d, 0x00000100},
 542	{0x0000008f, 0x00001c0a},
 543	{0x00000090, 0xff000001},
 544	{0x00000094, 0x00101101},
 545	{0x00000095, 0x00000fff},
 546	{0x00000096, 0x00116fff},
 547	{0x00000097, 0x60010000},
 548	{0x00000098, 0x10010000},
 549	{0x00000099, 0x00006000},
 550	{0x0000009a, 0x00001000},
 551	{0x0000009f, 0x00936a00}
 552};
 553
 554static const u32 caicos_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 555	{0x00000077, 0xff010100},
 556	{0x00000078, 0x00000000},
 557	{0x00000079, 0x00001434},
 558	{0x0000007a, 0xcc08ec08},
 559	{0x0000007b, 0x00040000},
 560	{0x0000007c, 0x000080c0},
 561	{0x0000007d, 0x09000000},
 562	{0x0000007e, 0x00210404},
 563	{0x00000081, 0x08a8e800},
 564	{0x00000082, 0x00030444},
 565	{0x00000083, 0x00000000},
 566	{0x00000085, 0x00000001},
 567	{0x00000086, 0x00000002},
 568	{0x00000087, 0x48490000},
 569	{0x00000088, 0x20244647},
 570	{0x00000089, 0x00000005},
 571	{0x0000008b, 0x66030000},
 572	{0x0000008c, 0x00006603},
 573	{0x0000008d, 0x00000100},
 574	{0x0000008f, 0x00001c0a},
 575	{0x00000090, 0xff000001},
 576	{0x00000094, 0x00101101},
 577	{0x00000095, 0x00000fff},
 578	{0x00000096, 0x00116fff},
 579	{0x00000097, 0x60010000},
 580	{0x00000098, 0x10010000},
 581	{0x00000099, 0x00006000},
 582	{0x0000009a, 0x00001000},
 583	{0x0000009f, 0x00916a00}
 584};
 585
 586static const u32 cayman_io_mc_regs[BTC_IO_MC_REGS_SIZE][2] = {
 587	{0x00000077, 0xff010100},
 588	{0x00000078, 0x00000000},
 589	{0x00000079, 0x00001434},
 590	{0x0000007a, 0xcc08ec08},
 591	{0x0000007b, 0x00040000},
 592	{0x0000007c, 0x000080c0},
 593	{0x0000007d, 0x09000000},
 594	{0x0000007e, 0x00210404},
 595	{0x00000081, 0x08a8e800},
 596	{0x00000082, 0x00030444},
 597	{0x00000083, 0x00000000},
 598	{0x00000085, 0x00000001},
 599	{0x00000086, 0x00000002},
 600	{0x00000087, 0x48490000},
 601	{0x00000088, 0x20244647},
 602	{0x00000089, 0x00000005},
 603	{0x0000008b, 0x66030000},
 604	{0x0000008c, 0x00006603},
 605	{0x0000008d, 0x00000100},
 606	{0x0000008f, 0x00001c0a},
 607	{0x00000090, 0xff000001},
 608	{0x00000094, 0x00101101},
 609	{0x00000095, 0x00000fff},
 610	{0x00000096, 0x00116fff},
 611	{0x00000097, 0x60010000},
 612	{0x00000098, 0x10010000},
 613	{0x00000099, 0x00006000},
 614	{0x0000009a, 0x00001000},
 615	{0x0000009f, 0x00976b00}
 616};
 617
 618int ni_mc_load_microcode(struct radeon_device *rdev)
 619{
 620	const __be32 *fw_data;
 621	u32 mem_type, running;
 622	u32 *io_mc_regs;
 623	int i, ucode_size, regs_size;
 624
 625	if (!rdev->mc_fw)
 626		return -EINVAL;
 627
 628	switch (rdev->family) {
 629	case CHIP_BARTS:
 630		io_mc_regs = (u32 *)&barts_io_mc_regs;
 631		ucode_size = BTC_MC_UCODE_SIZE;
 632		regs_size = BTC_IO_MC_REGS_SIZE;
 633		break;
 634	case CHIP_TURKS:
 635		io_mc_regs = (u32 *)&turks_io_mc_regs;
 636		ucode_size = BTC_MC_UCODE_SIZE;
 637		regs_size = BTC_IO_MC_REGS_SIZE;
 638		break;
 639	case CHIP_CAICOS:
 640	default:
 641		io_mc_regs = (u32 *)&caicos_io_mc_regs;
 642		ucode_size = BTC_MC_UCODE_SIZE;
 643		regs_size = BTC_IO_MC_REGS_SIZE;
 644		break;
 645	case CHIP_CAYMAN:
 646		io_mc_regs = (u32 *)&cayman_io_mc_regs;
 647		ucode_size = CAYMAN_MC_UCODE_SIZE;
 648		regs_size = BTC_IO_MC_REGS_SIZE;
 649		break;
 650	}
 651
 652	mem_type = (RREG32(MC_SEQ_MISC0) & MC_SEQ_MISC0_GDDR5_MASK) >> MC_SEQ_MISC0_GDDR5_SHIFT;
 653	running = RREG32(MC_SEQ_SUP_CNTL) & RUN_MASK;
 654
 655	if ((mem_type == MC_SEQ_MISC0_GDDR5_VALUE) && (running == 0)) {
 
 
 
 
 
 656		/* reset the engine and set to writable */
 657		WREG32(MC_SEQ_SUP_CNTL, 0x00000008);
 658		WREG32(MC_SEQ_SUP_CNTL, 0x00000010);
 659
 660		/* load mc io regs */
 661		for (i = 0; i < regs_size; i++) {
 662			WREG32(MC_SEQ_IO_DEBUG_INDEX, io_mc_regs[(i << 1)]);
 663			WREG32(MC_SEQ_IO_DEBUG_DATA, io_mc_regs[(i << 1) + 1]);
 664		}
 665		/* load the MC ucode */
 666		fw_data = (const __be32 *)rdev->mc_fw->data;
 667		for (i = 0; i < ucode_size; i++)
 668			WREG32(MC_SEQ_SUP_PGM, be32_to_cpup(fw_data++));
 669
 670		/* put the engine back into the active state */
 671		WREG32(MC_SEQ_SUP_CNTL, 0x00000008);
 672		WREG32(MC_SEQ_SUP_CNTL, 0x00000004);
 673		WREG32(MC_SEQ_SUP_CNTL, 0x00000001);
 674
 675		/* wait for training to complete */
 676		for (i = 0; i < rdev->usec_timeout; i++) {
 677			if (RREG32(MC_IO_PAD_CNTL_D0) & MEM_FALL_OUT_CMD)
 678				break;
 679			udelay(1);
 680		}
 681	}
 682
 683	return 0;
 684}
 685
 686int ni_init_microcode(struct radeon_device *rdev)
 687{
 
 688	const char *chip_name;
 689	const char *rlc_chip_name;
 690	size_t pfp_req_size, me_req_size, rlc_req_size, mc_req_size;
 691	size_t smc_req_size = 0;
 692	char fw_name[30];
 693	int err;
 694
 695	DRM_DEBUG("\n");
 696
 
 
 
 
 
 
 
 697	switch (rdev->family) {
 698	case CHIP_BARTS:
 699		chip_name = "BARTS";
 700		rlc_chip_name = "BTC";
 701		pfp_req_size = EVERGREEN_PFP_UCODE_SIZE * 4;
 702		me_req_size = EVERGREEN_PM4_UCODE_SIZE * 4;
 703		rlc_req_size = EVERGREEN_RLC_UCODE_SIZE * 4;
 704		mc_req_size = BTC_MC_UCODE_SIZE * 4;
 705		smc_req_size = ALIGN(BARTS_SMC_UCODE_SIZE, 4);
 706		break;
 707	case CHIP_TURKS:
 708		chip_name = "TURKS";
 709		rlc_chip_name = "BTC";
 710		pfp_req_size = EVERGREEN_PFP_UCODE_SIZE * 4;
 711		me_req_size = EVERGREEN_PM4_UCODE_SIZE * 4;
 712		rlc_req_size = EVERGREEN_RLC_UCODE_SIZE * 4;
 713		mc_req_size = BTC_MC_UCODE_SIZE * 4;
 714		smc_req_size = ALIGN(TURKS_SMC_UCODE_SIZE, 4);
 715		break;
 716	case CHIP_CAICOS:
 717		chip_name = "CAICOS";
 718		rlc_chip_name = "BTC";
 719		pfp_req_size = EVERGREEN_PFP_UCODE_SIZE * 4;
 720		me_req_size = EVERGREEN_PM4_UCODE_SIZE * 4;
 721		rlc_req_size = EVERGREEN_RLC_UCODE_SIZE * 4;
 722		mc_req_size = BTC_MC_UCODE_SIZE * 4;
 723		smc_req_size = ALIGN(CAICOS_SMC_UCODE_SIZE, 4);
 724		break;
 725	case CHIP_CAYMAN:
 726		chip_name = "CAYMAN";
 727		rlc_chip_name = "CAYMAN";
 728		pfp_req_size = CAYMAN_PFP_UCODE_SIZE * 4;
 729		me_req_size = CAYMAN_PM4_UCODE_SIZE * 4;
 730		rlc_req_size = CAYMAN_RLC_UCODE_SIZE * 4;
 731		mc_req_size = CAYMAN_MC_UCODE_SIZE * 4;
 732		smc_req_size = ALIGN(CAYMAN_SMC_UCODE_SIZE, 4);
 733		break;
 734	case CHIP_ARUBA:
 735		chip_name = "ARUBA";
 736		rlc_chip_name = "ARUBA";
 737		/* pfp/me same size as CAYMAN */
 738		pfp_req_size = CAYMAN_PFP_UCODE_SIZE * 4;
 739		me_req_size = CAYMAN_PM4_UCODE_SIZE * 4;
 740		rlc_req_size = ARUBA_RLC_UCODE_SIZE * 4;
 741		mc_req_size = 0;
 742		break;
 743	default:
 744		BUG();
 745	}
 746
 747	DRM_INFO("Loading %s Microcode\n", chip_name);
 748
 749	snprintf(fw_name, sizeof(fw_name), "radeon/%s_pfp.bin", chip_name);
 750	err = request_firmware(&rdev->pfp_fw, fw_name, rdev->dev);
 751	if (err)
 752		goto out;
 753	if (rdev->pfp_fw->size != pfp_req_size) {
 754		pr_err("ni_cp: Bogus length %zu in firmware \"%s\"\n",
 
 755		       rdev->pfp_fw->size, fw_name);
 756		err = -EINVAL;
 757		goto out;
 758	}
 759
 760	snprintf(fw_name, sizeof(fw_name), "radeon/%s_me.bin", chip_name);
 761	err = request_firmware(&rdev->me_fw, fw_name, rdev->dev);
 762	if (err)
 763		goto out;
 764	if (rdev->me_fw->size != me_req_size) {
 765		pr_err("ni_cp: Bogus length %zu in firmware \"%s\"\n",
 
 766		       rdev->me_fw->size, fw_name);
 767		err = -EINVAL;
 768	}
 769
 770	snprintf(fw_name, sizeof(fw_name), "radeon/%s_rlc.bin", rlc_chip_name);
 771	err = request_firmware(&rdev->rlc_fw, fw_name, rdev->dev);
 772	if (err)
 773		goto out;
 774	if (rdev->rlc_fw->size != rlc_req_size) {
 775		pr_err("ni_rlc: Bogus length %zu in firmware \"%s\"\n",
 
 776		       rdev->rlc_fw->size, fw_name);
 777		err = -EINVAL;
 778	}
 779
 780	/* no MC ucode on TN */
 781	if (!(rdev->flags & RADEON_IS_IGP)) {
 782		snprintf(fw_name, sizeof(fw_name), "radeon/%s_mc.bin", chip_name);
 783		err = request_firmware(&rdev->mc_fw, fw_name, rdev->dev);
 784		if (err)
 785			goto out;
 786		if (rdev->mc_fw->size != mc_req_size) {
 787			pr_err("ni_mc: Bogus length %zu in firmware \"%s\"\n",
 788			       rdev->mc_fw->size, fw_name);
 789			err = -EINVAL;
 790		}
 791	}
 792
 793	if ((rdev->family >= CHIP_BARTS) && (rdev->family <= CHIP_CAYMAN)) {
 794		snprintf(fw_name, sizeof(fw_name), "radeon/%s_smc.bin", chip_name);
 795		err = request_firmware(&rdev->smc_fw, fw_name, rdev->dev);
 796		if (err) {
 797			pr_err("smc: error loading firmware \"%s\"\n", fw_name);
 798			release_firmware(rdev->smc_fw);
 799			rdev->smc_fw = NULL;
 800			err = 0;
 801		} else if (rdev->smc_fw->size != smc_req_size) {
 802			pr_err("ni_mc: Bogus length %zu in firmware \"%s\"\n",
 803			       rdev->smc_fw->size, fw_name);
 804			err = -EINVAL;
 805		}
 806	}
 
 
 807
 808out:
 809	if (err) {
 810		if (err != -EINVAL)
 811			pr_err("ni_cp: Failed to load firmware \"%s\"\n",
 
 812			       fw_name);
 813		release_firmware(rdev->pfp_fw);
 814		rdev->pfp_fw = NULL;
 815		release_firmware(rdev->me_fw);
 816		rdev->me_fw = NULL;
 817		release_firmware(rdev->rlc_fw);
 818		rdev->rlc_fw = NULL;
 819		release_firmware(rdev->mc_fw);
 820		rdev->mc_fw = NULL;
 821	}
 822	return err;
 823}
 824
 825/**
 826 * cayman_get_allowed_info_register - fetch the register for the info ioctl
 827 *
 828 * @rdev: radeon_device pointer
 829 * @reg: register offset in bytes
 830 * @val: register value
 831 *
 832 * Returns 0 for success or -EINVAL for an invalid register
 833 *
 834 */
 835int cayman_get_allowed_info_register(struct radeon_device *rdev,
 836				     u32 reg, u32 *val)
 837{
 838	switch (reg) {
 839	case GRBM_STATUS:
 840	case GRBM_STATUS_SE0:
 841	case GRBM_STATUS_SE1:
 842	case SRBM_STATUS:
 843	case SRBM_STATUS2:
 844	case (DMA_STATUS_REG + DMA0_REGISTER_OFFSET):
 845	case (DMA_STATUS_REG + DMA1_REGISTER_OFFSET):
 846	case UVD_STATUS:
 847		*val = RREG32(reg);
 848		return 0;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 849	default:
 850		return -EINVAL;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 851	}
 
 
 852}
 853
 854int tn_get_temp(struct radeon_device *rdev)
 
 
 
 855{
 856	u32 temp = RREG32_SMC(TN_CURRENT_GNB_TEMP) & 0x7ff;
 857	int actual_temp = (temp / 8) - 49;
 858
 859	return actual_temp * 1000;
 
 
 
 
 
 860}
 861
 862/*
 863 * Core functions
 864 */
 865static void cayman_gpu_init(struct radeon_device *rdev)
 866{
 
 
 867	u32 gb_addr_config = 0;
 868	u32 mc_arb_ramcfg;
 
 869	u32 cgts_tcc_disable;
 870	u32 sx_debug_1;
 871	u32 smx_dc_ctl0;
 
 
 
 872	u32 cgts_sm_ctrl_reg;
 873	u32 hdp_host_path_cntl;
 874	u32 tmp;
 875	u32 disabled_rb_mask;
 876	int i, j;
 877
 878	switch (rdev->family) {
 879	case CHIP_CAYMAN:
 
 880		rdev->config.cayman.max_shader_engines = 2;
 881		rdev->config.cayman.max_pipes_per_simd = 4;
 882		rdev->config.cayman.max_tile_pipes = 8;
 883		rdev->config.cayman.max_simds_per_se = 12;
 884		rdev->config.cayman.max_backends_per_se = 4;
 885		rdev->config.cayman.max_texture_channel_caches = 8;
 886		rdev->config.cayman.max_gprs = 256;
 887		rdev->config.cayman.max_threads = 256;
 888		rdev->config.cayman.max_gs_threads = 32;
 889		rdev->config.cayman.max_stack_entries = 512;
 890		rdev->config.cayman.sx_num_of_sets = 8;
 891		rdev->config.cayman.sx_max_export_size = 256;
 892		rdev->config.cayman.sx_max_export_pos_size = 64;
 893		rdev->config.cayman.sx_max_export_smx_size = 192;
 894		rdev->config.cayman.max_hw_contexts = 8;
 895		rdev->config.cayman.sq_num_cf_insts = 2;
 896
 897		rdev->config.cayman.sc_prim_fifo_size = 0x100;
 898		rdev->config.cayman.sc_hiz_tile_fifo_size = 0x30;
 899		rdev->config.cayman.sc_earlyz_tile_fifo_size = 0x130;
 900		gb_addr_config = CAYMAN_GB_ADDR_CONFIG_GOLDEN;
 901		break;
 902	case CHIP_ARUBA:
 903	default:
 904		rdev->config.cayman.max_shader_engines = 1;
 905		rdev->config.cayman.max_pipes_per_simd = 4;
 906		rdev->config.cayman.max_tile_pipes = 2;
 907		if ((rdev->pdev->device == 0x9900) ||
 908		    (rdev->pdev->device == 0x9901) ||
 909		    (rdev->pdev->device == 0x9905) ||
 910		    (rdev->pdev->device == 0x9906) ||
 911		    (rdev->pdev->device == 0x9907) ||
 912		    (rdev->pdev->device == 0x9908) ||
 913		    (rdev->pdev->device == 0x9909) ||
 914		    (rdev->pdev->device == 0x990B) ||
 915		    (rdev->pdev->device == 0x990C) ||
 916		    (rdev->pdev->device == 0x990F) ||
 917		    (rdev->pdev->device == 0x9910) ||
 918		    (rdev->pdev->device == 0x9917) ||
 919		    (rdev->pdev->device == 0x9999) ||
 920		    (rdev->pdev->device == 0x999C)) {
 921			rdev->config.cayman.max_simds_per_se = 6;
 922			rdev->config.cayman.max_backends_per_se = 2;
 923			rdev->config.cayman.max_hw_contexts = 8;
 924			rdev->config.cayman.sx_max_export_size = 256;
 925			rdev->config.cayman.sx_max_export_pos_size = 64;
 926			rdev->config.cayman.sx_max_export_smx_size = 192;
 927		} else if ((rdev->pdev->device == 0x9903) ||
 928			   (rdev->pdev->device == 0x9904) ||
 929			   (rdev->pdev->device == 0x990A) ||
 930			   (rdev->pdev->device == 0x990D) ||
 931			   (rdev->pdev->device == 0x990E) ||
 932			   (rdev->pdev->device == 0x9913) ||
 933			   (rdev->pdev->device == 0x9918) ||
 934			   (rdev->pdev->device == 0x999D)) {
 935			rdev->config.cayman.max_simds_per_se = 4;
 936			rdev->config.cayman.max_backends_per_se = 2;
 937			rdev->config.cayman.max_hw_contexts = 8;
 938			rdev->config.cayman.sx_max_export_size = 256;
 939			rdev->config.cayman.sx_max_export_pos_size = 64;
 940			rdev->config.cayman.sx_max_export_smx_size = 192;
 941		} else if ((rdev->pdev->device == 0x9919) ||
 942			   (rdev->pdev->device == 0x9990) ||
 943			   (rdev->pdev->device == 0x9991) ||
 944			   (rdev->pdev->device == 0x9994) ||
 945			   (rdev->pdev->device == 0x9995) ||
 946			   (rdev->pdev->device == 0x9996) ||
 947			   (rdev->pdev->device == 0x999A) ||
 948			   (rdev->pdev->device == 0x99A0)) {
 949			rdev->config.cayman.max_simds_per_se = 3;
 950			rdev->config.cayman.max_backends_per_se = 1;
 951			rdev->config.cayman.max_hw_contexts = 4;
 952			rdev->config.cayman.sx_max_export_size = 128;
 953			rdev->config.cayman.sx_max_export_pos_size = 32;
 954			rdev->config.cayman.sx_max_export_smx_size = 96;
 955		} else {
 956			rdev->config.cayman.max_simds_per_se = 2;
 957			rdev->config.cayman.max_backends_per_se = 1;
 958			rdev->config.cayman.max_hw_contexts = 4;
 959			rdev->config.cayman.sx_max_export_size = 128;
 960			rdev->config.cayman.sx_max_export_pos_size = 32;
 961			rdev->config.cayman.sx_max_export_smx_size = 96;
 962		}
 963		rdev->config.cayman.max_texture_channel_caches = 2;
 964		rdev->config.cayman.max_gprs = 256;
 965		rdev->config.cayman.max_threads = 256;
 966		rdev->config.cayman.max_gs_threads = 32;
 967		rdev->config.cayman.max_stack_entries = 512;
 968		rdev->config.cayman.sx_num_of_sets = 8;
 969		rdev->config.cayman.sq_num_cf_insts = 2;
 970
 971		rdev->config.cayman.sc_prim_fifo_size = 0x40;
 972		rdev->config.cayman.sc_hiz_tile_fifo_size = 0x30;
 973		rdev->config.cayman.sc_earlyz_tile_fifo_size = 0x130;
 974		gb_addr_config = ARUBA_GB_ADDR_CONFIG_GOLDEN;
 975		break;
 976	}
 977
 978	/* Initialize HDP */
 979	for (i = 0, j = 0; i < 32; i++, j += 0x18) {
 980		WREG32((0x2c14 + j), 0x00000000);
 981		WREG32((0x2c18 + j), 0x00000000);
 982		WREG32((0x2c1c + j), 0x00000000);
 983		WREG32((0x2c20 + j), 0x00000000);
 984		WREG32((0x2c24 + j), 0x00000000);
 985	}
 986
 987	WREG32(GRBM_CNTL, GRBM_READ_TIMEOUT(0xff));
 988	WREG32(SRBM_INT_CNTL, 0x1);
 989	WREG32(SRBM_INT_ACK, 0x1);
 990
 991	evergreen_fix_pci_max_read_req_size(rdev);
 992
 993	RREG32(MC_SHARED_CHMAP);
 994	mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG);
 995
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 996	tmp = (mc_arb_ramcfg & NOOFCOLS_MASK) >> NOOFCOLS_SHIFT;
 997	rdev->config.cayman.mem_row_size_in_kb = (4 * (1 << (8 + tmp))) / 1024;
 998	if (rdev->config.cayman.mem_row_size_in_kb > 4)
 999		rdev->config.cayman.mem_row_size_in_kb = 4;
1000	/* XXX use MC settings? */
1001	rdev->config.cayman.shader_engine_tile_size = 32;
1002	rdev->config.cayman.num_gpus = 1;
1003	rdev->config.cayman.multi_gpu_tile_size = 64;
1004
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1005	tmp = (gb_addr_config & NUM_PIPES_MASK) >> NUM_PIPES_SHIFT;
1006	rdev->config.cayman.num_tile_pipes = (1 << tmp);
1007	tmp = (gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT;
1008	rdev->config.cayman.mem_max_burst_length_bytes = (tmp + 1) * 256;
1009	tmp = (gb_addr_config & NUM_SHADER_ENGINES_MASK) >> NUM_SHADER_ENGINES_SHIFT;
1010	rdev->config.cayman.num_shader_engines = tmp + 1;
1011	tmp = (gb_addr_config & NUM_GPUS_MASK) >> NUM_GPUS_SHIFT;
1012	rdev->config.cayman.num_gpus = tmp + 1;
1013	tmp = (gb_addr_config & MULTI_GPU_TILE_SIZE_MASK) >> MULTI_GPU_TILE_SIZE_SHIFT;
1014	rdev->config.cayman.multi_gpu_tile_size = 1 << tmp;
1015	tmp = (gb_addr_config & ROW_SIZE_MASK) >> ROW_SIZE_SHIFT;
1016	rdev->config.cayman.mem_row_size_in_kb = 1 << tmp;
1017
1018
 
 
 
 
 
 
 
 
 
 
1019	/* setup tiling info dword.  gb_addr_config is not adequate since it does
1020	 * not have bank info, so create a custom tiling dword.
1021	 * bits 3:0   num_pipes
1022	 * bits 7:4   num_banks
1023	 * bits 11:8  group_size
1024	 * bits 15:12 row_size
1025	 */
1026	rdev->config.cayman.tile_config = 0;
1027	switch (rdev->config.cayman.num_tile_pipes) {
1028	case 1:
1029	default:
1030		rdev->config.cayman.tile_config |= (0 << 0);
1031		break;
1032	case 2:
1033		rdev->config.cayman.tile_config |= (1 << 0);
1034		break;
1035	case 4:
1036		rdev->config.cayman.tile_config |= (2 << 0);
1037		break;
1038	case 8:
1039		rdev->config.cayman.tile_config |= (3 << 0);
1040		break;
1041	}
1042
1043	/* num banks is 8 on all fusion asics. 0 = 4, 1 = 8, 2 = 16 */
1044	if (rdev->flags & RADEON_IS_IGP)
1045		rdev->config.cayman.tile_config |= 1 << 4;
1046	else {
1047		switch ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) {
1048		case 0: /* four banks */
1049			rdev->config.cayman.tile_config |= 0 << 4;
1050			break;
1051		case 1: /* eight banks */
1052			rdev->config.cayman.tile_config |= 1 << 4;
1053			break;
1054		case 2: /* sixteen banks */
1055		default:
1056			rdev->config.cayman.tile_config |= 2 << 4;
1057			break;
1058		}
1059	}
1060	rdev->config.cayman.tile_config |=
1061		((gb_addr_config & PIPE_INTERLEAVE_SIZE_MASK) >> PIPE_INTERLEAVE_SIZE_SHIFT) << 8;
1062	rdev->config.cayman.tile_config |=
1063		((gb_addr_config & ROW_SIZE_MASK) >> ROW_SIZE_SHIFT) << 12;
1064
1065	tmp = 0;
1066	for (i = (rdev->config.cayman.max_shader_engines - 1); i >= 0; i--) {
1067		u32 rb_disable_bitmap;
1068
1069		WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
1070		WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
1071		rb_disable_bitmap = (RREG32(CC_RB_BACKEND_DISABLE) & 0x00ff0000) >> 16;
1072		tmp <<= 4;
1073		tmp |= rb_disable_bitmap;
1074	}
1075	/* enabled rb are just the one not disabled :) */
1076	disabled_rb_mask = tmp;
1077	tmp = 0;
1078	for (i = 0; i < (rdev->config.cayman.max_backends_per_se * rdev->config.cayman.max_shader_engines); i++)
1079		tmp |= (1 << i);
1080	/* if all the backends are disabled, fix it up here */
1081	if ((disabled_rb_mask & tmp) == tmp) {
1082		for (i = 0; i < (rdev->config.cayman.max_backends_per_se * rdev->config.cayman.max_shader_engines); i++)
1083			disabled_rb_mask &= ~(1 << i);
1084	}
1085
1086	for (i = 0; i < rdev->config.cayman.max_shader_engines; i++) {
1087		u32 simd_disable_bitmap;
1088
1089		WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
1090		WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
1091		simd_disable_bitmap = (RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffff0000) >> 16;
1092		simd_disable_bitmap |= 0xffffffff << rdev->config.cayman.max_simds_per_se;
1093		tmp <<= 16;
1094		tmp |= simd_disable_bitmap;
1095	}
1096	rdev->config.cayman.active_simds = hweight32(~tmp);
1097
1098	WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_BROADCAST_WRITES);
1099	WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_BROADCAST_WRITES);
1100
1101	WREG32(GB_ADDR_CONFIG, gb_addr_config);
1102	WREG32(DMIF_ADDR_CONFIG, gb_addr_config);
1103	if (ASIC_IS_DCE6(rdev))
1104		WREG32(DMIF_ADDR_CALC, gb_addr_config);
1105	WREG32(HDP_ADDR_CONFIG, gb_addr_config);
1106	WREG32(DMA_TILING_CONFIG + DMA0_REGISTER_OFFSET, gb_addr_config);
1107	WREG32(DMA_TILING_CONFIG + DMA1_REGISTER_OFFSET, gb_addr_config);
1108	WREG32(UVD_UDEC_ADDR_CONFIG, gb_addr_config);
1109	WREG32(UVD_UDEC_DB_ADDR_CONFIG, gb_addr_config);
1110	WREG32(UVD_UDEC_DBW_ADDR_CONFIG, gb_addr_config);
1111
1112	if ((rdev->config.cayman.max_backends_per_se == 1) &&
1113	    (rdev->flags & RADEON_IS_IGP)) {
1114		if ((disabled_rb_mask & 3) == 2) {
1115			/* RB1 disabled, RB0 enabled */
1116			tmp = 0x00000000;
1117		} else {
1118			/* RB0 disabled, RB1 enabled */
1119			tmp = 0x11111111;
1120		}
1121	} else {
1122		tmp = gb_addr_config & NUM_PIPES_MASK;
1123		tmp = r6xx_remap_render_backend(rdev, tmp,
1124						rdev->config.cayman.max_backends_per_se *
1125						rdev->config.cayman.max_shader_engines,
1126						CAYMAN_MAX_BACKENDS, disabled_rb_mask);
1127	}
1128	rdev->config.cayman.backend_map = tmp;
1129	WREG32(GB_BACKEND_MAP, tmp);
1130
1131	cgts_tcc_disable = 0xffff0000;
1132	for (i = 0; i < rdev->config.cayman.max_texture_channel_caches; i++)
1133		cgts_tcc_disable &= ~(1 << (16 + i));
1134	WREG32(CGTS_TCC_DISABLE, cgts_tcc_disable);
1135	WREG32(CGTS_SYS_TCC_DISABLE, cgts_tcc_disable);
 
 
 
 
 
 
1136	WREG32(CGTS_USER_SYS_TCC_DISABLE, cgts_tcc_disable);
1137	WREG32(CGTS_USER_TCC_DISABLE, cgts_tcc_disable);
1138
1139	/* reprogram the shader complex */
1140	cgts_sm_ctrl_reg = RREG32(CGTS_SM_CTRL_REG);
1141	for (i = 0; i < 16; i++)
1142		WREG32(CGTS_SM_CTRL_REG, OVERRIDE);
1143	WREG32(CGTS_SM_CTRL_REG, cgts_sm_ctrl_reg);
1144
1145	/* set HW defaults for 3D engine */
1146	WREG32(CP_MEQ_THRESHOLDS, MEQ1_START(0x30) | MEQ2_START(0x60));
1147
1148	sx_debug_1 = RREG32(SX_DEBUG_1);
1149	sx_debug_1 |= ENABLE_NEW_SMX_ADDRESS;
1150	WREG32(SX_DEBUG_1, sx_debug_1);
1151
1152	smx_dc_ctl0 = RREG32(SMX_DC_CTL0);
1153	smx_dc_ctl0 &= ~NUMBER_OF_SETS(0x1ff);
1154	smx_dc_ctl0 |= NUMBER_OF_SETS(rdev->config.cayman.sx_num_of_sets);
1155	WREG32(SMX_DC_CTL0, smx_dc_ctl0);
1156
1157	WREG32(SPI_CONFIG_CNTL_1, VTX_DONE_DELAY(4) | CRC_SIMD_ID_WADDR_DISABLE);
1158
1159	/* need to be explicitly zero-ed */
1160	WREG32(VGT_OFFCHIP_LDS_BASE, 0);
1161	WREG32(SQ_LSTMP_RING_BASE, 0);
1162	WREG32(SQ_HSTMP_RING_BASE, 0);
1163	WREG32(SQ_ESTMP_RING_BASE, 0);
1164	WREG32(SQ_GSTMP_RING_BASE, 0);
1165	WREG32(SQ_VSTMP_RING_BASE, 0);
1166	WREG32(SQ_PSTMP_RING_BASE, 0);
1167
1168	WREG32(TA_CNTL_AUX, DISABLE_CUBE_ANISO);
1169
1170	WREG32(SX_EXPORT_BUFFER_SIZES, (COLOR_BUFFER_SIZE((rdev->config.cayman.sx_max_export_size / 4) - 1) |
1171					POSITION_BUFFER_SIZE((rdev->config.cayman.sx_max_export_pos_size / 4) - 1) |
1172					SMX_BUFFER_SIZE((rdev->config.cayman.sx_max_export_smx_size / 4) - 1)));
1173
1174	WREG32(PA_SC_FIFO_SIZE, (SC_PRIM_FIFO_SIZE(rdev->config.cayman.sc_prim_fifo_size) |
1175				 SC_HIZ_TILE_FIFO_SIZE(rdev->config.cayman.sc_hiz_tile_fifo_size) |
1176				 SC_EARLYZ_TILE_FIFO_SIZE(rdev->config.cayman.sc_earlyz_tile_fifo_size)));
1177
1178
1179	WREG32(VGT_NUM_INSTANCES, 1);
1180
1181	WREG32(CP_PERFMON_CNTL, 0);
1182
1183	WREG32(SQ_MS_FIFO_SIZES, (CACHE_FIFO_SIZE(16 * rdev->config.cayman.sq_num_cf_insts) |
1184				  FETCH_FIFO_HIWATER(0x4) |
1185				  DONE_FIFO_HIWATER(0xe0) |
1186				  ALU_UPDATE_FIFO_HIWATER(0x8)));
1187
1188	WREG32(SQ_GPR_RESOURCE_MGMT_1, NUM_CLAUSE_TEMP_GPRS(4));
1189	WREG32(SQ_CONFIG, (VC_ENABLE |
1190			   EXPORT_SRC_C |
1191			   GFX_PRIO(0) |
1192			   CS1_PRIO(0) |
1193			   CS2_PRIO(1)));
1194	WREG32(SQ_DYN_GPR_CNTL_PS_FLUSH_REQ, DYN_GPR_ENABLE);
1195
1196	WREG32(PA_SC_FORCE_EOV_MAX_CNTS, (FORCE_EOV_MAX_CLK_CNT(4095) |
1197					  FORCE_EOV_MAX_REZ_CNT(255)));
1198
1199	WREG32(VGT_CACHE_INVALIDATION, CACHE_INVALIDATION(VC_AND_TC) |
1200	       AUTO_INVLD_EN(ES_AND_GS_AUTO));
1201
1202	WREG32(VGT_GS_VERTEX_REUSE, 16);
1203	WREG32(PA_SC_LINE_STIPPLE_STATE, 0);
1204
1205	WREG32(CB_PERF_CTR0_SEL_0, 0);
1206	WREG32(CB_PERF_CTR0_SEL_1, 0);
1207	WREG32(CB_PERF_CTR1_SEL_0, 0);
1208	WREG32(CB_PERF_CTR1_SEL_1, 0);
1209	WREG32(CB_PERF_CTR2_SEL_0, 0);
1210	WREG32(CB_PERF_CTR2_SEL_1, 0);
1211	WREG32(CB_PERF_CTR3_SEL_0, 0);
1212	WREG32(CB_PERF_CTR3_SEL_1, 0);
1213
1214	tmp = RREG32(HDP_MISC_CNTL);
1215	tmp |= HDP_FLUSH_INVALIDATE_CACHE;
1216	WREG32(HDP_MISC_CNTL, tmp);
1217
1218	hdp_host_path_cntl = RREG32(HDP_HOST_PATH_CNTL);
1219	WREG32(HDP_HOST_PATH_CNTL, hdp_host_path_cntl);
1220
1221	WREG32(PA_CL_ENHANCE, CLIP_VTX_REORDER_ENA | NUM_CLIP_SEQ(3));
1222
1223	udelay(50);
1224
1225	/* set clockgating golden values on TN */
1226	if (rdev->family == CHIP_ARUBA) {
1227		tmp = RREG32_CG(CG_CGTT_LOCAL_0);
1228		tmp &= ~0x00380000;
1229		WREG32_CG(CG_CGTT_LOCAL_0, tmp);
1230		tmp = RREG32_CG(CG_CGTT_LOCAL_1);
1231		tmp &= ~0x0e000000;
1232		WREG32_CG(CG_CGTT_LOCAL_1, tmp);
1233	}
1234}
1235
1236/*
1237 * GART
1238 */
1239void cayman_pcie_gart_tlb_flush(struct radeon_device *rdev)
1240{
1241	/* flush hdp cache */
1242	WREG32(HDP_MEM_COHERENCY_FLUSH_CNTL, 0x1);
1243
1244	/* bits 0-7 are the VM contexts0-7 */
1245	WREG32(VM_INVALIDATE_REQUEST, 1);
1246}
1247
1248static int cayman_pcie_gart_enable(struct radeon_device *rdev)
1249{
1250	int i, r;
1251
1252	if (rdev->gart.robj == NULL) {
1253		dev_err(rdev->dev, "No VRAM object for PCIE GART.\n");
1254		return -EINVAL;
1255	}
1256	r = radeon_gart_table_vram_pin(rdev);
1257	if (r)
1258		return r;
 
1259	/* Setup TLB control */
1260	WREG32(MC_VM_MX_L1_TLB_CNTL,
1261	       (0xA << 7) |
1262	       ENABLE_L1_TLB |
1263	       ENABLE_L1_FRAGMENT_PROCESSING |
1264	       SYSTEM_ACCESS_MODE_NOT_IN_SYS |
1265	       ENABLE_ADVANCED_DRIVER_MODEL |
1266	       SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU);
1267	/* Setup L2 cache */
1268	WREG32(VM_L2_CNTL, ENABLE_L2_CACHE |
1269	       ENABLE_L2_FRAGMENT_PROCESSING |
1270	       ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
1271	       ENABLE_L2_PDE0_CACHE_LRU_UPDATE_BY_WRITE |
1272	       EFFECTIVE_L2_QUEUE_SIZE(7) |
1273	       CONTEXT1_IDENTITY_ACCESS_MODE(1));
1274	WREG32(VM_L2_CNTL2, INVALIDATE_ALL_L1_TLBS | INVALIDATE_L2_CACHE);
1275	WREG32(VM_L2_CNTL3, L2_CACHE_BIGK_ASSOCIATIVITY |
1276	       BANK_SELECT(6) |
1277	       L2_CACHE_BIGK_FRAGMENT_SIZE(6));
1278	/* setup context0 */
1279	WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR, rdev->mc.gtt_start >> 12);
1280	WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR, rdev->mc.gtt_end >> 12);
1281	WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12);
1282	WREG32(VM_CONTEXT0_PROTECTION_FAULT_DEFAULT_ADDR,
1283			(u32)(rdev->dummy_page.addr >> 12));
1284	WREG32(VM_CONTEXT0_CNTL2, 0);
1285	WREG32(VM_CONTEXT0_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) |
1286				RANGE_PROTECTION_FAULT_ENABLE_DEFAULT);
1287
1288	WREG32(0x15D4, 0);
1289	WREG32(0x15D8, 0);
1290	WREG32(0x15DC, 0);
1291
1292	/* empty context1-7 */
1293	/* Assign the pt base to something valid for now; the pts used for
1294	 * the VMs are determined by the application and setup and assigned
1295	 * on the fly in the vm part of radeon_gart.c
1296	 */
1297	for (i = 1; i < 8; i++) {
1298		WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR + (i << 2), 0);
1299		WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR + (i << 2),
1300			rdev->vm_manager.max_pfn - 1);
1301		WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (i << 2),
1302		       rdev->vm_manager.saved_table_addr[i]);
1303	}
1304
1305	/* enable context1-7 */
1306	WREG32(VM_CONTEXT1_PROTECTION_FAULT_DEFAULT_ADDR,
1307	       (u32)(rdev->dummy_page.addr >> 12));
1308	WREG32(VM_CONTEXT1_CNTL2, 4);
1309	WREG32(VM_CONTEXT1_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(1) |
1310				PAGE_TABLE_BLOCK_SIZE(radeon_vm_block_size - 9) |
1311				RANGE_PROTECTION_FAULT_ENABLE_INTERRUPT |
1312				RANGE_PROTECTION_FAULT_ENABLE_DEFAULT |
1313				DUMMY_PAGE_PROTECTION_FAULT_ENABLE_INTERRUPT |
1314				DUMMY_PAGE_PROTECTION_FAULT_ENABLE_DEFAULT |
1315				PDE0_PROTECTION_FAULT_ENABLE_INTERRUPT |
1316				PDE0_PROTECTION_FAULT_ENABLE_DEFAULT |
1317				VALID_PROTECTION_FAULT_ENABLE_INTERRUPT |
1318				VALID_PROTECTION_FAULT_ENABLE_DEFAULT |
1319				READ_PROTECTION_FAULT_ENABLE_INTERRUPT |
1320				READ_PROTECTION_FAULT_ENABLE_DEFAULT |
1321				WRITE_PROTECTION_FAULT_ENABLE_INTERRUPT |
1322				WRITE_PROTECTION_FAULT_ENABLE_DEFAULT);
1323
1324	cayman_pcie_gart_tlb_flush(rdev);
1325	DRM_INFO("PCIE GART of %uM enabled (table at 0x%016llX).\n",
1326		 (unsigned)(rdev->mc.gtt_size >> 20),
1327		 (unsigned long long)rdev->gart.table_addr);
1328	rdev->gart.ready = true;
1329	return 0;
1330}
1331
1332static void cayman_pcie_gart_disable(struct radeon_device *rdev)
1333{
1334	unsigned i;
1335
1336	for (i = 1; i < 8; ++i) {
1337		rdev->vm_manager.saved_table_addr[i] = RREG32(
1338			VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (i << 2));
1339	}
1340
1341	/* Disable all tables */
1342	WREG32(VM_CONTEXT0_CNTL, 0);
1343	WREG32(VM_CONTEXT1_CNTL, 0);
1344	/* Setup TLB control */
1345	WREG32(MC_VM_MX_L1_TLB_CNTL, ENABLE_L1_FRAGMENT_PROCESSING |
1346	       SYSTEM_ACCESS_MODE_NOT_IN_SYS |
1347	       SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU);
1348	/* Setup L2 cache */
1349	WREG32(VM_L2_CNTL, ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
1350	       ENABLE_L2_PDE0_CACHE_LRU_UPDATE_BY_WRITE |
1351	       EFFECTIVE_L2_QUEUE_SIZE(7) |
1352	       CONTEXT1_IDENTITY_ACCESS_MODE(1));
1353	WREG32(VM_L2_CNTL2, 0);
1354	WREG32(VM_L2_CNTL3, L2_CACHE_BIGK_ASSOCIATIVITY |
1355	       L2_CACHE_BIGK_FRAGMENT_SIZE(6));
1356	radeon_gart_table_vram_unpin(rdev);
 
 
 
 
 
 
 
1357}
1358
1359static void cayman_pcie_gart_fini(struct radeon_device *rdev)
1360{
1361	cayman_pcie_gart_disable(rdev);
1362	radeon_gart_table_vram_free(rdev);
1363	radeon_gart_fini(rdev);
1364}
1365
1366void cayman_cp_int_cntl_setup(struct radeon_device *rdev,
1367			      int ring, u32 cp_int_cntl)
1368{
1369	WREG32(SRBM_GFX_CNTL, RINGID(ring));
1370	WREG32(CP_INT_CNTL, cp_int_cntl);
1371}
1372
1373/*
1374 * CP.
1375 */
1376void cayman_fence_ring_emit(struct radeon_device *rdev,
1377			    struct radeon_fence *fence)
1378{
1379	struct radeon_ring *ring = &rdev->ring[fence->ring];
1380	u64 addr = rdev->fence_drv[fence->ring].gpu_addr;
1381	u32 cp_coher_cntl = PACKET3_FULL_CACHE_ENA | PACKET3_TC_ACTION_ENA |
1382		PACKET3_SH_ACTION_ENA;
1383
1384	/* flush read cache over gart for this vmid */
1385	radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_SYNC, 3));
1386	radeon_ring_write(ring, PACKET3_ENGINE_ME | cp_coher_cntl);
1387	radeon_ring_write(ring, 0xFFFFFFFF);
1388	radeon_ring_write(ring, 0);
1389	radeon_ring_write(ring, 10); /* poll interval */
1390	/* EVENT_WRITE_EOP - flush caches, send int */
1391	radeon_ring_write(ring, PACKET3(PACKET3_EVENT_WRITE_EOP, 4));
1392	radeon_ring_write(ring, EVENT_TYPE(CACHE_FLUSH_AND_INV_EVENT_TS) | EVENT_INDEX(5));
1393	radeon_ring_write(ring, lower_32_bits(addr));
1394	radeon_ring_write(ring, (upper_32_bits(addr) & 0xff) | DATA_SEL(1) | INT_SEL(2));
1395	radeon_ring_write(ring, fence->seq);
1396	radeon_ring_write(ring, 0);
1397}
1398
1399void cayman_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib)
1400{
1401	struct radeon_ring *ring = &rdev->ring[ib->ring];
1402	unsigned vm_id = ib->vm ? ib->vm->ids[ib->ring].id : 0;
1403	u32 cp_coher_cntl = PACKET3_FULL_CACHE_ENA | PACKET3_TC_ACTION_ENA |
1404		PACKET3_SH_ACTION_ENA;
1405
1406	/* set to DX10/11 mode */
1407	radeon_ring_write(ring, PACKET3(PACKET3_MODE_CONTROL, 0));
1408	radeon_ring_write(ring, 1);
1409
1410	if (ring->rptr_save_reg) {
1411		uint32_t next_rptr = ring->wptr + 3 + 4 + 8;
1412		radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1));
1413		radeon_ring_write(ring, ((ring->rptr_save_reg - 
1414					  PACKET3_SET_CONFIG_REG_START) >> 2));
1415		radeon_ring_write(ring, next_rptr);
1416	}
1417
1418	radeon_ring_write(ring, PACKET3(PACKET3_INDIRECT_BUFFER, 2));
1419	radeon_ring_write(ring,
1420#ifdef __BIG_ENDIAN
1421			  (2 << 0) |
1422#endif
1423			  (ib->gpu_addr & 0xFFFFFFFC));
1424	radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF);
1425	radeon_ring_write(ring, ib->length_dw | (vm_id << 24));
1426
1427	/* flush read cache over gart for this vmid */
1428	radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_SYNC, 3));
1429	radeon_ring_write(ring, PACKET3_ENGINE_ME | cp_coher_cntl);
1430	radeon_ring_write(ring, 0xFFFFFFFF);
1431	radeon_ring_write(ring, 0);
1432	radeon_ring_write(ring, (vm_id << 24) | 10); /* poll interval */
1433}
1434
1435static void cayman_cp_enable(struct radeon_device *rdev, bool enable)
1436{
1437	if (enable)
1438		WREG32(CP_ME_CNTL, 0);
1439	else {
1440		if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
1441			radeon_ttm_set_active_vram_size(rdev, rdev->mc.visible_vram_size);
1442		WREG32(CP_ME_CNTL, (CP_ME_HALT | CP_PFP_HALT));
1443		WREG32(SCRATCH_UMSK, 0);
1444		rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ready = false;
1445	}
1446}
1447
1448u32 cayman_gfx_get_rptr(struct radeon_device *rdev,
1449			struct radeon_ring *ring)
1450{
1451	u32 rptr;
1452
1453	if (rdev->wb.enabled)
1454		rptr = rdev->wb.wb[ring->rptr_offs/4];
1455	else {
1456		if (ring->idx == RADEON_RING_TYPE_GFX_INDEX)
1457			rptr = RREG32(CP_RB0_RPTR);
1458		else if (ring->idx == CAYMAN_RING_TYPE_CP1_INDEX)
1459			rptr = RREG32(CP_RB1_RPTR);
1460		else
1461			rptr = RREG32(CP_RB2_RPTR);
1462	}
1463
1464	return rptr;
1465}
1466
1467u32 cayman_gfx_get_wptr(struct radeon_device *rdev,
1468			struct radeon_ring *ring)
1469{
1470	u32 wptr;
1471
1472	if (ring->idx == RADEON_RING_TYPE_GFX_INDEX)
1473		wptr = RREG32(CP_RB0_WPTR);
1474	else if (ring->idx == CAYMAN_RING_TYPE_CP1_INDEX)
1475		wptr = RREG32(CP_RB1_WPTR);
1476	else
1477		wptr = RREG32(CP_RB2_WPTR);
1478
1479	return wptr;
1480}
1481
1482void cayman_gfx_set_wptr(struct radeon_device *rdev,
1483			 struct radeon_ring *ring)
1484{
1485	if (ring->idx == RADEON_RING_TYPE_GFX_INDEX) {
1486		WREG32(CP_RB0_WPTR, ring->wptr);
1487		(void)RREG32(CP_RB0_WPTR);
1488	} else if (ring->idx == CAYMAN_RING_TYPE_CP1_INDEX) {
1489		WREG32(CP_RB1_WPTR, ring->wptr);
1490		(void)RREG32(CP_RB1_WPTR);
1491	} else {
1492		WREG32(CP_RB2_WPTR, ring->wptr);
1493		(void)RREG32(CP_RB2_WPTR);
1494	}
1495}
1496
1497static int cayman_cp_load_microcode(struct radeon_device *rdev)
1498{
1499	const __be32 *fw_data;
1500	int i;
1501
1502	if (!rdev->me_fw || !rdev->pfp_fw)
1503		return -EINVAL;
1504
1505	cayman_cp_enable(rdev, false);
1506
1507	fw_data = (const __be32 *)rdev->pfp_fw->data;
1508	WREG32(CP_PFP_UCODE_ADDR, 0);
1509	for (i = 0; i < CAYMAN_PFP_UCODE_SIZE; i++)
1510		WREG32(CP_PFP_UCODE_DATA, be32_to_cpup(fw_data++));
1511	WREG32(CP_PFP_UCODE_ADDR, 0);
1512
1513	fw_data = (const __be32 *)rdev->me_fw->data;
1514	WREG32(CP_ME_RAM_WADDR, 0);
1515	for (i = 0; i < CAYMAN_PM4_UCODE_SIZE; i++)
1516		WREG32(CP_ME_RAM_DATA, be32_to_cpup(fw_data++));
1517
1518	WREG32(CP_PFP_UCODE_ADDR, 0);
1519	WREG32(CP_ME_RAM_WADDR, 0);
1520	WREG32(CP_ME_RAM_RADDR, 0);
1521	return 0;
1522}
1523
1524static int cayman_cp_start(struct radeon_device *rdev)
1525{
1526	struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
1527	int r, i;
1528
1529	r = radeon_ring_lock(rdev, ring, 7);
1530	if (r) {
1531		DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
1532		return r;
1533	}
1534	radeon_ring_write(ring, PACKET3(PACKET3_ME_INITIALIZE, 5));
1535	radeon_ring_write(ring, 0x1);
1536	radeon_ring_write(ring, 0x0);
1537	radeon_ring_write(ring, rdev->config.cayman.max_hw_contexts - 1);
1538	radeon_ring_write(ring, PACKET3_ME_INITIALIZE_DEVICE_ID(1));
1539	radeon_ring_write(ring, 0);
1540	radeon_ring_write(ring, 0);
1541	radeon_ring_unlock_commit(rdev, ring, false);
1542
1543	cayman_cp_enable(rdev, true);
1544
1545	r = radeon_ring_lock(rdev, ring, cayman_default_size + 19);
1546	if (r) {
1547		DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
1548		return r;
1549	}
1550
1551	/* setup clear context state */
1552	radeon_ring_write(ring, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
1553	radeon_ring_write(ring, PACKET3_PREAMBLE_BEGIN_CLEAR_STATE);
1554
1555	for (i = 0; i < cayman_default_size; i++)
1556		radeon_ring_write(ring, cayman_default_state[i]);
1557
1558	radeon_ring_write(ring, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
1559	radeon_ring_write(ring, PACKET3_PREAMBLE_END_CLEAR_STATE);
1560
1561	/* set clear context state */
1562	radeon_ring_write(ring, PACKET3(PACKET3_CLEAR_STATE, 0));
1563	radeon_ring_write(ring, 0);
1564
1565	/* SQ_VTX_BASE_VTX_LOC */
1566	radeon_ring_write(ring, 0xc0026f00);
1567	radeon_ring_write(ring, 0x00000000);
1568	radeon_ring_write(ring, 0x00000000);
1569	radeon_ring_write(ring, 0x00000000);
1570
1571	/* Clear consts */
1572	radeon_ring_write(ring, 0xc0036f00);
1573	radeon_ring_write(ring, 0x00000bc4);
1574	radeon_ring_write(ring, 0xffffffff);
1575	radeon_ring_write(ring, 0xffffffff);
1576	radeon_ring_write(ring, 0xffffffff);
1577
1578	radeon_ring_write(ring, 0xc0026900);
1579	radeon_ring_write(ring, 0x00000316);
1580	radeon_ring_write(ring, 0x0000000e); /* VGT_VERTEX_REUSE_BLOCK_CNTL */
1581	radeon_ring_write(ring, 0x00000010); /*  */
1582
1583	radeon_ring_unlock_commit(rdev, ring, false);
1584
1585	/* XXX init other rings */
1586
1587	return 0;
1588}
1589
1590static void cayman_cp_fini(struct radeon_device *rdev)
1591{
1592	struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
1593	cayman_cp_enable(rdev, false);
1594	radeon_ring_fini(rdev, ring);
1595	radeon_scratch_free(rdev, ring->rptr_save_reg);
1596}
1597
1598static int cayman_cp_resume(struct radeon_device *rdev)
1599{
1600	static const int ridx[] = {
1601		RADEON_RING_TYPE_GFX_INDEX,
1602		CAYMAN_RING_TYPE_CP1_INDEX,
1603		CAYMAN_RING_TYPE_CP2_INDEX
1604	};
1605	static const unsigned cp_rb_cntl[] = {
1606		CP_RB0_CNTL,
1607		CP_RB1_CNTL,
1608		CP_RB2_CNTL,
1609	};
1610	static const unsigned cp_rb_rptr_addr[] = {
1611		CP_RB0_RPTR_ADDR,
1612		CP_RB1_RPTR_ADDR,
1613		CP_RB2_RPTR_ADDR
1614	};
1615	static const unsigned cp_rb_rptr_addr_hi[] = {
1616		CP_RB0_RPTR_ADDR_HI,
1617		CP_RB1_RPTR_ADDR_HI,
1618		CP_RB2_RPTR_ADDR_HI
1619	};
1620	static const unsigned cp_rb_base[] = {
1621		CP_RB0_BASE,
1622		CP_RB1_BASE,
1623		CP_RB2_BASE
1624	};
1625	static const unsigned cp_rb_rptr[] = {
1626		CP_RB0_RPTR,
1627		CP_RB1_RPTR,
1628		CP_RB2_RPTR
1629	};
1630	static const unsigned cp_rb_wptr[] = {
1631		CP_RB0_WPTR,
1632		CP_RB1_WPTR,
1633		CP_RB2_WPTR
1634	};
1635	struct radeon_ring *ring;
1636	int i, r;
1637
1638	/* Reset cp; if cp is reset, then PA, SH, VGT also need to be reset */
1639	WREG32(GRBM_SOFT_RESET, (SOFT_RESET_CP |
1640				 SOFT_RESET_PA |
1641				 SOFT_RESET_SH |
1642				 SOFT_RESET_VGT |
1643				 SOFT_RESET_SPI |
1644				 SOFT_RESET_SX));
1645	RREG32(GRBM_SOFT_RESET);
1646	mdelay(15);
1647	WREG32(GRBM_SOFT_RESET, 0);
1648	RREG32(GRBM_SOFT_RESET);
1649
1650	WREG32(CP_SEM_WAIT_TIMER, 0x0);
1651	WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0);
1652
1653	/* Set the write pointer delay */
1654	WREG32(CP_RB_WPTR_DELAY, 0);
1655
1656	WREG32(CP_DEBUG, (1 << 27));
1657
1658	/* set the wb address whether it's enabled or not */
1659	WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
1660	WREG32(SCRATCH_UMSK, 0xff);
1661
1662	for (i = 0; i < 3; ++i) {
1663		uint32_t rb_cntl;
1664		uint64_t addr;
1665
1666		/* Set ring buffer size */
1667		ring = &rdev->ring[ridx[i]];
1668		rb_cntl = order_base_2(ring->ring_size / 8);
1669		rb_cntl |= order_base_2(RADEON_GPU_PAGE_SIZE/8) << 8;
1670#ifdef __BIG_ENDIAN
1671		rb_cntl |= BUF_SWAP_32BIT;
1672#endif
1673		WREG32(cp_rb_cntl[i], rb_cntl);
1674
1675		/* set the wb address whether it's enabled or not */
1676		addr = rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET;
1677		WREG32(cp_rb_rptr_addr[i], addr & 0xFFFFFFFC);
1678		WREG32(cp_rb_rptr_addr_hi[i], upper_32_bits(addr) & 0xFF);
1679	}
 
 
 
 
1680
1681	/* set the rb base addr, this causes an internal reset of ALL rings */
1682	for (i = 0; i < 3; ++i) {
1683		ring = &rdev->ring[ridx[i]];
1684		WREG32(cp_rb_base[i], ring->gpu_addr >> 8);
 
1685	}
1686
1687	for (i = 0; i < 3; ++i) {
1688		/* Initialize the ring buffer's read and write pointers */
1689		ring = &rdev->ring[ridx[i]];
1690		WREG32_P(cp_rb_cntl[i], RB_RPTR_WR_ENA, ~RB_RPTR_WR_ENA);
1691
1692		ring->wptr = 0;
1693		WREG32(cp_rb_rptr[i], 0);
1694		WREG32(cp_rb_wptr[i], ring->wptr);
1695
1696		mdelay(1);
1697		WREG32_P(cp_rb_cntl[i], 0, ~RB_RPTR_WR_ENA);
1698	}
1699
1700	/* start the rings */
1701	cayman_cp_start(rdev);
1702	rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ready = true;
1703	rdev->ring[CAYMAN_RING_TYPE_CP1_INDEX].ready = false;
1704	rdev->ring[CAYMAN_RING_TYPE_CP2_INDEX].ready = false;
1705	/* this only test cp0 */
1706	r = radeon_ring_test(rdev, RADEON_RING_TYPE_GFX_INDEX, &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]);
1707	if (r) {
1708		rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ready = false;
1709		rdev->ring[CAYMAN_RING_TYPE_CP1_INDEX].ready = false;
1710		rdev->ring[CAYMAN_RING_TYPE_CP2_INDEX].ready = false;
1711		return r;
1712	}
1713
1714	if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
1715		radeon_ttm_set_active_vram_size(rdev, rdev->mc.real_vram_size);
 
 
1716
1717	return 0;
1718}
 
1719
1720u32 cayman_gpu_check_soft_reset(struct radeon_device *rdev)
1721{
1722	u32 reset_mask = 0;
1723	u32 tmp;
1724
1725	/* GRBM_STATUS */
1726	tmp = RREG32(GRBM_STATUS);
1727	if (tmp & (PA_BUSY | SC_BUSY |
1728		   SH_BUSY | SX_BUSY |
1729		   TA_BUSY | VGT_BUSY |
1730		   DB_BUSY | CB_BUSY |
1731		   GDS_BUSY | SPI_BUSY |
1732		   IA_BUSY | IA_BUSY_NO_DMA))
1733		reset_mask |= RADEON_RESET_GFX;
1734
1735	if (tmp & (CF_RQ_PENDING | PF_RQ_PENDING |
1736		   CP_BUSY | CP_COHERENCY_BUSY))
1737		reset_mask |= RADEON_RESET_CP;
1738
1739	if (tmp & GRBM_EE_BUSY)
1740		reset_mask |= RADEON_RESET_GRBM | RADEON_RESET_GFX | RADEON_RESET_CP;
 
 
 
 
 
 
1741
1742	/* DMA_STATUS_REG 0 */
1743	tmp = RREG32(DMA_STATUS_REG + DMA0_REGISTER_OFFSET);
1744	if (!(tmp & DMA_IDLE))
1745		reset_mask |= RADEON_RESET_DMA;
1746
1747	/* DMA_STATUS_REG 1 */
1748	tmp = RREG32(DMA_STATUS_REG + DMA1_REGISTER_OFFSET);
1749	if (!(tmp & DMA_IDLE))
1750		reset_mask |= RADEON_RESET_DMA1;
1751
1752	/* SRBM_STATUS2 */
1753	tmp = RREG32(SRBM_STATUS2);
1754	if (tmp & DMA_BUSY)
1755		reset_mask |= RADEON_RESET_DMA;
1756
1757	if (tmp & DMA1_BUSY)
1758		reset_mask |= RADEON_RESET_DMA1;
1759
1760	/* SRBM_STATUS */
1761	tmp = RREG32(SRBM_STATUS);
1762	if (tmp & (RLC_RQ_PENDING | RLC_BUSY))
1763		reset_mask |= RADEON_RESET_RLC;
1764
1765	if (tmp & IH_BUSY)
1766		reset_mask |= RADEON_RESET_IH;
 
 
 
 
 
 
 
 
 
 
 
1767
1768	if (tmp & SEM_BUSY)
1769		reset_mask |= RADEON_RESET_SEM;
1770
1771	if (tmp & GRBM_RQ_PENDING)
1772		reset_mask |= RADEON_RESET_GRBM;
 
 
 
 
 
1773
1774	if (tmp & VMC_BUSY)
1775		reset_mask |= RADEON_RESET_VMC;
1776
1777	if (tmp & (MCB_BUSY | MCB_NON_DISPLAY_BUSY |
1778		   MCC_BUSY | MCD_BUSY))
1779		reset_mask |= RADEON_RESET_MC;
1780
1781	if (evergreen_is_display_hung(rdev))
1782		reset_mask |= RADEON_RESET_DISPLAY;
1783
1784	/* VM_L2_STATUS */
1785	tmp = RREG32(VM_L2_STATUS);
1786	if (tmp & L2_BUSY)
1787		reset_mask |= RADEON_RESET_VMC;
1788
1789	/* Skip MC reset as it's mostly likely not hung, just busy */
1790	if (reset_mask & RADEON_RESET_MC) {
1791		DRM_DEBUG("MC busy: 0x%08X, clearing.\n", reset_mask);
1792		reset_mask &= ~RADEON_RESET_MC;
1793	}
1794
1795	return reset_mask;
 
 
 
 
 
 
 
 
 
1796}
1797
1798static void cayman_gpu_soft_reset(struct radeon_device *rdev, u32 reset_mask)
1799{
1800	struct evergreen_mc_save save;
1801	u32 grbm_soft_reset = 0, srbm_soft_reset = 0;
1802	u32 tmp;
1803
1804	if (reset_mask == 0)
1805		return;
1806
1807	dev_info(rdev->dev, "GPU softreset: 0x%08X\n", reset_mask);
1808
1809	evergreen_print_gpu_status_regs(rdev);
1810	dev_info(rdev->dev, "  VM_CONTEXT0_PROTECTION_FAULT_ADDR   0x%08X\n",
1811		 RREG32(0x14F8));
1812	dev_info(rdev->dev, "  VM_CONTEXT0_PROTECTION_FAULT_STATUS 0x%08X\n",
1813		 RREG32(0x14D8));
1814	dev_info(rdev->dev, "  VM_CONTEXT1_PROTECTION_FAULT_ADDR   0x%08X\n",
1815		 RREG32(0x14FC));
1816	dev_info(rdev->dev, "  VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n",
1817		 RREG32(0x14DC));
1818
1819	/* Disable CP parsing/prefetching */
1820	WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT);
1821
1822	if (reset_mask & RADEON_RESET_DMA) {
1823		/* dma0 */
1824		tmp = RREG32(DMA_RB_CNTL + DMA0_REGISTER_OFFSET);
1825		tmp &= ~DMA_RB_ENABLE;
1826		WREG32(DMA_RB_CNTL + DMA0_REGISTER_OFFSET, tmp);
1827	}
1828
1829	if (reset_mask & RADEON_RESET_DMA1) {
1830		/* dma1 */
1831		tmp = RREG32(DMA_RB_CNTL + DMA1_REGISTER_OFFSET);
1832		tmp &= ~DMA_RB_ENABLE;
1833		WREG32(DMA_RB_CNTL + DMA1_REGISTER_OFFSET, tmp);
1834	}
1835
1836	udelay(50);
1837
 
 
 
 
 
 
 
 
 
1838	evergreen_mc_stop(rdev, &save);
1839	if (evergreen_mc_wait_for_idle(rdev)) {
1840		dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
1841	}
 
 
1842
1843	if (reset_mask & (RADEON_RESET_GFX | RADEON_RESET_COMPUTE)) {
1844		grbm_soft_reset = SOFT_RESET_CB |
1845			SOFT_RESET_DB |
1846			SOFT_RESET_GDS |
1847			SOFT_RESET_PA |
1848			SOFT_RESET_SC |
1849			SOFT_RESET_SPI |
1850			SOFT_RESET_SH |
1851			SOFT_RESET_SX |
1852			SOFT_RESET_TC |
1853			SOFT_RESET_TA |
1854			SOFT_RESET_VGT |
1855			SOFT_RESET_IA;
1856	}
1857
1858	if (reset_mask & RADEON_RESET_CP) {
1859		grbm_soft_reset |= SOFT_RESET_CP | SOFT_RESET_VGT;
1860
1861		srbm_soft_reset |= SOFT_RESET_GRBM;
1862	}
1863
1864	if (reset_mask & RADEON_RESET_DMA)
1865		srbm_soft_reset |= SOFT_RESET_DMA;
1866
1867	if (reset_mask & RADEON_RESET_DMA1)
1868		srbm_soft_reset |= SOFT_RESET_DMA1;
1869
1870	if (reset_mask & RADEON_RESET_DISPLAY)
1871		srbm_soft_reset |= SOFT_RESET_DC;
1872
1873	if (reset_mask & RADEON_RESET_RLC)
1874		srbm_soft_reset |= SOFT_RESET_RLC;
1875
1876	if (reset_mask & RADEON_RESET_SEM)
1877		srbm_soft_reset |= SOFT_RESET_SEM;
1878
1879	if (reset_mask & RADEON_RESET_IH)
1880		srbm_soft_reset |= SOFT_RESET_IH;
1881
1882	if (reset_mask & RADEON_RESET_GRBM)
1883		srbm_soft_reset |= SOFT_RESET_GRBM;
1884
1885	if (reset_mask & RADEON_RESET_VMC)
1886		srbm_soft_reset |= SOFT_RESET_VMC;
1887
1888	if (!(rdev->flags & RADEON_IS_IGP)) {
1889		if (reset_mask & RADEON_RESET_MC)
1890			srbm_soft_reset |= SOFT_RESET_MC;
1891	}
1892
1893	if (grbm_soft_reset) {
1894		tmp = RREG32(GRBM_SOFT_RESET);
1895		tmp |= grbm_soft_reset;
1896		dev_info(rdev->dev, "GRBM_SOFT_RESET=0x%08X\n", tmp);
1897		WREG32(GRBM_SOFT_RESET, tmp);
1898		tmp = RREG32(GRBM_SOFT_RESET);
1899
1900		udelay(50);
1901
1902		tmp &= ~grbm_soft_reset;
1903		WREG32(GRBM_SOFT_RESET, tmp);
1904		tmp = RREG32(GRBM_SOFT_RESET);
1905	}
1906
1907	if (srbm_soft_reset) {
1908		tmp = RREG32(SRBM_SOFT_RESET);
1909		tmp |= srbm_soft_reset;
1910		dev_info(rdev->dev, "SRBM_SOFT_RESET=0x%08X\n", tmp);
1911		WREG32(SRBM_SOFT_RESET, tmp);
1912		tmp = RREG32(SRBM_SOFT_RESET);
1913
1914		udelay(50);
1915
1916		tmp &= ~srbm_soft_reset;
1917		WREG32(SRBM_SOFT_RESET, tmp);
1918		tmp = RREG32(SRBM_SOFT_RESET);
1919	}
1920
1921	/* Wait a little for things to settle down */
1922	udelay(50);
1923
 
 
 
 
 
 
 
1924	evergreen_mc_resume(rdev, &save);
1925	udelay(50);
1926
1927	evergreen_print_gpu_status_regs(rdev);
1928}
1929
1930int cayman_asic_reset(struct radeon_device *rdev, bool hard)
1931{
1932	u32 reset_mask;
1933
1934	if (hard) {
1935		evergreen_gpu_pci_config_reset(rdev);
1936		return 0;
1937	}
1938
1939	reset_mask = cayman_gpu_check_soft_reset(rdev);
1940
1941	if (reset_mask)
1942		r600_set_bios_scratch_engine_hung(rdev, true);
1943
1944	cayman_gpu_soft_reset(rdev, reset_mask);
1945
1946	reset_mask = cayman_gpu_check_soft_reset(rdev);
1947
1948	if (reset_mask)
1949		evergreen_gpu_pci_config_reset(rdev);
1950
1951	r600_set_bios_scratch_engine_hung(rdev, false);
1952
1953	return 0;
1954}
1955
1956/**
1957 * cayman_gfx_is_lockup - Check if the GFX engine is locked up
1958 *
1959 * @rdev: radeon_device pointer
1960 * @ring: radeon_ring structure holding ring information
1961 *
1962 * Check if the GFX engine is locked up.
1963 * Returns true if the engine appears to be locked up, false if not.
1964 */
1965bool cayman_gfx_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
1966{
1967	u32 reset_mask = cayman_gpu_check_soft_reset(rdev);
1968
1969	if (!(reset_mask & (RADEON_RESET_GFX |
1970			    RADEON_RESET_COMPUTE |
1971			    RADEON_RESET_CP))) {
1972		radeon_ring_lockup_update(rdev, ring);
1973		return false;
1974	}
1975	return radeon_ring_test_lockup(rdev, ring);
1976}
1977
1978static void cayman_uvd_init(struct radeon_device *rdev)
1979{
1980	int r;
1981
1982	if (!rdev->has_uvd)
1983		return;
1984
1985	r = radeon_uvd_init(rdev);
1986	if (r) {
1987		dev_err(rdev->dev, "failed UVD (%d) init.\n", r);
1988		/*
1989		 * At this point rdev->uvd.vcpu_bo is NULL which trickles down
1990		 * to early fails uvd_v2_2_resume() and thus nothing happens
1991		 * there. So it is pointless to try to go through that code
1992		 * hence why we disable uvd here.
1993		 */
1994		rdev->has_uvd = false;
1995		return;
1996	}
1997	rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL;
1998	r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096);
1999}
2000
2001static void cayman_uvd_start(struct radeon_device *rdev)
2002{
2003	int r;
2004
2005	if (!rdev->has_uvd)
2006		return;
2007
2008	r = uvd_v2_2_resume(rdev);
2009	if (r) {
2010		dev_err(rdev->dev, "failed UVD resume (%d).\n", r);
2011		goto error;
2012	}
2013	r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX);
2014	if (r) {
2015		dev_err(rdev->dev, "failed initializing UVD fences (%d).\n", r);
2016		goto error;
2017	}
2018	return;
2019
2020error:
2021	rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0;
2022}
2023
2024static void cayman_uvd_resume(struct radeon_device *rdev)
2025{
2026	struct radeon_ring *ring;
2027	int r;
2028
2029	if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size)
2030		return;
2031
2032	ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
2033	r = radeon_ring_init(rdev, ring, ring->ring_size, 0, PACKET0(UVD_NO_OP, 0));
2034	if (r) {
2035		dev_err(rdev->dev, "failed initializing UVD ring (%d).\n", r);
2036		return;
2037	}
2038	r = uvd_v1_0_init(rdev);
2039	if (r) {
2040		dev_err(rdev->dev, "failed initializing UVD (%d).\n", r);
2041		return;
2042	}
2043}
2044
2045static void cayman_vce_init(struct radeon_device *rdev)
2046{
2047	int r;
2048
2049	/* Only set for CHIP_ARUBA */
2050	if (!rdev->has_vce)
2051		return;
2052
2053	r = radeon_vce_init(rdev);
2054	if (r) {
2055		dev_err(rdev->dev, "failed VCE (%d) init.\n", r);
2056		/*
2057		 * At this point rdev->vce.vcpu_bo is NULL which trickles down
2058		 * to early fails cayman_vce_start() and thus nothing happens
2059		 * there. So it is pointless to try to go through that code
2060		 * hence why we disable vce here.
2061		 */
2062		rdev->has_vce = false;
2063		return;
2064	}
2065	rdev->ring[TN_RING_TYPE_VCE1_INDEX].ring_obj = NULL;
2066	r600_ring_init(rdev, &rdev->ring[TN_RING_TYPE_VCE1_INDEX], 4096);
2067	rdev->ring[TN_RING_TYPE_VCE2_INDEX].ring_obj = NULL;
2068	r600_ring_init(rdev, &rdev->ring[TN_RING_TYPE_VCE2_INDEX], 4096);
2069}
2070
2071static void cayman_vce_start(struct radeon_device *rdev)
2072{
2073	int r;
2074
2075	if (!rdev->has_vce)
2076		return;
2077
2078	r = radeon_vce_resume(rdev);
2079	if (r) {
2080		dev_err(rdev->dev, "failed VCE resume (%d).\n", r);
2081		goto error;
2082	}
2083	r = vce_v1_0_resume(rdev);
2084	if (r) {
2085		dev_err(rdev->dev, "failed VCE resume (%d).\n", r);
2086		goto error;
2087	}
2088	r = radeon_fence_driver_start_ring(rdev, TN_RING_TYPE_VCE1_INDEX);
2089	if (r) {
2090		dev_err(rdev->dev, "failed initializing VCE1 fences (%d).\n", r);
2091		goto error;
2092	}
2093	r = radeon_fence_driver_start_ring(rdev, TN_RING_TYPE_VCE2_INDEX);
2094	if (r) {
2095		dev_err(rdev->dev, "failed initializing VCE2 fences (%d).\n", r);
2096		goto error;
2097	}
2098	return;
2099
2100error:
2101	rdev->ring[TN_RING_TYPE_VCE1_INDEX].ring_size = 0;
2102	rdev->ring[TN_RING_TYPE_VCE2_INDEX].ring_size = 0;
2103}
2104
2105static void cayman_vce_resume(struct radeon_device *rdev)
2106{
2107	struct radeon_ring *ring;
2108	int r;
2109
2110	if (!rdev->has_vce || !rdev->ring[TN_RING_TYPE_VCE1_INDEX].ring_size)
2111		return;
2112
2113	ring = &rdev->ring[TN_RING_TYPE_VCE1_INDEX];
2114	r = radeon_ring_init(rdev, ring, ring->ring_size, 0, 0x0);
2115	if (r) {
2116		dev_err(rdev->dev, "failed initializing VCE1 ring (%d).\n", r);
2117		return;
2118	}
2119	ring = &rdev->ring[TN_RING_TYPE_VCE2_INDEX];
2120	r = radeon_ring_init(rdev, ring, ring->ring_size, 0, 0x0);
2121	if (r) {
2122		dev_err(rdev->dev, "failed initializing VCE1 ring (%d).\n", r);
2123		return;
2124	}
2125	r = vce_v1_0_init(rdev);
2126	if (r) {
2127		dev_err(rdev->dev, "failed initializing VCE (%d).\n", r);
2128		return;
2129	}
2130}
2131
2132static int cayman_startup(struct radeon_device *rdev)
2133{
2134	struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
2135	int r;
2136
2137	/* enable pcie gen2 link */
2138	evergreen_pcie_gen2_enable(rdev);
2139	/* enable aspm */
2140	evergreen_program_aspm(rdev);
2141
2142	/* scratch needs to be initialized before MC */
2143	r = r600_vram_scratch_init(rdev);
2144	if (r)
2145		return r;
2146
2147	evergreen_mc_program(rdev);
2148
2149	if (!(rdev->flags & RADEON_IS_IGP) && !rdev->pm.dpm_enabled) {
2150		r = ni_mc_load_microcode(rdev);
2151		if (r) {
2152			DRM_ERROR("Failed to load MC firmware!\n");
2153			return r;
2154		}
2155	}
 
 
 
 
 
2156
 
2157	r = cayman_pcie_gart_enable(rdev);
2158	if (r)
2159		return r;
2160	cayman_gpu_init(rdev);
2161
2162	/* allocate rlc buffers */
2163	if (rdev->flags & RADEON_IS_IGP) {
2164		rdev->rlc.reg_list = tn_rlc_save_restore_register_list;
2165		rdev->rlc.reg_list_size =
2166			(u32)ARRAY_SIZE(tn_rlc_save_restore_register_list);
2167		rdev->rlc.cs_data = cayman_cs_data;
2168		r = sumo_rlc_init(rdev);
2169		if (r) {
2170			DRM_ERROR("Failed to init rlc BOs!\n");
2171			return r;
2172		}
2173	}
2174
2175	/* allocate wb buffer */
2176	r = radeon_wb_init(rdev);
2177	if (r)
2178		return r;
2179
2180	r = radeon_fence_driver_start_ring(rdev, RADEON_RING_TYPE_GFX_INDEX);
2181	if (r) {
2182		dev_err(rdev->dev, "failed initializing CP fences (%d).\n", r);
2183		return r;
2184	}
2185
2186	cayman_uvd_start(rdev);
2187	cayman_vce_start(rdev);
2188
2189	r = radeon_fence_driver_start_ring(rdev, CAYMAN_RING_TYPE_CP1_INDEX);
2190	if (r) {
2191		dev_err(rdev->dev, "failed initializing CP fences (%d).\n", r);
2192		return r;
2193	}
2194
2195	r = radeon_fence_driver_start_ring(rdev, CAYMAN_RING_TYPE_CP2_INDEX);
2196	if (r) {
2197		dev_err(rdev->dev, "failed initializing CP fences (%d).\n", r);
2198		return r;
2199	}
2200
2201	r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX);
2202	if (r) {
2203		dev_err(rdev->dev, "failed initializing DMA fences (%d).\n", r);
2204		return r;
2205	}
2206
2207	r = radeon_fence_driver_start_ring(rdev, CAYMAN_RING_TYPE_DMA1_INDEX);
2208	if (r) {
2209		dev_err(rdev->dev, "failed initializing DMA fences (%d).\n", r);
2210		return r;
2211	}
2212
2213	/* Enable IRQ */
2214	if (!rdev->irq.installed) {
2215		r = radeon_irq_kms_init(rdev);
2216		if (r)
2217			return r;
2218	}
2219
2220	r = r600_irq_init(rdev);
2221	if (r) {
2222		DRM_ERROR("radeon: IH init failed (%d).\n", r);
2223		radeon_irq_kms_fini(rdev);
2224		return r;
2225	}
2226	evergreen_irq_set(rdev);
2227
2228	r = radeon_ring_init(rdev, ring, ring->ring_size, RADEON_WB_CP_RPTR_OFFSET,
2229			     RADEON_CP_PACKET2);
2230	if (r)
2231		return r;
2232
2233	ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX];
2234	r = radeon_ring_init(rdev, ring, ring->ring_size, R600_WB_DMA_RPTR_OFFSET,
2235			     DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0));
2236	if (r)
2237		return r;
2238
2239	ring = &rdev->ring[CAYMAN_RING_TYPE_DMA1_INDEX];
2240	r = radeon_ring_init(rdev, ring, ring->ring_size, CAYMAN_WB_DMA1_RPTR_OFFSET,
2241			     DMA_PACKET(DMA_PACKET_NOP, 0, 0, 0));
2242	if (r)
2243		return r;
2244
2245	r = cayman_cp_load_microcode(rdev);
2246	if (r)
2247		return r;
2248	r = cayman_cp_resume(rdev);
2249	if (r)
2250		return r;
2251
2252	r = cayman_dma_resume(rdev);
2253	if (r)
2254		return r;
2255
2256	cayman_uvd_resume(rdev);
2257	cayman_vce_resume(rdev);
2258
2259	r = radeon_ib_pool_init(rdev);
2260	if (r) {
2261		dev_err(rdev->dev, "IB initialization failed (%d).\n", r);
2262		return r;
2263	}
2264
2265	r = radeon_vm_manager_init(rdev);
2266	if (r) {
2267		dev_err(rdev->dev, "vm manager initialization failed (%d).\n", r);
2268		return r;
2269	}
2270
2271	r = radeon_audio_init(rdev);
2272	if (r)
2273		return r;
2274
2275	return 0;
2276}
2277
2278int cayman_resume(struct radeon_device *rdev)
2279{
2280	int r;
2281
2282	/* Do not reset GPU before posting, on rv770 hw unlike on r500 hw,
2283	 * posting will perform necessary task to bring back GPU into good
2284	 * shape.
2285	 */
2286	/* post card */
2287	atom_asic_init(rdev->mode_info.atom_context);
2288
2289	/* init golden registers */
2290	ni_init_golden_registers(rdev);
2291
2292	if (rdev->pm.pm_method == PM_METHOD_DPM)
2293		radeon_pm_resume(rdev);
2294
2295	rdev->accel_working = true;
2296	r = cayman_startup(rdev);
2297	if (r) {
2298		DRM_ERROR("cayman startup failed on resume\n");
2299		rdev->accel_working = false;
2300		return r;
2301	}
 
 
 
 
 
 
 
2302	return r;
 
2303}
2304
2305int cayman_suspend(struct radeon_device *rdev)
2306{
2307	radeon_pm_suspend(rdev);
2308	radeon_audio_fini(rdev);
2309	radeon_vm_manager_fini(rdev);
2310	cayman_cp_enable(rdev, false);
2311	cayman_dma_stop(rdev);
2312	if (rdev->has_uvd) {
2313		radeon_uvd_suspend(rdev);
2314		uvd_v1_0_fini(rdev);
2315	}
2316	evergreen_irq_suspend(rdev);
2317	radeon_wb_disable(rdev);
2318	cayman_pcie_gart_disable(rdev);
 
 
 
 
 
 
 
 
2319	return 0;
2320}
2321
2322/* Plan is to move initialization in that function and use
2323 * helper function so that radeon_device_init pretty much
2324 * do nothing more than calling asic specific function. This
2325 * should also allow to remove a bunch of callback function
2326 * like vram_info.
2327 */
2328int cayman_init(struct radeon_device *rdev)
2329{
2330	struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
2331	int r;
2332
 
 
 
 
2333	/* Read BIOS */
2334	if (!radeon_get_bios(rdev)) {
2335		if (ASIC_IS_AVIVO(rdev))
2336			return -EINVAL;
2337	}
2338	/* Must be an ATOMBIOS */
2339	if (!rdev->is_atom_bios) {
2340		dev_err(rdev->dev, "Expecting atombios for cayman GPU\n");
2341		return -EINVAL;
2342	}
2343	r = radeon_atombios_init(rdev);
2344	if (r)
2345		return r;
2346
2347	/* Post card if necessary */
2348	if (!radeon_card_posted(rdev)) {
2349		if (!rdev->bios) {
2350			dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n");
2351			return -EINVAL;
2352		}
2353		DRM_INFO("GPU not posted. posting now...\n");
2354		atom_asic_init(rdev->mode_info.atom_context);
2355	}
2356	/* init golden registers */
2357	ni_init_golden_registers(rdev);
2358	/* Initialize scratch registers */
2359	r600_scratch_init(rdev);
2360	/* Initialize surface registers */
2361	radeon_surface_init(rdev);
2362	/* Initialize clocks */
2363	radeon_get_clock_info(rdev_to_drm(rdev));
2364	/* Fence driver */
2365	radeon_fence_driver_init(rdev);
 
 
2366	/* initialize memory controller */
2367	r = evergreen_mc_init(rdev);
2368	if (r)
2369		return r;
2370	/* Memory manager */
2371	r = radeon_bo_init(rdev);
2372	if (r)
2373		return r;
2374
2375	if (rdev->flags & RADEON_IS_IGP) {
2376		if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) {
2377			r = ni_init_microcode(rdev);
2378			if (r) {
2379				DRM_ERROR("Failed to load firmware!\n");
2380				return r;
2381			}
2382		}
2383	} else {
2384		if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw || !rdev->mc_fw) {
2385			r = ni_init_microcode(rdev);
2386			if (r) {
2387				DRM_ERROR("Failed to load firmware!\n");
2388				return r;
2389			}
2390		}
2391	}
2392
2393	/* Initialize power management */
2394	radeon_pm_init(rdev);
2395
2396	ring->ring_obj = NULL;
2397	r600_ring_init(rdev, ring, 1024 * 1024);
2398
2399	ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX];
2400	ring->ring_obj = NULL;
2401	r600_ring_init(rdev, ring, 64 * 1024);
2402
2403	ring = &rdev->ring[CAYMAN_RING_TYPE_DMA1_INDEX];
2404	ring->ring_obj = NULL;
2405	r600_ring_init(rdev, ring, 64 * 1024);
2406
2407	cayman_uvd_init(rdev);
2408	cayman_vce_init(rdev);
2409
2410	rdev->ih.ring_obj = NULL;
2411	r600_ih_ring_init(rdev, 64 * 1024);
2412
2413	r = r600_pcie_gart_init(rdev);
2414	if (r)
2415		return r;
2416
2417	rdev->accel_working = true;
2418	r = cayman_startup(rdev);
2419	if (r) {
2420		dev_err(rdev->dev, "disabling GPU acceleration\n");
2421		cayman_cp_fini(rdev);
2422		cayman_dma_fini(rdev);
2423		r600_irq_fini(rdev);
2424		if (rdev->flags & RADEON_IS_IGP)
2425			sumo_rlc_fini(rdev);
2426		radeon_wb_fini(rdev);
2427		radeon_ib_pool_fini(rdev);
2428		radeon_vm_manager_fini(rdev);
2429		radeon_irq_kms_fini(rdev);
2430		cayman_pcie_gart_fini(rdev);
2431		rdev->accel_working = false;
2432	}
 
 
 
 
 
 
 
 
 
 
 
 
2433
2434	/* Don't start up if the MC ucode is missing.
2435	 * The default clocks and voltages before the MC ucode
2436	 * is loaded are not suffient for advanced operations.
2437	 *
2438	 * We can skip this check for TN, because there is no MC
2439	 * ucode.
2440	 */
2441	if (!rdev->mc_fw && !(rdev->flags & RADEON_IS_IGP)) {
2442		DRM_ERROR("radeon: MC ucode required for NI+.\n");
2443		return -EINVAL;
2444	}
2445
2446	return 0;
2447}
2448
2449void cayman_fini(struct radeon_device *rdev)
2450{
2451	radeon_pm_fini(rdev);
2452	cayman_cp_fini(rdev);
2453	cayman_dma_fini(rdev);
2454	r600_irq_fini(rdev);
2455	if (rdev->flags & RADEON_IS_IGP)
2456		sumo_rlc_fini(rdev);
2457	radeon_wb_fini(rdev);
2458	radeon_vm_manager_fini(rdev);
2459	radeon_ib_pool_fini(rdev);
2460	radeon_irq_kms_fini(rdev);
2461	uvd_v1_0_fini(rdev);
2462	radeon_uvd_fini(rdev);
2463	if (rdev->has_vce)
2464		radeon_vce_fini(rdev);
2465	cayman_pcie_gart_fini(rdev);
2466	r600_vram_scratch_fini(rdev);
2467	radeon_gem_fini(rdev);
2468	radeon_fence_driver_fini(rdev);
2469	radeon_bo_fini(rdev);
2470	radeon_atombios_fini(rdev);
2471	kfree(rdev->bios);
2472	rdev->bios = NULL;
2473}
2474
2475/*
2476 * vm
2477 */
2478int cayman_vm_init(struct radeon_device *rdev)
2479{
2480	/* number of VMs */
2481	rdev->vm_manager.nvm = 8;
2482	/* base offset of vram pages */
2483	if (rdev->flags & RADEON_IS_IGP) {
2484		u64 tmp = RREG32(FUS_MC_VM_FB_OFFSET);
2485		tmp <<= 22;
2486		rdev->vm_manager.vram_base_offset = tmp;
2487	} else
2488		rdev->vm_manager.vram_base_offset = 0;
2489	return 0;
2490}
2491
2492void cayman_vm_fini(struct radeon_device *rdev)
2493{
2494}
2495
2496/**
2497 * cayman_vm_decode_fault - print human readable fault info
2498 *
2499 * @rdev: radeon_device pointer
2500 * @status: VM_CONTEXT1_PROTECTION_FAULT_STATUS register value
2501 * @addr: VM_CONTEXT1_PROTECTION_FAULT_ADDR register value
2502 *
2503 * Print human readable fault information (cayman/TN).
2504 */
2505void cayman_vm_decode_fault(struct radeon_device *rdev,
2506			    u32 status, u32 addr)
2507{
2508	u32 mc_id = (status & MEMORY_CLIENT_ID_MASK) >> MEMORY_CLIENT_ID_SHIFT;
2509	u32 vmid = (status & FAULT_VMID_MASK) >> FAULT_VMID_SHIFT;
2510	u32 protections = (status & PROTECTIONS_MASK) >> PROTECTIONS_SHIFT;
2511	char *block;
2512
2513	switch (mc_id) {
2514	case 32:
2515	case 16:
2516	case 96:
2517	case 80:
2518	case 160:
2519	case 144:
2520	case 224:
2521	case 208:
2522		block = "CB";
2523		break;
2524	case 33:
2525	case 17:
2526	case 97:
2527	case 81:
2528	case 161:
2529	case 145:
2530	case 225:
2531	case 209:
2532		block = "CB_FMASK";
2533		break;
2534	case 34:
2535	case 18:
2536	case 98:
2537	case 82:
2538	case 162:
2539	case 146:
2540	case 226:
2541	case 210:
2542		block = "CB_CMASK";
2543		break;
2544	case 35:
2545	case 19:
2546	case 99:
2547	case 83:
2548	case 163:
2549	case 147:
2550	case 227:
2551	case 211:
2552		block = "CB_IMMED";
2553		break;
2554	case 36:
2555	case 20:
2556	case 100:
2557	case 84:
2558	case 164:
2559	case 148:
2560	case 228:
2561	case 212:
2562		block = "DB";
2563		break;
2564	case 37:
2565	case 21:
2566	case 101:
2567	case 85:
2568	case 165:
2569	case 149:
2570	case 229:
2571	case 213:
2572		block = "DB_HTILE";
2573		break;
2574	case 38:
2575	case 22:
2576	case 102:
2577	case 86:
2578	case 166:
2579	case 150:
2580	case 230:
2581	case 214:
2582		block = "SX";
2583		break;
2584	case 39:
2585	case 23:
2586	case 103:
2587	case 87:
2588	case 167:
2589	case 151:
2590	case 231:
2591	case 215:
2592		block = "DB_STEN";
2593		break;
2594	case 40:
2595	case 24:
2596	case 104:
2597	case 88:
2598	case 232:
2599	case 216:
2600	case 168:
2601	case 152:
2602		block = "TC_TFETCH";
2603		break;
2604	case 41:
2605	case 25:
2606	case 105:
2607	case 89:
2608	case 233:
2609	case 217:
2610	case 169:
2611	case 153:
2612		block = "TC_VFETCH";
2613		break;
2614	case 42:
2615	case 26:
2616	case 106:
2617	case 90:
2618	case 234:
2619	case 218:
2620	case 170:
2621	case 154:
2622		block = "VC";
2623		break;
2624	case 112:
2625		block = "CP";
2626		break;
2627	case 113:
2628	case 114:
2629		block = "SH";
2630		break;
2631	case 115:
2632		block = "VGT";
2633		break;
2634	case 178:
2635		block = "IH";
2636		break;
2637	case 51:
2638		block = "RLC";
2639		break;
2640	case 55:
2641		block = "DMA";
2642		break;
2643	case 56:
2644		block = "HDP";
2645		break;
2646	default:
2647		block = "unknown";
2648		break;
2649	}
2650
2651	printk("VM fault (0x%02x, vmid %d) at page %u, %s from %s (%d)\n",
2652	       protections, vmid, addr,
2653	       (status & MEMORY_CLIENT_RW_MASK) ? "write" : "read",
2654	       block, mc_id);
2655}
2656
2657/*
2658 * cayman_vm_flush - vm flush using the CP
2659 *
2660 * Update the page table base and flush the VM TLB
2661 * using the CP (cayman-si).
2662 */
2663void cayman_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring,
2664		     unsigned vm_id, uint64_t pd_addr)
2665{
2666	radeon_ring_write(ring, PACKET0(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR + (vm_id << 2), 0));
2667	radeon_ring_write(ring, pd_addr >> 12);
2668
2669	/* flush hdp cache */
2670	radeon_ring_write(ring, PACKET0(HDP_MEM_COHERENCY_FLUSH_CNTL, 0));
2671	radeon_ring_write(ring, 0x1);
2672
2673	/* bits 0-7 are the VM contexts0-7 */
2674	radeon_ring_write(ring, PACKET0(VM_INVALIDATE_REQUEST, 0));
2675	radeon_ring_write(ring, 1 << vm_id);
2676
2677	/* wait for the invalidate to complete */
2678	radeon_ring_write(ring, PACKET3(PACKET3_WAIT_REG_MEM, 5));
2679	radeon_ring_write(ring, (WAIT_REG_MEM_FUNCTION(0) |  /* always */
2680				 WAIT_REG_MEM_ENGINE(0))); /* me */
2681	radeon_ring_write(ring, VM_INVALIDATE_REQUEST >> 2);
2682	radeon_ring_write(ring, 0);
2683	radeon_ring_write(ring, 0); /* ref */
2684	radeon_ring_write(ring, 0); /* mask */
2685	radeon_ring_write(ring, 0x20); /* poll interval */
2686
2687	/* sync PFP to ME, otherwise we might get invalid PFP reads */
2688	radeon_ring_write(ring, PACKET3(PACKET3_PFP_SYNC_ME, 0));
2689	radeon_ring_write(ring, 0x0);
2690}
2691
2692int tn_set_vce_clocks(struct radeon_device *rdev, u32 evclk, u32 ecclk)
2693{
2694	struct atom_clock_dividers dividers;
2695	int r, i;
2696
2697	r = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM,
2698					   ecclk, false, &dividers);
2699	if (r)
2700		return r;
2701
2702	for (i = 0; i < 100; i++) {
2703		if (RREG32(CG_ECLK_STATUS) & ECLK_STATUS)
2704			break;
2705		mdelay(10);
2706	}
2707	if (i == 100)
2708		return -ETIMEDOUT;
2709
2710	WREG32_P(CG_ECLK_CNTL, dividers.post_div, ~(ECLK_DIR_CNTL_EN|ECLK_DIVIDER_MASK));
2711
2712	for (i = 0; i < 100; i++) {
2713		if (RREG32(CG_ECLK_STATUS) & ECLK_STATUS)
2714			break;
2715		mdelay(10);
2716	}
2717	if (i == 100)
2718		return -ETIMEDOUT;
2719
2720	return 0;
2721}