diff options
Diffstat (limited to 'drivers/gpu/drm/msm/adreno/a3xx_gpu.c')
| -rw-r--r-- | drivers/gpu/drm/msm/adreno/a3xx_gpu.c | 709 | 
1 files changed, 709 insertions, 0 deletions
diff --git a/drivers/gpu/drm/msm/adreno/a3xx_gpu.c b/drivers/gpu/drm/msm/adreno/a3xx_gpu.c new file mode 100644 index 00000000000..942e09d898a --- /dev/null +++ b/drivers/gpu/drm/msm/adreno/a3xx_gpu.c @@ -0,0 +1,709 @@ +/* + * Copyright (C) 2013 Red Hat + * Author: Rob Clark <robdclark@gmail.com> + * + * This program is free software; you can redistribute it and/or modify it + * under the terms of the GNU General Public License version 2 as published by + * the Free Software Foundation. + * + * This program is distributed in the hope that it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for + * more details. + * + * You should have received a copy of the GNU General Public License along with + * this program.  If not, see <http://www.gnu.org/licenses/>. + */ + +#ifdef CONFIG_MSM_OCMEM +#  include <mach/ocmem.h> +#endif + +#include "a3xx_gpu.h" + +#define A3XX_INT0_MASK \ +	(A3XX_INT0_RBBM_AHB_ERROR |        \ +	 A3XX_INT0_RBBM_ATB_BUS_OVERFLOW | \ +	 A3XX_INT0_CP_T0_PACKET_IN_IB |    \ +	 A3XX_INT0_CP_OPCODE_ERROR |       \ +	 A3XX_INT0_CP_RESERVED_BIT_ERROR | \ +	 A3XX_INT0_CP_HW_FAULT |           \ +	 A3XX_INT0_CP_IB1_INT |            \ +	 A3XX_INT0_CP_IB2_INT |            \ +	 A3XX_INT0_CP_RB_INT |             \ +	 A3XX_INT0_CP_REG_PROTECT_FAULT |  \ +	 A3XX_INT0_CP_AHB_ERROR_HALT |     \ +	 A3XX_INT0_UCHE_OOB_ACCESS) + + +static bool hang_debug = false; +MODULE_PARM_DESC(hang_debug, "Dump registers when hang is detected (can be slow!)"); +module_param_named(hang_debug, hang_debug, bool, 0600); +static void a3xx_dump(struct msm_gpu *gpu); + +static void a3xx_me_init(struct msm_gpu *gpu) +{ +	struct msm_ringbuffer *ring = gpu->rb; + +	OUT_PKT3(ring, CP_ME_INIT, 17); +	OUT_RING(ring, 0x000003f7); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000080); +	OUT_RING(ring, 0x00000100); +	OUT_RING(ring, 0x00000180); +	OUT_RING(ring, 0x00006600); +	OUT_RING(ring, 0x00000150); +	OUT_RING(ring, 0x0000014e); +	OUT_RING(ring, 0x00000154); +	OUT_RING(ring, 0x00000001); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000000); +	OUT_RING(ring, 0x00000000); + +	gpu->funcs->flush(gpu); +	gpu->funcs->idle(gpu); +} + +static int a3xx_hw_init(struct msm_gpu *gpu) +{ +	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); +	struct a3xx_gpu *a3xx_gpu = to_a3xx_gpu(adreno_gpu); +	uint32_t *ptr, len; +	int i, ret; + +	DBG("%s", gpu->name); + +	if (adreno_is_a305(adreno_gpu)) { +		/* Set up 16 deep read/write request queues: */ +		gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF1, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_RD_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_WR_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_DDR_OUT_MAX_BURST, 0x0000303); +		gpu_write(gpu, REG_A3XX_VBIF_IN_WR_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_IN_WR_LIM_CONF1, 0x10101010); +		/* Enable WR-REQ: */ +		gpu_write(gpu, REG_A3XX_VBIF_GATE_OFF_WRREQ_EN, 0x0000ff); +		/* Set up round robin arbitration between both AXI ports: */ +		gpu_write(gpu, REG_A3XX_VBIF_ARB_CTL, 0x00000030); +		/* Set up AOOO: */ +		gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO_EN, 0x0000003c); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO, 0x003c003c); + +	} else if (adreno_is_a320(adreno_gpu)) { +		/* Set up 16 deep read/write request queues: */ +		gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF1, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_RD_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_WR_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_DDR_OUT_MAX_BURST, 0x0000303); +		gpu_write(gpu, REG_A3XX_VBIF_IN_WR_LIM_CONF0, 0x10101010); +		gpu_write(gpu, REG_A3XX_VBIF_IN_WR_LIM_CONF1, 0x10101010); +		/* Enable WR-REQ: */ +		gpu_write(gpu, REG_A3XX_VBIF_GATE_OFF_WRREQ_EN, 0x0000ff); +		/* Set up round robin arbitration between both AXI ports: */ +		gpu_write(gpu, REG_A3XX_VBIF_ARB_CTL, 0x00000030); +		/* Set up AOOO: */ +		gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO_EN, 0x0000003c); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO, 0x003c003c); +		/* Enable 1K sort: */ +		gpu_write(gpu, REG_A3XX_VBIF_ABIT_SORT, 0x000000ff); +		gpu_write(gpu, REG_A3XX_VBIF_ABIT_SORT_CONF, 0x000000a4); + +	} else if (adreno_is_a330v2(adreno_gpu)) { +		/* +		 * Most of the VBIF registers on 8974v2 have the correct +		 * values at power on, so we won't modify those if we don't +		 * need to +		 */ +		/* Enable 1k sort: */ +		gpu_write(gpu, REG_A3XX_VBIF_ABIT_SORT, 0x0001003f); +		gpu_write(gpu, REG_A3XX_VBIF_ABIT_SORT_CONF, 0x000000a4); +		/* Enable WR-REQ: */ +		gpu_write(gpu, REG_A3XX_VBIF_GATE_OFF_WRREQ_EN, 0x00003f); +		gpu_write(gpu, REG_A3XX_VBIF_DDR_OUT_MAX_BURST, 0x0000303); +		/* Set up VBIF_ROUND_ROBIN_QOS_ARB: */ +		gpu_write(gpu, REG_A3XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x0003); + +	} else if (adreno_is_a330(adreno_gpu)) { +		/* Set up 16 deep read/write request queues: */ +		gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF0, 0x18181818); +		gpu_write(gpu, REG_A3XX_VBIF_IN_RD_LIM_CONF1, 0x18181818); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_RD_LIM_CONF0, 0x18181818); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_WR_LIM_CONF0, 0x18181818); +		gpu_write(gpu, REG_A3XX_VBIF_DDR_OUT_MAX_BURST, 0x0000303); +		gpu_write(gpu, REG_A3XX_VBIF_IN_WR_LIM_CONF0, 0x18181818); +		gpu_write(gpu, REG_A3XX_VBIF_IN_WR_LIM_CONF1, 0x18181818); +		/* Enable WR-REQ: */ +		gpu_write(gpu, REG_A3XX_VBIF_GATE_OFF_WRREQ_EN, 0x00003f); +		/* Set up round robin arbitration between both AXI ports: */ +		gpu_write(gpu, REG_A3XX_VBIF_ARB_CTL, 0x00000030); +		/* Set up VBIF_ROUND_ROBIN_QOS_ARB: */ +		gpu_write(gpu, REG_A3XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x0001); +		/* Set up AOOO: */ +		gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO_EN, 0x0000003f); +		gpu_write(gpu, REG_A3XX_VBIF_OUT_AXI_AOOO, 0x003f003f); +		/* Enable 1K sort: */ +		gpu_write(gpu, REG_A3XX_VBIF_ABIT_SORT, 0x0001003f); +		gpu_write(gpu, REG_A3XX_VBIF_ABIT_SORT_CONF, 0x000000a4); +		/* Disable VBIF clock gating. This is to enable AXI running +		 * higher frequency than GPU: +		 */ +		gpu_write(gpu, REG_A3XX_VBIF_CLKON, 0x00000001); + +	} else { +		BUG(); +	} + +	/* Make all blocks contribute to the GPU BUSY perf counter: */ +	gpu_write(gpu, REG_A3XX_RBBM_GPU_BUSY_MASKED, 0xffffffff); + +	/* Tune the hystersis counters for SP and CP idle detection: */ +	gpu_write(gpu, REG_A3XX_RBBM_SP_HYST_CNT, 0x10); +	gpu_write(gpu, REG_A3XX_RBBM_WAIT_IDLE_CLOCKS_CTL, 0x10); + +	/* Enable the RBBM error reporting bits.  This lets us get +	 * useful information on failure: +	 */ +	gpu_write(gpu, REG_A3XX_RBBM_AHB_CTL0, 0x00000001); + +	/* Enable AHB error reporting: */ +	gpu_write(gpu, REG_A3XX_RBBM_AHB_CTL1, 0xa6ffffff); + +	/* Turn on the power counters: */ +	gpu_write(gpu, REG_A3XX_RBBM_RBBM_CTL, 0x00030000); + +	/* Turn on hang detection - this spews a lot of useful information +	 * into the RBBM registers on a hang: +	 */ +	gpu_write(gpu, REG_A3XX_RBBM_INTERFACE_HANG_INT_CTL, 0x00010fff); + +	/* Enable 64-byte cacheline size. HW Default is 32-byte (0x000000E0): */ +	gpu_write(gpu, REG_A3XX_UCHE_CACHE_MODE_CONTROL_REG, 0x00000001); + +	/* Enable Clock gating: */ +	if (adreno_is_a320(adreno_gpu)) +		gpu_write(gpu, REG_A3XX_RBBM_CLOCK_CTL, 0xbfffffff); +	else if (adreno_is_a330v2(adreno_gpu)) +		gpu_write(gpu, REG_A3XX_RBBM_CLOCK_CTL, 0xaaaaaaaa); +	else if (adreno_is_a330(adreno_gpu)) +		gpu_write(gpu, REG_A3XX_RBBM_CLOCK_CTL, 0xbffcffff); + +	if (adreno_is_a330v2(adreno_gpu)) +		gpu_write(gpu, REG_A3XX_RBBM_GPR0_CTL, 0x05515455); +	else if (adreno_is_a330(adreno_gpu)) +		gpu_write(gpu, REG_A3XX_RBBM_GPR0_CTL, 0x00000000); + +	/* Set the OCMEM base address for A330, etc */ +	if (a3xx_gpu->ocmem_hdl) { +		gpu_write(gpu, REG_A3XX_RB_GMEM_BASE_ADDR, +			(unsigned int)(a3xx_gpu->ocmem_base >> 14)); +	} + +	/* Turn on performance counters: */ +	gpu_write(gpu, REG_A3XX_RBBM_PERFCTR_CTL, 0x01); + +	/* Enable the perfcntrs that we use.. */ +	for (i = 0; i < gpu->num_perfcntrs; i++) { +		const struct msm_gpu_perfcntr *perfcntr = &gpu->perfcntrs[i]; +		gpu_write(gpu, perfcntr->select_reg, perfcntr->select_val); +	} + +	gpu_write(gpu, REG_A3XX_RBBM_INT_0_MASK, A3XX_INT0_MASK); + +	ret = adreno_hw_init(gpu); +	if (ret) +		return ret; + +	/* setup access protection: */ +	gpu_write(gpu, REG_A3XX_CP_PROTECT_CTRL, 0x00000007); + +	/* RBBM registers */ +	gpu_write(gpu, REG_A3XX_CP_PROTECT(0), 0x63000040); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(1), 0x62000080); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(2), 0x600000cc); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(3), 0x60000108); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(4), 0x64000140); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(5), 0x66000400); + +	/* CP registers */ +	gpu_write(gpu, REG_A3XX_CP_PROTECT(6), 0x65000700); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(7), 0x610007d8); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(8), 0x620007e0); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(9), 0x61001178); +	gpu_write(gpu, REG_A3XX_CP_PROTECT(10), 0x64001180); + +	/* RB registers */ +	gpu_write(gpu, REG_A3XX_CP_PROTECT(11), 0x60003300); + +	/* VBIF registers */ +	gpu_write(gpu, REG_A3XX_CP_PROTECT(12), 0x6b00c000); + +	/* NOTE: PM4/micro-engine firmware registers look to be the same +	 * for a2xx and a3xx.. we could possibly push that part down to +	 * adreno_gpu base class.  Or push both PM4 and PFP but +	 * parameterize the pfp ucode addr/data registers.. +	 */ + +	/* Load PM4: */ +	ptr = (uint32_t *)(adreno_gpu->pm4->data); +	len = adreno_gpu->pm4->size / 4; +	DBG("loading PM4 ucode version: %x", ptr[1]); + +	gpu_write(gpu, REG_AXXX_CP_DEBUG, +			AXXX_CP_DEBUG_DYNAMIC_CLK_DISABLE | +			AXXX_CP_DEBUG_MIU_128BIT_WRITE_ENABLE); +	gpu_write(gpu, REG_AXXX_CP_ME_RAM_WADDR, 0); +	for (i = 1; i < len; i++) +		gpu_write(gpu, REG_AXXX_CP_ME_RAM_DATA, ptr[i]); + +	/* Load PFP: */ +	ptr = (uint32_t *)(adreno_gpu->pfp->data); +	len = adreno_gpu->pfp->size / 4; +	DBG("loading PFP ucode version: %x", ptr[5]); + +	gpu_write(gpu, REG_A3XX_CP_PFP_UCODE_ADDR, 0); +	for (i = 1; i < len; i++) +		gpu_write(gpu, REG_A3XX_CP_PFP_UCODE_DATA, ptr[i]); + +	/* CP ROQ queue sizes (bytes) - RB:16, ST:16, IB1:32, IB2:64 */ +	if (adreno_is_a305(adreno_gpu) || adreno_is_a320(adreno_gpu)) { +		gpu_write(gpu, REG_AXXX_CP_QUEUE_THRESHOLDS, +				AXXX_CP_QUEUE_THRESHOLDS_CSQ_IB1_START(2) | +				AXXX_CP_QUEUE_THRESHOLDS_CSQ_IB2_START(6) | +				AXXX_CP_QUEUE_THRESHOLDS_CSQ_ST_START(14)); +	} else if (adreno_is_a330(adreno_gpu)) { +		/* NOTE: this (value take from downstream android driver) +		 * includes some bits outside of the known bitfields.  But +		 * A330 has this "MERCIU queue" thing too, which might +		 * explain a new bitfield or reshuffling: +		 */ +		gpu_write(gpu, REG_AXXX_CP_QUEUE_THRESHOLDS, 0x003e2008); +	} + +	/* clear ME_HALT to start micro engine */ +	gpu_write(gpu, REG_AXXX_CP_ME_CNTL, 0); + +	a3xx_me_init(gpu); + +	return 0; +} + +static void a3xx_recover(struct msm_gpu *gpu) +{ +	/* dump registers before resetting gpu, if enabled: */ +	if (hang_debug) +		a3xx_dump(gpu); +	gpu_write(gpu, REG_A3XX_RBBM_SW_RESET_CMD, 1); +	gpu_read(gpu, REG_A3XX_RBBM_SW_RESET_CMD); +	gpu_write(gpu, REG_A3XX_RBBM_SW_RESET_CMD, 0); +	adreno_recover(gpu); +} + +static void a3xx_destroy(struct msm_gpu *gpu) +{ +	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); +	struct a3xx_gpu *a3xx_gpu = to_a3xx_gpu(adreno_gpu); + +	DBG("%s", gpu->name); + +	adreno_gpu_cleanup(adreno_gpu); + +#ifdef CONFIG_MSM_OCMEM +	if (a3xx_gpu->ocmem_base) +		ocmem_free(OCMEM_GRAPHICS, a3xx_gpu->ocmem_hdl); +#endif + +	kfree(a3xx_gpu); +} + +static void a3xx_idle(struct msm_gpu *gpu) +{ +	/* wait for ringbuffer to drain: */ +	adreno_idle(gpu); + +	/* then wait for GPU to finish: */ +	if (spin_until(!(gpu_read(gpu, REG_A3XX_RBBM_STATUS) & +			A3XX_RBBM_STATUS_GPU_BUSY))) +		DRM_ERROR("%s: timeout waiting for GPU to idle!\n", gpu->name); + +	/* TODO maybe we need to reset GPU here to recover from hang? */ +} + +static irqreturn_t a3xx_irq(struct msm_gpu *gpu) +{ +	uint32_t status; + +	status = gpu_read(gpu, REG_A3XX_RBBM_INT_0_STATUS); +	DBG("%s: %08x", gpu->name, status); + +	// TODO + +	gpu_write(gpu, REG_A3XX_RBBM_INT_CLEAR_CMD, status); + +	msm_gpu_retire(gpu); + +	return IRQ_HANDLED; +} + +static const unsigned int a3xx_registers[] = { +	0x0000, 0x0002, 0x0010, 0x0012, 0x0018, 0x0018, 0x0020, 0x0027, +	0x0029, 0x002b, 0x002e, 0x0033, 0x0040, 0x0042, 0x0050, 0x005c, +	0x0060, 0x006c, 0x0080, 0x0082, 0x0084, 0x0088, 0x0090, 0x00e5, +	0x00ea, 0x00ed, 0x0100, 0x0100, 0x0110, 0x0123, 0x01c0, 0x01c1, +	0x01c3, 0x01c5, 0x01c7, 0x01c7, 0x01d5, 0x01d9, 0x01dc, 0x01dd, +	0x01ea, 0x01ea, 0x01ee, 0x01f1, 0x01f5, 0x01f5, 0x01fc, 0x01ff, +	0x0440, 0x0440, 0x0443, 0x0443, 0x0445, 0x0445, 0x044d, 0x044f, +	0x0452, 0x0452, 0x0454, 0x046f, 0x047c, 0x047c, 0x047f, 0x047f, +	0x0578, 0x057f, 0x0600, 0x0602, 0x0605, 0x0607, 0x060a, 0x060e, +	0x0612, 0x0614, 0x0c01, 0x0c02, 0x0c06, 0x0c1d, 0x0c3d, 0x0c3f, +	0x0c48, 0x0c4b, 0x0c80, 0x0c80, 0x0c88, 0x0c8b, 0x0ca0, 0x0cb7, +	0x0cc0, 0x0cc1, 0x0cc6, 0x0cc7, 0x0ce4, 0x0ce5, 0x0e00, 0x0e05, +	0x0e0c, 0x0e0c, 0x0e22, 0x0e23, 0x0e41, 0x0e45, 0x0e64, 0x0e65, +	0x0e80, 0x0e82, 0x0e84, 0x0e89, 0x0ea0, 0x0ea1, 0x0ea4, 0x0ea7, +	0x0ec4, 0x0ecb, 0x0ee0, 0x0ee0, 0x0f00, 0x0f01, 0x0f03, 0x0f09, +	0x2040, 0x2040, 0x2044, 0x2044, 0x2048, 0x204d, 0x2068, 0x2069, +	0x206c, 0x206d, 0x2070, 0x2070, 0x2072, 0x2072, 0x2074, 0x2075, +	0x2079, 0x207a, 0x20c0, 0x20d3, 0x20e4, 0x20ef, 0x2100, 0x2109, +	0x210c, 0x210c, 0x210e, 0x210e, 0x2110, 0x2111, 0x2114, 0x2115, +	0x21e4, 0x21e4, 0x21ea, 0x21ea, 0x21ec, 0x21ed, 0x21f0, 0x21f0, +	0x2200, 0x2212, 0x2214, 0x2217, 0x221a, 0x221a, 0x2240, 0x227e, +	0x2280, 0x228b, 0x22c0, 0x22c0, 0x22c4, 0x22ce, 0x22d0, 0x22d8, +	0x22df, 0x22e6, 0x22e8, 0x22e9, 0x22ec, 0x22ec, 0x22f0, 0x22f7, +	0x22ff, 0x22ff, 0x2340, 0x2343, 0x2348, 0x2349, 0x2350, 0x2356, +	0x2360, 0x2360, 0x2440, 0x2440, 0x2444, 0x2444, 0x2448, 0x244d, +	0x2468, 0x2469, 0x246c, 0x246d, 0x2470, 0x2470, 0x2472, 0x2472, +	0x2474, 0x2475, 0x2479, 0x247a, 0x24c0, 0x24d3, 0x24e4, 0x24ef, +	0x2500, 0x2509, 0x250c, 0x250c, 0x250e, 0x250e, 0x2510, 0x2511, +	0x2514, 0x2515, 0x25e4, 0x25e4, 0x25ea, 0x25ea, 0x25ec, 0x25ed, +	0x25f0, 0x25f0, 0x2600, 0x2612, 0x2614, 0x2617, 0x261a, 0x261a, +	0x2640, 0x267e, 0x2680, 0x268b, 0x26c0, 0x26c0, 0x26c4, 0x26ce, +	0x26d0, 0x26d8, 0x26df, 0x26e6, 0x26e8, 0x26e9, 0x26ec, 0x26ec, +	0x26f0, 0x26f7, 0x26ff, 0x26ff, 0x2740, 0x2743, 0x2748, 0x2749, +	0x2750, 0x2756, 0x2760, 0x2760, 0x300c, 0x300e, 0x301c, 0x301d, +	0x302a, 0x302a, 0x302c, 0x302d, 0x3030, 0x3031, 0x3034, 0x3036, +	0x303c, 0x303c, 0x305e, 0x305f, +}; + +#ifdef CONFIG_DEBUG_FS +static void a3xx_show(struct msm_gpu *gpu, struct seq_file *m) +{ +	struct drm_device *dev = gpu->dev; +	int i; + +	adreno_show(gpu, m); + +	mutex_lock(&dev->struct_mutex); + +	gpu->funcs->pm_resume(gpu); + +	seq_printf(m, "status:   %08x\n", +			gpu_read(gpu, REG_A3XX_RBBM_STATUS)); + +	/* dump these out in a form that can be parsed by demsm: */ +	seq_printf(m, "IO:region %s 00000000 00020000\n", gpu->name); +	for (i = 0; i < ARRAY_SIZE(a3xx_registers); i += 2) { +		uint32_t start = a3xx_registers[i]; +		uint32_t end   = a3xx_registers[i+1]; +		uint32_t addr; + +		for (addr = start; addr <= end; addr++) { +			uint32_t val = gpu_read(gpu, addr); +			seq_printf(m, "IO:R %08x %08x\n", addr<<2, val); +		} +	} + +	gpu->funcs->pm_suspend(gpu); + +	mutex_unlock(&dev->struct_mutex); +} +#endif + +/* would be nice to not have to duplicate the _show() stuff with printk(): */ +static void a3xx_dump(struct msm_gpu *gpu) +{ +	int i; + +	adreno_dump(gpu); +	printk("status:   %08x\n", +			gpu_read(gpu, REG_A3XX_RBBM_STATUS)); + +	/* dump these out in a form that can be parsed by demsm: */ +	printk("IO:region %s 00000000 00020000\n", gpu->name); +	for (i = 0; i < ARRAY_SIZE(a3xx_registers); i += 2) { +		uint32_t start = a3xx_registers[i]; +		uint32_t end   = a3xx_registers[i+1]; +		uint32_t addr; + +		for (addr = start; addr <= end; addr++) { +			uint32_t val = gpu_read(gpu, addr); +			printk("IO:R %08x %08x\n", addr<<2, val); +		} +	} +} + +static const struct adreno_gpu_funcs funcs = { +	.base = { +		.get_param = adreno_get_param, +		.hw_init = a3xx_hw_init, +		.pm_suspend = msm_gpu_pm_suspend, +		.pm_resume = msm_gpu_pm_resume, +		.recover = a3xx_recover, +		.last_fence = adreno_last_fence, +		.submit = adreno_submit, +		.flush = adreno_flush, +		.idle = a3xx_idle, +		.irq = a3xx_irq, +		.destroy = a3xx_destroy, +#ifdef CONFIG_DEBUG_FS +		.show = a3xx_show, +#endif +	}, +}; + +static const struct msm_gpu_perfcntr perfcntrs[] = { +	{ REG_A3XX_SP_PERFCOUNTER6_SELECT, REG_A3XX_RBBM_PERFCTR_SP_6_LO, +			SP_ALU_ACTIVE_CYCLES, "ALUACTIVE" }, +	{ REG_A3XX_SP_PERFCOUNTER7_SELECT, REG_A3XX_RBBM_PERFCTR_SP_7_LO, +			SP_FS_FULL_ALU_INSTRUCTIONS, "ALUFULL" }, +}; + +struct msm_gpu *a3xx_gpu_init(struct drm_device *dev) +{ +	struct a3xx_gpu *a3xx_gpu = NULL; +	struct adreno_gpu *adreno_gpu; +	struct msm_gpu *gpu; +	struct msm_drm_private *priv = dev->dev_private; +	struct platform_device *pdev = priv->gpu_pdev; +	struct adreno_platform_config *config; +	int ret; + +	if (!pdev) { +		dev_err(dev->dev, "no a3xx device\n"); +		ret = -ENXIO; +		goto fail; +	} + +	config = pdev->dev.platform_data; + +	a3xx_gpu = kzalloc(sizeof(*a3xx_gpu), GFP_KERNEL); +	if (!a3xx_gpu) { +		ret = -ENOMEM; +		goto fail; +	} + +	adreno_gpu = &a3xx_gpu->base; +	gpu = &adreno_gpu->base; + +	a3xx_gpu->pdev = pdev; + +	gpu->fast_rate = config->fast_rate; +	gpu->slow_rate = config->slow_rate; +	gpu->bus_freq  = config->bus_freq; +#ifdef CONFIG_MSM_BUS_SCALING +	gpu->bus_scale_table = config->bus_scale_table; +#endif + +	DBG("fast_rate=%u, slow_rate=%u, bus_freq=%u", +			gpu->fast_rate, gpu->slow_rate, gpu->bus_freq); + +	gpu->perfcntrs = perfcntrs; +	gpu->num_perfcntrs = ARRAY_SIZE(perfcntrs); + +	ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs, config->rev); +	if (ret) +		goto fail; + +	/* if needed, allocate gmem: */ +	if (adreno_is_a330(adreno_gpu)) { +#ifdef CONFIG_MSM_OCMEM +		/* TODO this is different/missing upstream: */ +		struct ocmem_buf *ocmem_hdl = +				ocmem_allocate(OCMEM_GRAPHICS, adreno_gpu->gmem); + +		a3xx_gpu->ocmem_hdl = ocmem_hdl; +		a3xx_gpu->ocmem_base = ocmem_hdl->addr; +		adreno_gpu->gmem = ocmem_hdl->len; +		DBG("using %dK of OCMEM at 0x%08x", adreno_gpu->gmem / 1024, +				a3xx_gpu->ocmem_base); +#endif +	} + +	if (!gpu->mmu) { +		/* TODO we think it is possible to configure the GPU to +		 * restrict access to VRAM carveout.  But the required +		 * registers are unknown.  For now just bail out and +		 * limp along with just modesetting.  If it turns out +		 * to not be possible to restrict access, then we must +		 * implement a cmdstream validator. +		 */ +		dev_err(dev->dev, "No memory protection without IOMMU\n"); +		ret = -ENXIO; +		goto fail; +	} + +	return gpu; + +fail: +	if (a3xx_gpu) +		a3xx_destroy(&a3xx_gpu->base.base); + +	return ERR_PTR(ret); +} + +/* + * The a3xx device: + */ + +#if defined(CONFIG_MSM_BUS_SCALING) && !defined(CONFIG_OF) +#  include <mach/kgsl.h> +#endif + +static void set_gpu_pdev(struct drm_device *dev, +		struct platform_device *pdev) +{ +	struct msm_drm_private *priv = dev->dev_private; +	priv->gpu_pdev = pdev; +} + +static int a3xx_bind(struct device *dev, struct device *master, void *data) +{ +	static struct adreno_platform_config config = {}; +#ifdef CONFIG_OF +	struct device_node *child, *node = dev->of_node; +	u32 val; +	int ret; + +	ret = of_property_read_u32(node, "qcom,chipid", &val); +	if (ret) { +		dev_err(dev, "could not find chipid: %d\n", ret); +		return ret; +	} + +	config.rev = ADRENO_REV((val >> 24) & 0xff, +			(val >> 16) & 0xff, (val >> 8) & 0xff, val & 0xff); + +	/* find clock rates: */ +	config.fast_rate = 0; +	config.slow_rate = ~0; +	for_each_child_of_node(node, child) { +		if (of_device_is_compatible(child, "qcom,gpu-pwrlevels")) { +			struct device_node *pwrlvl; +			for_each_child_of_node(child, pwrlvl) { +				ret = of_property_read_u32(pwrlvl, "qcom,gpu-freq", &val); +				if (ret) { +					dev_err(dev, "could not find gpu-freq: %d\n", ret); +					return ret; +				} +				config.fast_rate = max(config.fast_rate, val); +				config.slow_rate = min(config.slow_rate, val); +			} +		} +	} + +	if (!config.fast_rate) { +		dev_err(dev, "could not find clk rates\n"); +		return -ENXIO; +	} + +#else +	struct kgsl_device_platform_data *pdata = dev->platform_data; +	uint32_t version = socinfo_get_version(); +	if (cpu_is_apq8064ab()) { +		config.fast_rate = 450000000; +		config.slow_rate = 27000000; +		config.bus_freq  = 4; +		config.rev = ADRENO_REV(3, 2, 1, 0); +	} else if (cpu_is_apq8064()) { +		config.fast_rate = 400000000; +		config.slow_rate = 27000000; +		config.bus_freq  = 4; + +		if (SOCINFO_VERSION_MAJOR(version) == 2) +			config.rev = ADRENO_REV(3, 2, 0, 2); +		else if ((SOCINFO_VERSION_MAJOR(version) == 1) && +				(SOCINFO_VERSION_MINOR(version) == 1)) +			config.rev = ADRENO_REV(3, 2, 0, 1); +		else +			config.rev = ADRENO_REV(3, 2, 0, 0); + +	} else if (cpu_is_msm8960ab()) { +		config.fast_rate = 400000000; +		config.slow_rate = 320000000; +		config.bus_freq  = 4; + +		if (SOCINFO_VERSION_MINOR(version) == 0) +			config.rev = ADRENO_REV(3, 2, 1, 0); +		else +			config.rev = ADRENO_REV(3, 2, 1, 1); + +	} else if (cpu_is_msm8930()) { +		config.fast_rate = 400000000; +		config.slow_rate = 27000000; +		config.bus_freq  = 3; + +		if ((SOCINFO_VERSION_MAJOR(version) == 1) && +			(SOCINFO_VERSION_MINOR(version) == 2)) +			config.rev = ADRENO_REV(3, 0, 5, 2); +		else +			config.rev = ADRENO_REV(3, 0, 5, 0); + +	} +#  ifdef CONFIG_MSM_BUS_SCALING +	config.bus_scale_table = pdata->bus_scale_table; +#  endif +#endif +	dev->platform_data = &config; +	set_gpu_pdev(dev_get_drvdata(master), to_platform_device(dev)); +	return 0; +} + +static void a3xx_unbind(struct device *dev, struct device *master, +		void *data) +{ +	set_gpu_pdev(dev_get_drvdata(master), NULL); +} + +static const struct component_ops a3xx_ops = { +		.bind   = a3xx_bind, +		.unbind = a3xx_unbind, +}; + +static int a3xx_probe(struct platform_device *pdev) +{ +	return component_add(&pdev->dev, &a3xx_ops); +} + +static int a3xx_remove(struct platform_device *pdev) +{ +	component_del(&pdev->dev, &a3xx_ops); +	return 0; +} + +static const struct of_device_id dt_match[] = { +	{ .compatible = "qcom,kgsl-3d0" }, +	{} +}; + +static struct platform_driver a3xx_driver = { +	.probe = a3xx_probe, +	.remove = a3xx_remove, +	.driver = { +		.name = "kgsl-3d0", +		.of_match_table = dt_match, +	}, +}; + +void __init a3xx_register(void) +{ +	platform_driver_register(&a3xx_driver); +} + +void __exit a3xx_unregister(void) +{ +	platform_driver_unregister(&a3xx_driver); +}  | 
