1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (c) 2016-2017 The Linux Foundation. All rights reserved.
5 #include <linux/kernel.h>
6 #include <linux/types.h>
7 #include <linux/cpumask.h>
8 #include <linux/qcom_scm.h>
9 #include <linux/pm_opp.h>
10 #include <linux/nvmem-consumer.h>
11 #include <linux/slab.h>
16 extern bool hang_debug;
17 static void a5xx_dump(struct msm_gpu *gpu);
21 void a5xx_flush(struct msm_gpu *gpu, struct msm_ringbuffer *ring,
24 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
25 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
30 * Most flush operations need to issue a WHERE_AM_I opcode to sync up
33 if (a5xx_gpu->has_whereami && sync) {
34 OUT_PKT7(ring, CP_WHERE_AM_I, 2);
35 OUT_RING(ring, lower_32_bits(shadowptr(a5xx_gpu, ring)));
36 OUT_RING(ring, upper_32_bits(shadowptr(a5xx_gpu, ring)));
39 spin_lock_irqsave(&ring->preempt_lock, flags);
41 /* Copy the shadow to the actual register */
42 ring->cur = ring->next;
44 /* Make sure to wrap wptr if we need to */
45 wptr = get_wptr(ring);
47 spin_unlock_irqrestore(&ring->preempt_lock, flags);
49 /* Make sure everything is posted before making a decision */
52 /* Update HW if this is the current ring and we are not in preempt */
53 if (a5xx_gpu->cur_ring == ring && !a5xx_in_preempt(a5xx_gpu))
54 gpu_write(gpu, REG_A5XX_CP_RB_WPTR, wptr);
57 static void a5xx_submit_in_rb(struct msm_gpu *gpu, struct msm_gem_submit *submit)
59 struct msm_drm_private *priv = gpu->dev->dev_private;
60 struct msm_ringbuffer *ring = submit->ring;
61 struct msm_gem_object *obj;
62 uint32_t *ptr, dwords;
65 for (i = 0; i < submit->nr_cmds; i++) {
66 switch (submit->cmd[i].type) {
67 case MSM_SUBMIT_CMD_IB_TARGET_BUF:
69 case MSM_SUBMIT_CMD_CTX_RESTORE_BUF:
70 if (priv->lastctx == submit->queue->ctx)
73 case MSM_SUBMIT_CMD_BUF:
74 /* copy commands into RB: */
75 obj = submit->bos[submit->cmd[i].idx].obj;
76 dwords = submit->cmd[i].size;
78 ptr = msm_gem_get_vaddr(&obj->base);
80 /* _get_vaddr() shouldn't fail at this point,
81 * since we've already mapped it once in
87 for (i = 0; i < dwords; i++) {
88 /* normally the OUT_PKTn() would wait
89 * for space for the packet. But since
90 * we just OUT_RING() the whole thing,
91 * need to call adreno_wait_ring()
94 adreno_wait_ring(ring, 1);
95 OUT_RING(ring, ptr[i]);
98 msm_gem_put_vaddr(&obj->base);
104 a5xx_flush(gpu, ring, true);
105 a5xx_preempt_trigger(gpu);
107 /* we might not necessarily have a cmd from userspace to
108 * trigger an event to know that submit has completed, so
111 a5xx_idle(gpu, ring);
112 ring->memptrs->fence = submit->seqno;
116 static void a5xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit)
118 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
119 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
120 struct msm_drm_private *priv = gpu->dev->dev_private;
121 struct msm_ringbuffer *ring = submit->ring;
122 unsigned int i, ibs = 0;
124 if (IS_ENABLED(CONFIG_DRM_MSM_GPU_SUDO) && submit->in_rb) {
125 priv->lastctx = NULL;
126 a5xx_submit_in_rb(gpu, submit);
130 OUT_PKT7(ring, CP_PREEMPT_ENABLE_GLOBAL, 1);
131 OUT_RING(ring, 0x02);
133 /* Turn off protected mode to write to special registers */
134 OUT_PKT7(ring, CP_SET_PROTECTED_MODE, 1);
137 /* Set the save preemption record for the ring/command */
138 OUT_PKT4(ring, REG_A5XX_CP_CONTEXT_SWITCH_SAVE_ADDR_LO, 2);
139 OUT_RING(ring, lower_32_bits(a5xx_gpu->preempt_iova[submit->ring->id]));
140 OUT_RING(ring, upper_32_bits(a5xx_gpu->preempt_iova[submit->ring->id]));
142 /* Turn back on protected mode */
143 OUT_PKT7(ring, CP_SET_PROTECTED_MODE, 1);
146 /* Enable local preemption for finegrain preemption */
147 OUT_PKT7(ring, CP_PREEMPT_ENABLE_GLOBAL, 1);
148 OUT_RING(ring, 0x02);
150 /* Allow CP_CONTEXT_SWITCH_YIELD packets in the IB2 */
151 OUT_PKT7(ring, CP_YIELD_ENABLE, 1);
152 OUT_RING(ring, 0x02);
154 /* Submit the commands */
155 for (i = 0; i < submit->nr_cmds; i++) {
156 switch (submit->cmd[i].type) {
157 case MSM_SUBMIT_CMD_IB_TARGET_BUF:
159 case MSM_SUBMIT_CMD_CTX_RESTORE_BUF:
160 if (priv->lastctx == submit->queue->ctx)
163 case MSM_SUBMIT_CMD_BUF:
164 OUT_PKT7(ring, CP_INDIRECT_BUFFER_PFE, 3);
165 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova));
166 OUT_RING(ring, upper_32_bits(submit->cmd[i].iova));
167 OUT_RING(ring, submit->cmd[i].size);
174 * Write the render mode to NULL (0) to indicate to the CP that the IBs
175 * are done rendering - otherwise a lucky preemption would start
176 * replaying from the last checkpoint
178 OUT_PKT7(ring, CP_SET_RENDER_MODE, 5);
185 /* Turn off IB level preemptions */
186 OUT_PKT7(ring, CP_YIELD_ENABLE, 1);
187 OUT_RING(ring, 0x01);
189 /* Write the fence to the scratch register */
190 OUT_PKT4(ring, REG_A5XX_CP_SCRATCH_REG(2), 1);
191 OUT_RING(ring, submit->seqno);
194 * Execute a CACHE_FLUSH_TS event. This will ensure that the
195 * timestamp is written to the memory and then triggers the interrupt
197 OUT_PKT7(ring, CP_EVENT_WRITE, 4);
198 OUT_RING(ring, CP_EVENT_WRITE_0_EVENT(CACHE_FLUSH_TS) |
199 CP_EVENT_WRITE_0_IRQ);
200 OUT_RING(ring, lower_32_bits(rbmemptr(ring, fence)));
201 OUT_RING(ring, upper_32_bits(rbmemptr(ring, fence)));
202 OUT_RING(ring, submit->seqno);
204 /* Yield the floor on command completion */
205 OUT_PKT7(ring, CP_CONTEXT_SWITCH_YIELD, 4);
207 * If dword[2:1] are non zero, they specify an address for the CP to
208 * write the value of dword[3] to on preemption complete. Write 0 to
211 OUT_RING(ring, 0x00);
212 OUT_RING(ring, 0x00);
213 /* Data value - not used if the address above is 0 */
214 OUT_RING(ring, 0x01);
215 /* Set bit 0 to trigger an interrupt on preempt complete */
216 OUT_RING(ring, 0x01);
218 /* A WHERE_AM_I packet is not needed after a YIELD */
219 a5xx_flush(gpu, ring, false);
221 /* Check to see if we need to start preemption */
222 a5xx_preempt_trigger(gpu);
225 static const struct adreno_five_hwcg_regs {
229 {REG_A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
230 {REG_A5XX_RBBM_CLOCK_CNTL_SP1, 0x02222222},
231 {REG_A5XX_RBBM_CLOCK_CNTL_SP2, 0x02222222},
232 {REG_A5XX_RBBM_CLOCK_CNTL_SP3, 0x02222222},
233 {REG_A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
234 {REG_A5XX_RBBM_CLOCK_CNTL2_SP1, 0x02222220},
235 {REG_A5XX_RBBM_CLOCK_CNTL2_SP2, 0x02222220},
236 {REG_A5XX_RBBM_CLOCK_CNTL2_SP3, 0x02222220},
237 {REG_A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
238 {REG_A5XX_RBBM_CLOCK_HYST_SP1, 0x0000F3CF},
239 {REG_A5XX_RBBM_CLOCK_HYST_SP2, 0x0000F3CF},
240 {REG_A5XX_RBBM_CLOCK_HYST_SP3, 0x0000F3CF},
241 {REG_A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
242 {REG_A5XX_RBBM_CLOCK_DELAY_SP1, 0x00000080},
243 {REG_A5XX_RBBM_CLOCK_DELAY_SP2, 0x00000080},
244 {REG_A5XX_RBBM_CLOCK_DELAY_SP3, 0x00000080},
245 {REG_A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
246 {REG_A5XX_RBBM_CLOCK_CNTL_TP1, 0x22222222},
247 {REG_A5XX_RBBM_CLOCK_CNTL_TP2, 0x22222222},
248 {REG_A5XX_RBBM_CLOCK_CNTL_TP3, 0x22222222},
249 {REG_A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
250 {REG_A5XX_RBBM_CLOCK_CNTL2_TP1, 0x22222222},
251 {REG_A5XX_RBBM_CLOCK_CNTL2_TP2, 0x22222222},
252 {REG_A5XX_RBBM_CLOCK_CNTL2_TP3, 0x22222222},
253 {REG_A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
254 {REG_A5XX_RBBM_CLOCK_CNTL3_TP1, 0x00002222},
255 {REG_A5XX_RBBM_CLOCK_CNTL3_TP2, 0x00002222},
256 {REG_A5XX_RBBM_CLOCK_CNTL3_TP3, 0x00002222},
257 {REG_A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
258 {REG_A5XX_RBBM_CLOCK_HYST_TP1, 0x77777777},
259 {REG_A5XX_RBBM_CLOCK_HYST_TP2, 0x77777777},
260 {REG_A5XX_RBBM_CLOCK_HYST_TP3, 0x77777777},
261 {REG_A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
262 {REG_A5XX_RBBM_CLOCK_HYST2_TP1, 0x77777777},
263 {REG_A5XX_RBBM_CLOCK_HYST2_TP2, 0x77777777},
264 {REG_A5XX_RBBM_CLOCK_HYST2_TP3, 0x77777777},
265 {REG_A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
266 {REG_A5XX_RBBM_CLOCK_HYST3_TP1, 0x00007777},
267 {REG_A5XX_RBBM_CLOCK_HYST3_TP2, 0x00007777},
268 {REG_A5XX_RBBM_CLOCK_HYST3_TP3, 0x00007777},
269 {REG_A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
270 {REG_A5XX_RBBM_CLOCK_DELAY_TP1, 0x11111111},
271 {REG_A5XX_RBBM_CLOCK_DELAY_TP2, 0x11111111},
272 {REG_A5XX_RBBM_CLOCK_DELAY_TP3, 0x11111111},
273 {REG_A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
274 {REG_A5XX_RBBM_CLOCK_DELAY2_TP1, 0x11111111},
275 {REG_A5XX_RBBM_CLOCK_DELAY2_TP2, 0x11111111},
276 {REG_A5XX_RBBM_CLOCK_DELAY2_TP3, 0x11111111},
277 {REG_A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
278 {REG_A5XX_RBBM_CLOCK_DELAY3_TP1, 0x00001111},
279 {REG_A5XX_RBBM_CLOCK_DELAY3_TP2, 0x00001111},
280 {REG_A5XX_RBBM_CLOCK_DELAY3_TP3, 0x00001111},
281 {REG_A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
282 {REG_A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
283 {REG_A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
284 {REG_A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
285 {REG_A5XX_RBBM_CLOCK_HYST_UCHE, 0x00444444},
286 {REG_A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
287 {REG_A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
288 {REG_A5XX_RBBM_CLOCK_CNTL_RB1, 0x22222222},
289 {REG_A5XX_RBBM_CLOCK_CNTL_RB2, 0x22222222},
290 {REG_A5XX_RBBM_CLOCK_CNTL_RB3, 0x22222222},
291 {REG_A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
292 {REG_A5XX_RBBM_CLOCK_CNTL2_RB1, 0x00222222},
293 {REG_A5XX_RBBM_CLOCK_CNTL2_RB2, 0x00222222},
294 {REG_A5XX_RBBM_CLOCK_CNTL2_RB3, 0x00222222},
295 {REG_A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
296 {REG_A5XX_RBBM_CLOCK_CNTL_CCU1, 0x00022220},
297 {REG_A5XX_RBBM_CLOCK_CNTL_CCU2, 0x00022220},
298 {REG_A5XX_RBBM_CLOCK_CNTL_CCU3, 0x00022220},
299 {REG_A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
300 {REG_A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
301 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
302 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU1, 0x04040404},
303 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU2, 0x04040404},
304 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU3, 0x04040404},
305 {REG_A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
306 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
307 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_1, 0x00000002},
308 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_2, 0x00000002},
309 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_3, 0x00000002},
310 {REG_A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
311 {REG_A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
312 {REG_A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
313 {REG_A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
314 {REG_A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
315 {REG_A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
316 {REG_A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
317 {REG_A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
318 {REG_A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
319 {REG_A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
320 {REG_A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222}
322 {REG_A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
323 {REG_A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
324 {REG_A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
325 {REG_A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
326 {REG_A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
327 {REG_A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
328 {REG_A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
329 {REG_A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
330 {REG_A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
331 {REG_A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
332 {REG_A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
333 {REG_A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
334 {REG_A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
335 {REG_A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
336 {REG_A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
337 {REG_A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
338 {REG_A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
339 {REG_A5XX_RBBM_CLOCK_HYST_UCHE, 0x00FFFFF4},
340 {REG_A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
341 {REG_A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
342 {REG_A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
343 {REG_A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
344 {REG_A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
345 {REG_A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
346 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
347 {REG_A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
348 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
349 {REG_A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
350 {REG_A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
351 {REG_A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
352 {REG_A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
353 {REG_A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
354 {REG_A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
355 {REG_A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
356 {REG_A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
357 {REG_A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
358 {REG_A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
359 {REG_A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222},
361 {REG_A5XX_RBBM_CLOCK_CNTL_SP0, 0x02222222},
362 {REG_A5XX_RBBM_CLOCK_CNTL_SP1, 0x02222222},
363 {REG_A5XX_RBBM_CLOCK_CNTL2_SP0, 0x02222220},
364 {REG_A5XX_RBBM_CLOCK_CNTL2_SP1, 0x02222220},
365 {REG_A5XX_RBBM_CLOCK_HYST_SP0, 0x0000F3CF},
366 {REG_A5XX_RBBM_CLOCK_HYST_SP1, 0x0000F3CF},
367 {REG_A5XX_RBBM_CLOCK_DELAY_SP0, 0x00000080},
368 {REG_A5XX_RBBM_CLOCK_DELAY_SP1, 0x00000080},
369 {REG_A5XX_RBBM_CLOCK_CNTL_TP0, 0x22222222},
370 {REG_A5XX_RBBM_CLOCK_CNTL_TP1, 0x22222222},
371 {REG_A5XX_RBBM_CLOCK_CNTL2_TP0, 0x22222222},
372 {REG_A5XX_RBBM_CLOCK_CNTL2_TP1, 0x22222222},
373 {REG_A5XX_RBBM_CLOCK_CNTL3_TP0, 0x00002222},
374 {REG_A5XX_RBBM_CLOCK_CNTL3_TP1, 0x00002222},
375 {REG_A5XX_RBBM_CLOCK_HYST_TP0, 0x77777777},
376 {REG_A5XX_RBBM_CLOCK_HYST_TP1, 0x77777777},
377 {REG_A5XX_RBBM_CLOCK_HYST2_TP0, 0x77777777},
378 {REG_A5XX_RBBM_CLOCK_HYST2_TP1, 0x77777777},
379 {REG_A5XX_RBBM_CLOCK_HYST3_TP0, 0x00007777},
380 {REG_A5XX_RBBM_CLOCK_HYST3_TP1, 0x00007777},
381 {REG_A5XX_RBBM_CLOCK_DELAY_TP0, 0x11111111},
382 {REG_A5XX_RBBM_CLOCK_DELAY_TP1, 0x11111111},
383 {REG_A5XX_RBBM_CLOCK_DELAY2_TP0, 0x11111111},
384 {REG_A5XX_RBBM_CLOCK_DELAY2_TP1, 0x11111111},
385 {REG_A5XX_RBBM_CLOCK_DELAY3_TP0, 0x00001111},
386 {REG_A5XX_RBBM_CLOCK_DELAY3_TP1, 0x00001111},
387 {REG_A5XX_RBBM_CLOCK_CNTL_UCHE, 0x22222222},
388 {REG_A5XX_RBBM_CLOCK_CNTL2_UCHE, 0x22222222},
389 {REG_A5XX_RBBM_CLOCK_CNTL3_UCHE, 0x22222222},
390 {REG_A5XX_RBBM_CLOCK_CNTL4_UCHE, 0x00222222},
391 {REG_A5XX_RBBM_CLOCK_HYST_UCHE, 0x00444444},
392 {REG_A5XX_RBBM_CLOCK_DELAY_UCHE, 0x00000002},
393 {REG_A5XX_RBBM_CLOCK_CNTL_RB0, 0x22222222},
394 {REG_A5XX_RBBM_CLOCK_CNTL_RB1, 0x22222222},
395 {REG_A5XX_RBBM_CLOCK_CNTL2_RB0, 0x00222222},
396 {REG_A5XX_RBBM_CLOCK_CNTL2_RB1, 0x00222222},
397 {REG_A5XX_RBBM_CLOCK_CNTL_CCU0, 0x00022220},
398 {REG_A5XX_RBBM_CLOCK_CNTL_CCU1, 0x00022220},
399 {REG_A5XX_RBBM_CLOCK_CNTL_RAC, 0x05522222},
400 {REG_A5XX_RBBM_CLOCK_CNTL2_RAC, 0x00505555},
401 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU0, 0x04040404},
402 {REG_A5XX_RBBM_CLOCK_HYST_RB_CCU1, 0x04040404},
403 {REG_A5XX_RBBM_CLOCK_HYST_RAC, 0x07444044},
404 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_0, 0x00000002},
405 {REG_A5XX_RBBM_CLOCK_DELAY_RB_CCU_L1_1, 0x00000002},
406 {REG_A5XX_RBBM_CLOCK_DELAY_RAC, 0x00010011},
407 {REG_A5XX_RBBM_CLOCK_CNTL_TSE_RAS_RBBM, 0x04222222},
408 {REG_A5XX_RBBM_CLOCK_MODE_GPC, 0x02222222},
409 {REG_A5XX_RBBM_CLOCK_MODE_VFD, 0x00002222},
410 {REG_A5XX_RBBM_CLOCK_HYST_TSE_RAS_RBBM, 0x00000000},
411 {REG_A5XX_RBBM_CLOCK_HYST_GPC, 0x04104004},
412 {REG_A5XX_RBBM_CLOCK_HYST_VFD, 0x00000000},
413 {REG_A5XX_RBBM_CLOCK_DELAY_HLSQ, 0x00000000},
414 {REG_A5XX_RBBM_CLOCK_DELAY_TSE_RAS_RBBM, 0x00004000},
415 {REG_A5XX_RBBM_CLOCK_DELAY_GPC, 0x00000200},
416 {REG_A5XX_RBBM_CLOCK_DELAY_VFD, 0x00002222},
419 void a5xx_set_hwcg(struct msm_gpu *gpu, bool state)
421 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
422 const struct adreno_five_hwcg_regs *regs;
425 if (adreno_is_a508(adreno_gpu)) {
427 sz = ARRAY_SIZE(a50x_hwcg);
428 } else if (adreno_is_a509(adreno_gpu) || adreno_is_a512(adreno_gpu)) {
430 sz = ARRAY_SIZE(a512_hwcg);
433 sz = ARRAY_SIZE(a5xx_hwcg);
436 for (i = 0; i < sz; i++)
437 gpu_write(gpu, regs[i].offset,
438 state ? regs[i].value : 0);
440 if (adreno_is_a540(adreno_gpu)) {
441 gpu_write(gpu, REG_A5XX_RBBM_CLOCK_DELAY_GPMU, state ? 0x00000770 : 0);
442 gpu_write(gpu, REG_A5XX_RBBM_CLOCK_HYST_GPMU, state ? 0x00000004 : 0);
445 gpu_write(gpu, REG_A5XX_RBBM_CLOCK_CNTL, state ? 0xAAA8AA00 : 0);
446 gpu_write(gpu, REG_A5XX_RBBM_ISDB_CNT, state ? 0x182 : 0x180);
449 static int a5xx_me_init(struct msm_gpu *gpu)
451 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
452 struct msm_ringbuffer *ring = gpu->rb[0];
454 OUT_PKT7(ring, CP_ME_INIT, 8);
456 OUT_RING(ring, 0x0000002F);
458 /* Enable multiple hardware contexts */
459 OUT_RING(ring, 0x00000003);
461 /* Enable error detection */
462 OUT_RING(ring, 0x20000000);
464 /* Don't enable header dump */
465 OUT_RING(ring, 0x00000000);
466 OUT_RING(ring, 0x00000000);
468 /* Specify workarounds for various microcode issues */
469 if (adreno_is_a530(adreno_gpu)) {
470 /* Workaround for token end syncs
471 * Force a WFI after every direct-render 3D mode draw and every
474 OUT_RING(ring, 0x0000000B);
475 } else if (adreno_is_a510(adreno_gpu)) {
476 /* Workaround for token and syncs */
477 OUT_RING(ring, 0x00000001);
479 /* No workarounds enabled */
480 OUT_RING(ring, 0x00000000);
483 OUT_RING(ring, 0x00000000);
484 OUT_RING(ring, 0x00000000);
486 a5xx_flush(gpu, ring, true);
487 return a5xx_idle(gpu, ring) ? 0 : -EINVAL;
490 static int a5xx_preempt_start(struct msm_gpu *gpu)
492 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
493 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
494 struct msm_ringbuffer *ring = gpu->rb[0];
496 if (gpu->nr_rings == 1)
499 /* Turn off protected mode to write to special registers */
500 OUT_PKT7(ring, CP_SET_PROTECTED_MODE, 1);
503 /* Set the save preemption record for the ring/command */
504 OUT_PKT4(ring, REG_A5XX_CP_CONTEXT_SWITCH_SAVE_ADDR_LO, 2);
505 OUT_RING(ring, lower_32_bits(a5xx_gpu->preempt_iova[ring->id]));
506 OUT_RING(ring, upper_32_bits(a5xx_gpu->preempt_iova[ring->id]));
508 /* Turn back on protected mode */
509 OUT_PKT7(ring, CP_SET_PROTECTED_MODE, 1);
512 OUT_PKT7(ring, CP_PREEMPT_ENABLE_GLOBAL, 1);
513 OUT_RING(ring, 0x00);
515 OUT_PKT7(ring, CP_PREEMPT_ENABLE_LOCAL, 1);
516 OUT_RING(ring, 0x01);
518 OUT_PKT7(ring, CP_YIELD_ENABLE, 1);
519 OUT_RING(ring, 0x01);
521 /* Yield the floor on command completion */
522 OUT_PKT7(ring, CP_CONTEXT_SWITCH_YIELD, 4);
523 OUT_RING(ring, 0x00);
524 OUT_RING(ring, 0x00);
525 OUT_RING(ring, 0x01);
526 OUT_RING(ring, 0x01);
528 /* The WHERE_AMI_I packet is not needed after a YIELD is issued */
529 a5xx_flush(gpu, ring, false);
531 return a5xx_idle(gpu, ring) ? 0 : -EINVAL;
534 static void a5xx_ucode_check_version(struct a5xx_gpu *a5xx_gpu,
535 struct drm_gem_object *obj)
537 u32 *buf = msm_gem_get_vaddr(obj);
543 * If the lowest nibble is 0xa that is an indication that this microcode
544 * has been patched. The actual version is in dword [3] but we only care
545 * about the patchlevel which is the lowest nibble of dword [3]
547 if (((buf[0] & 0xf) == 0xa) && (buf[2] & 0xf) >= 1)
548 a5xx_gpu->has_whereami = true;
550 msm_gem_put_vaddr(obj);
553 static int a5xx_ucode_init(struct msm_gpu *gpu)
555 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
556 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
559 if (!a5xx_gpu->pm4_bo) {
560 a5xx_gpu->pm4_bo = adreno_fw_create_bo(gpu,
561 adreno_gpu->fw[ADRENO_FW_PM4], &a5xx_gpu->pm4_iova);
564 if (IS_ERR(a5xx_gpu->pm4_bo)) {
565 ret = PTR_ERR(a5xx_gpu->pm4_bo);
566 a5xx_gpu->pm4_bo = NULL;
567 DRM_DEV_ERROR(gpu->dev->dev, "could not allocate PM4: %d\n",
572 msm_gem_object_set_name(a5xx_gpu->pm4_bo, "pm4fw");
575 if (!a5xx_gpu->pfp_bo) {
576 a5xx_gpu->pfp_bo = adreno_fw_create_bo(gpu,
577 adreno_gpu->fw[ADRENO_FW_PFP], &a5xx_gpu->pfp_iova);
579 if (IS_ERR(a5xx_gpu->pfp_bo)) {
580 ret = PTR_ERR(a5xx_gpu->pfp_bo);
581 a5xx_gpu->pfp_bo = NULL;
582 DRM_DEV_ERROR(gpu->dev->dev, "could not allocate PFP: %d\n",
587 msm_gem_object_set_name(a5xx_gpu->pfp_bo, "pfpfw");
588 a5xx_ucode_check_version(a5xx_gpu, a5xx_gpu->pfp_bo);
591 gpu_write64(gpu, REG_A5XX_CP_ME_INSTR_BASE_LO,
592 REG_A5XX_CP_ME_INSTR_BASE_HI, a5xx_gpu->pm4_iova);
594 gpu_write64(gpu, REG_A5XX_CP_PFP_INSTR_BASE_LO,
595 REG_A5XX_CP_PFP_INSTR_BASE_HI, a5xx_gpu->pfp_iova);
600 #define SCM_GPU_ZAP_SHADER_RESUME 0
602 static int a5xx_zap_shader_resume(struct msm_gpu *gpu)
606 ret = qcom_scm_set_remote_state(SCM_GPU_ZAP_SHADER_RESUME, GPU_PAS_ID);
608 DRM_ERROR("%s: zap-shader resume failed: %d\n",
614 static int a5xx_zap_shader_init(struct msm_gpu *gpu)
620 * If the zap shader is already loaded into memory we just need to kick
621 * the remote processor to reinitialize it
624 return a5xx_zap_shader_resume(gpu);
626 ret = adreno_zap_shader_load(gpu, GPU_PAS_ID);
632 #define A5XX_INT_MASK (A5XX_RBBM_INT_0_MASK_RBBM_AHB_ERROR | \
633 A5XX_RBBM_INT_0_MASK_RBBM_TRANSFER_TIMEOUT | \
634 A5XX_RBBM_INT_0_MASK_RBBM_ME_MS_TIMEOUT | \
635 A5XX_RBBM_INT_0_MASK_RBBM_PFP_MS_TIMEOUT | \
636 A5XX_RBBM_INT_0_MASK_RBBM_ETS_MS_TIMEOUT | \
637 A5XX_RBBM_INT_0_MASK_RBBM_ATB_ASYNC_OVERFLOW | \
638 A5XX_RBBM_INT_0_MASK_CP_HW_ERROR | \
639 A5XX_RBBM_INT_0_MASK_MISC_HANG_DETECT | \
640 A5XX_RBBM_INT_0_MASK_CP_SW | \
641 A5XX_RBBM_INT_0_MASK_CP_CACHE_FLUSH_TS | \
642 A5XX_RBBM_INT_0_MASK_UCHE_OOB_ACCESS | \
643 A5XX_RBBM_INT_0_MASK_GPMU_VOLTAGE_DROOP)
645 static int a5xx_hw_init(struct msm_gpu *gpu)
647 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
648 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
652 gpu_write(gpu, REG_A5XX_VBIF_ROUND_ROBIN_QOS_ARB, 0x00000003);
654 if (adreno_is_a509(adreno_gpu) || adreno_is_a512(adreno_gpu) ||
655 adreno_is_a540(adreno_gpu))
656 gpu_write(gpu, REG_A5XX_VBIF_GATE_OFF_WRREQ_EN, 0x00000009);
658 /* Make all blocks contribute to the GPU BUSY perf counter */
659 gpu_write(gpu, REG_A5XX_RBBM_PERFCTR_GPU_BUSY_MASKED, 0xFFFFFFFF);
661 /* Enable RBBM error reporting bits */
662 gpu_write(gpu, REG_A5XX_RBBM_AHB_CNTL0, 0x00000001);
664 if (adreno_gpu->info->quirks & ADRENO_QUIRK_FAULT_DETECT_MASK) {
666 * Mask out the activity signals from RB1-3 to avoid false
670 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL11,
672 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL12,
674 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL13,
676 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL14,
678 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL15,
680 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL16,
682 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL17,
684 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_MASK_CNTL18,
688 /* Enable fault detection */
689 gpu_write(gpu, REG_A5XX_RBBM_INTERFACE_HANG_INT_CNTL,
692 /* Turn on performance counters */
693 gpu_write(gpu, REG_A5XX_RBBM_PERFCTR_CNTL, 0x01);
695 /* Select CP0 to always count cycles */
696 gpu_write(gpu, REG_A5XX_CP_PERFCTR_CP_SEL_0, PERF_CP_ALWAYS_COUNT);
698 /* Select RBBM0 to countable 6 to get the busy status for devfreq */
699 gpu_write(gpu, REG_A5XX_RBBM_PERFCTR_RBBM_SEL_0, 6);
701 /* Increase VFD cache access so LRZ and other data gets evicted less */
702 gpu_write(gpu, REG_A5XX_UCHE_CACHE_WAYS, 0x02);
704 /* Disable L2 bypass in the UCHE */
705 gpu_write(gpu, REG_A5XX_UCHE_TRAP_BASE_LO, 0xFFFF0000);
706 gpu_write(gpu, REG_A5XX_UCHE_TRAP_BASE_HI, 0x0001FFFF);
707 gpu_write(gpu, REG_A5XX_UCHE_WRITE_THRU_BASE_LO, 0xFFFF0000);
708 gpu_write(gpu, REG_A5XX_UCHE_WRITE_THRU_BASE_HI, 0x0001FFFF);
710 /* Set the GMEM VA range (0 to gpu->gmem) */
711 gpu_write(gpu, REG_A5XX_UCHE_GMEM_RANGE_MIN_LO, 0x00100000);
712 gpu_write(gpu, REG_A5XX_UCHE_GMEM_RANGE_MIN_HI, 0x00000000);
713 gpu_write(gpu, REG_A5XX_UCHE_GMEM_RANGE_MAX_LO,
714 0x00100000 + adreno_gpu->gmem - 1);
715 gpu_write(gpu, REG_A5XX_UCHE_GMEM_RANGE_MAX_HI, 0x00000000);
717 if (adreno_is_a508(adreno_gpu) || adreno_is_a510(adreno_gpu)) {
718 gpu_write(gpu, REG_A5XX_CP_MEQ_THRESHOLDS, 0x20);
719 if (adreno_is_a508(adreno_gpu))
720 gpu_write(gpu, REG_A5XX_CP_MERCIU_SIZE, 0x400);
722 gpu_write(gpu, REG_A5XX_CP_MERCIU_SIZE, 0x20);
723 gpu_write(gpu, REG_A5XX_CP_ROQ_THRESHOLDS_2, 0x40000030);
724 gpu_write(gpu, REG_A5XX_CP_ROQ_THRESHOLDS_1, 0x20100D0A);
726 gpu_write(gpu, REG_A5XX_CP_MEQ_THRESHOLDS, 0x40);
727 if (adreno_is_a530(adreno_gpu))
728 gpu_write(gpu, REG_A5XX_CP_MERCIU_SIZE, 0x40);
730 gpu_write(gpu, REG_A5XX_CP_MERCIU_SIZE, 0x400);
731 gpu_write(gpu, REG_A5XX_CP_ROQ_THRESHOLDS_2, 0x80000060);
732 gpu_write(gpu, REG_A5XX_CP_ROQ_THRESHOLDS_1, 0x40201B16);
735 if (adreno_is_a508(adreno_gpu))
736 gpu_write(gpu, REG_A5XX_PC_DBG_ECO_CNTL,
737 (0x100 << 11 | 0x100 << 22));
738 else if (adreno_is_a509(adreno_gpu) || adreno_is_a510(adreno_gpu) ||
739 adreno_is_a512(adreno_gpu))
740 gpu_write(gpu, REG_A5XX_PC_DBG_ECO_CNTL,
741 (0x200 << 11 | 0x200 << 22));
743 gpu_write(gpu, REG_A5XX_PC_DBG_ECO_CNTL,
744 (0x400 << 11 | 0x300 << 22));
746 if (adreno_gpu->info->quirks & ADRENO_QUIRK_TWO_PASS_USE_WFI)
747 gpu_rmw(gpu, REG_A5XX_PC_DBG_ECO_CNTL, 0, (1 << 8));
750 * Disable the RB sampler datapath DP2 clock gating optimization
751 * for 1-SP GPUs, as it is enabled by default.
753 if (adreno_is_a508(adreno_gpu) || adreno_is_a509(adreno_gpu) ||
754 adreno_is_a512(adreno_gpu))
755 gpu_rmw(gpu, REG_A5XX_RB_DBG_ECO_CNTL, 0, (1 << 9));
757 /* Disable UCHE global filter as SP can invalidate/flush independently */
758 gpu_write(gpu, REG_A5XX_UCHE_MODE_CNTL, BIT(29));
760 /* Enable USE_RETENTION_FLOPS */
761 gpu_write(gpu, REG_A5XX_CP_CHICKEN_DBG, 0x02000000);
763 /* Enable ME/PFP split notification */
764 gpu_write(gpu, REG_A5XX_RBBM_AHB_CNTL1, 0xA6FFFFFF);
767 * In A5x, CCU can send context_done event of a particular context to
768 * UCHE which ultimately reaches CP even when there is valid
769 * transaction of that context inside CCU. This can let CP to program
770 * config registers, which will make the "valid transaction" inside
771 * CCU to be interpreted differently. This can cause gpu fault. This
772 * bug is fixed in latest A510 revision. To enable this bug fix -
773 * bit[11] of RB_DBG_ECO_CNTL need to be set to 0, default is 1
774 * (disable). For older A510 version this bit is unused.
776 if (adreno_is_a510(adreno_gpu))
777 gpu_rmw(gpu, REG_A5XX_RB_DBG_ECO_CNTL, (1 << 11), 0);
780 a5xx_set_hwcg(gpu, true);
782 gpu_write(gpu, REG_A5XX_RBBM_AHB_CNTL2, 0x0000003F);
784 /* Set the highest bank bit */
785 if (adreno_is_a540(adreno_gpu))
790 gpu_write(gpu, REG_A5XX_TPL1_MODE_CNTL, regbit << 7);
791 gpu_write(gpu, REG_A5XX_RB_MODE_CNTL, regbit << 1);
793 if (adreno_is_a509(adreno_gpu) || adreno_is_a512(adreno_gpu) ||
794 adreno_is_a540(adreno_gpu))
795 gpu_write(gpu, REG_A5XX_UCHE_DBG_ECO_CNTL_2, regbit);
797 /* Disable All flat shading optimization (ALLFLATOPTDIS) */
798 gpu_rmw(gpu, REG_A5XX_VPC_DBG_ECO_CNTL, 0, (1 << 10));
800 /* Protect registers from the CP */
801 gpu_write(gpu, REG_A5XX_CP_PROTECT_CNTL, 0x00000007);
804 gpu_write(gpu, REG_A5XX_CP_PROTECT(0), ADRENO_PROTECT_RW(0x04, 4));
805 gpu_write(gpu, REG_A5XX_CP_PROTECT(1), ADRENO_PROTECT_RW(0x08, 8));
806 gpu_write(gpu, REG_A5XX_CP_PROTECT(2), ADRENO_PROTECT_RW(0x10, 16));
807 gpu_write(gpu, REG_A5XX_CP_PROTECT(3), ADRENO_PROTECT_RW(0x20, 32));
808 gpu_write(gpu, REG_A5XX_CP_PROTECT(4), ADRENO_PROTECT_RW(0x40, 64));
809 gpu_write(gpu, REG_A5XX_CP_PROTECT(5), ADRENO_PROTECT_RW(0x80, 64));
811 /* Content protect */
812 gpu_write(gpu, REG_A5XX_CP_PROTECT(6),
813 ADRENO_PROTECT_RW(REG_A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO,
815 gpu_write(gpu, REG_A5XX_CP_PROTECT(7),
816 ADRENO_PROTECT_RW(REG_A5XX_RBBM_SECVID_TRUST_CNTL, 2));
819 gpu_write(gpu, REG_A5XX_CP_PROTECT(8), ADRENO_PROTECT_RW(0x800, 64));
820 gpu_write(gpu, REG_A5XX_CP_PROTECT(9), ADRENO_PROTECT_RW(0x840, 8));
821 gpu_write(gpu, REG_A5XX_CP_PROTECT(10), ADRENO_PROTECT_RW(0x880, 32));
822 gpu_write(gpu, REG_A5XX_CP_PROTECT(11), ADRENO_PROTECT_RW(0xAA0, 1));
825 gpu_write(gpu, REG_A5XX_CP_PROTECT(12), ADRENO_PROTECT_RW(0xCC0, 1));
826 gpu_write(gpu, REG_A5XX_CP_PROTECT(13), ADRENO_PROTECT_RW(0xCF0, 2));
829 gpu_write(gpu, REG_A5XX_CP_PROTECT(14), ADRENO_PROTECT_RW(0xE68, 8));
830 gpu_write(gpu, REG_A5XX_CP_PROTECT(15), ADRENO_PROTECT_RW(0xE70, 16));
833 gpu_write(gpu, REG_A5XX_CP_PROTECT(16), ADRENO_PROTECT_RW(0xE80, 16));
835 if (adreno_is_a508(adreno_gpu) || adreno_is_a509(adreno_gpu) ||
836 adreno_is_a510(adreno_gpu) || adreno_is_a512(adreno_gpu) ||
837 adreno_is_a530(adreno_gpu))
838 gpu_write(gpu, REG_A5XX_CP_PROTECT(17),
839 ADRENO_PROTECT_RW(0x10000, 0x8000));
841 gpu_write(gpu, REG_A5XX_RBBM_SECVID_TSB_CNTL, 0);
843 * Disable the trusted memory range - we don't actually supported secure
844 * memory rendering at this point in time and we don't want to block off
845 * part of the virtual memory space.
847 gpu_write64(gpu, REG_A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_LO,
848 REG_A5XX_RBBM_SECVID_TSB_TRUSTED_BASE_HI, 0x00000000);
849 gpu_write(gpu, REG_A5XX_RBBM_SECVID_TSB_TRUSTED_SIZE, 0x00000000);
851 /* Put the GPU into 64 bit by default */
852 gpu_write(gpu, REG_A5XX_CP_ADDR_MODE_CNTL, 0x1);
853 gpu_write(gpu, REG_A5XX_VSC_ADDR_MODE_CNTL, 0x1);
854 gpu_write(gpu, REG_A5XX_GRAS_ADDR_MODE_CNTL, 0x1);
855 gpu_write(gpu, REG_A5XX_RB_ADDR_MODE_CNTL, 0x1);
856 gpu_write(gpu, REG_A5XX_PC_ADDR_MODE_CNTL, 0x1);
857 gpu_write(gpu, REG_A5XX_HLSQ_ADDR_MODE_CNTL, 0x1);
858 gpu_write(gpu, REG_A5XX_VFD_ADDR_MODE_CNTL, 0x1);
859 gpu_write(gpu, REG_A5XX_VPC_ADDR_MODE_CNTL, 0x1);
860 gpu_write(gpu, REG_A5XX_UCHE_ADDR_MODE_CNTL, 0x1);
861 gpu_write(gpu, REG_A5XX_SP_ADDR_MODE_CNTL, 0x1);
862 gpu_write(gpu, REG_A5XX_TPL1_ADDR_MODE_CNTL, 0x1);
863 gpu_write(gpu, REG_A5XX_RBBM_SECVID_TSB_ADDR_MODE_CNTL, 0x1);
866 * VPC corner case with local memory load kill leads to corrupt
867 * internal state. Normal Disable does not work for all a5x chips.
868 * So do the following setting to disable it.
870 if (adreno_gpu->info->quirks & ADRENO_QUIRK_LMLOADKILL_DISABLE) {
871 gpu_rmw(gpu, REG_A5XX_VPC_DBG_ECO_CNTL, 0, BIT(23));
872 gpu_rmw(gpu, REG_A5XX_HLSQ_DBG_ECO_CNTL, BIT(18), 0);
875 ret = adreno_hw_init(gpu);
879 if (!(adreno_is_a508(adreno_gpu) || adreno_is_a509(adreno_gpu) ||
880 adreno_is_a510(adreno_gpu) || adreno_is_a512(adreno_gpu)))
881 a5xx_gpmu_ucode_init(gpu);
883 ret = a5xx_ucode_init(gpu);
887 /* Set the ringbuffer address */
888 gpu_write64(gpu, REG_A5XX_CP_RB_BASE, REG_A5XX_CP_RB_BASE_HI,
892 * If the microcode supports the WHERE_AM_I opcode then we can use that
893 * in lieu of the RPTR shadow and enable preemption. Otherwise, we
894 * can't safely use the RPTR shadow or preemption. In either case, the
895 * RPTR shadow should be disabled in hardware.
897 gpu_write(gpu, REG_A5XX_CP_RB_CNTL,
898 MSM_GPU_RB_CNTL_DEFAULT | AXXX_CP_RB_CNTL_NO_UPDATE);
900 /* Create a privileged buffer for the RPTR shadow */
901 if (a5xx_gpu->has_whereami) {
902 if (!a5xx_gpu->shadow_bo) {
903 a5xx_gpu->shadow = msm_gem_kernel_new(gpu->dev,
904 sizeof(u32) * gpu->nr_rings,
905 MSM_BO_WC | MSM_BO_MAP_PRIV,
906 gpu->aspace, &a5xx_gpu->shadow_bo,
907 &a5xx_gpu->shadow_iova);
909 if (IS_ERR(a5xx_gpu->shadow))
910 return PTR_ERR(a5xx_gpu->shadow);
913 gpu_write64(gpu, REG_A5XX_CP_RB_RPTR_ADDR,
914 REG_A5XX_CP_RB_RPTR_ADDR_HI, shadowptr(a5xx_gpu, gpu->rb[0]));
915 } else if (gpu->nr_rings > 1) {
916 /* Disable preemption if WHERE_AM_I isn't available */
917 a5xx_preempt_fini(gpu);
921 a5xx_preempt_hw_init(gpu);
923 /* Disable the interrupts through the initial bringup stage */
924 gpu_write(gpu, REG_A5XX_RBBM_INT_0_MASK, A5XX_INT_MASK);
926 /* Clear ME_HALT to start the micro engine */
927 gpu_write(gpu, REG_A5XX_CP_PFP_ME_CNTL, 0);
928 ret = a5xx_me_init(gpu);
932 ret = a5xx_power_init(gpu);
937 * Send a pipeline event stat to get misbehaving counters to start
940 if (adreno_is_a530(adreno_gpu)) {
941 OUT_PKT7(gpu->rb[0], CP_EVENT_WRITE, 1);
942 OUT_RING(gpu->rb[0], CP_EVENT_WRITE_0_EVENT(STAT_EVENT));
944 a5xx_flush(gpu, gpu->rb[0], true);
945 if (!a5xx_idle(gpu, gpu->rb[0]))
950 * If the chip that we are using does support loading one, then
951 * try to load a zap shader into the secure world. If successful
952 * we can use the CP to switch out of secure mode. If not then we
953 * have no resource but to try to switch ourselves out manually. If we
954 * guessed wrong then access to the RBBM_SECVID_TRUST_CNTL register will
955 * be blocked and a permissions violation will soon follow.
957 ret = a5xx_zap_shader_init(gpu);
959 OUT_PKT7(gpu->rb[0], CP_SET_SECURE_MODE, 1);
960 OUT_RING(gpu->rb[0], 0x00000000);
962 a5xx_flush(gpu, gpu->rb[0], true);
963 if (!a5xx_idle(gpu, gpu->rb[0]))
965 } else if (ret == -ENODEV) {
967 * This device does not use zap shader (but print a warning
968 * just in case someone got their dt wrong.. hopefully they
969 * have a debug UART to realize the error of their ways...
970 * if you mess this up you are about to crash horribly)
972 dev_warn_once(gpu->dev->dev,
973 "Zap shader not enabled - using SECVID_TRUST_CNTL instead\n");
974 gpu_write(gpu, REG_A5XX_RBBM_SECVID_TRUST_CNTL, 0x0);
979 /* Last step - yield the ringbuffer */
980 a5xx_preempt_start(gpu);
985 static void a5xx_recover(struct msm_gpu *gpu)
989 adreno_dump_info(gpu);
991 for (i = 0; i < 8; i++) {
992 printk("CP_SCRATCH_REG%d: %u\n", i,
993 gpu_read(gpu, REG_A5XX_CP_SCRATCH_REG(i)));
999 gpu_write(gpu, REG_A5XX_RBBM_SW_RESET_CMD, 1);
1000 gpu_read(gpu, REG_A5XX_RBBM_SW_RESET_CMD);
1001 gpu_write(gpu, REG_A5XX_RBBM_SW_RESET_CMD, 0);
1002 adreno_recover(gpu);
1005 static void a5xx_destroy(struct msm_gpu *gpu)
1007 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
1008 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
1010 DBG("%s", gpu->name);
1012 a5xx_preempt_fini(gpu);
1014 if (a5xx_gpu->pm4_bo) {
1015 msm_gem_unpin_iova(a5xx_gpu->pm4_bo, gpu->aspace);
1016 drm_gem_object_put(a5xx_gpu->pm4_bo);
1019 if (a5xx_gpu->pfp_bo) {
1020 msm_gem_unpin_iova(a5xx_gpu->pfp_bo, gpu->aspace);
1021 drm_gem_object_put(a5xx_gpu->pfp_bo);
1024 if (a5xx_gpu->gpmu_bo) {
1025 msm_gem_unpin_iova(a5xx_gpu->gpmu_bo, gpu->aspace);
1026 drm_gem_object_put(a5xx_gpu->gpmu_bo);
1029 if (a5xx_gpu->shadow_bo) {
1030 msm_gem_unpin_iova(a5xx_gpu->shadow_bo, gpu->aspace);
1031 drm_gem_object_put(a5xx_gpu->shadow_bo);
1034 adreno_gpu_cleanup(adreno_gpu);
1038 static inline bool _a5xx_check_idle(struct msm_gpu *gpu)
1040 if (gpu_read(gpu, REG_A5XX_RBBM_STATUS) & ~A5XX_RBBM_STATUS_HI_BUSY)
1044 * Nearly every abnormality ends up pausing the GPU and triggering a
1045 * fault so we can safely just watch for this one interrupt to fire
1047 return !(gpu_read(gpu, REG_A5XX_RBBM_INT_0_STATUS) &
1048 A5XX_RBBM_INT_0_MASK_MISC_HANG_DETECT);
1051 bool a5xx_idle(struct msm_gpu *gpu, struct msm_ringbuffer *ring)
1053 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
1054 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
1056 if (ring != a5xx_gpu->cur_ring) {
1057 WARN(1, "Tried to idle a non-current ringbuffer\n");
1061 /* wait for CP to drain ringbuffer: */
1062 if (!adreno_idle(gpu, ring))
1065 if (spin_until(_a5xx_check_idle(gpu))) {
1066 DRM_ERROR("%s: %ps: timeout waiting for GPU to idle: status %8.8X irq %8.8X rptr/wptr %d/%d\n",
1067 gpu->name, __builtin_return_address(0),
1068 gpu_read(gpu, REG_A5XX_RBBM_STATUS),
1069 gpu_read(gpu, REG_A5XX_RBBM_INT_0_STATUS),
1070 gpu_read(gpu, REG_A5XX_CP_RB_RPTR),
1071 gpu_read(gpu, REG_A5XX_CP_RB_WPTR));
1078 static int a5xx_fault_handler(void *arg, unsigned long iova, int flags, void *data)
1080 struct msm_gpu *gpu = arg;
1081 pr_warn_ratelimited("*** gpu fault: iova=%08lx, flags=%d (%u,%u,%u,%u)\n",
1083 gpu_read(gpu, REG_A5XX_CP_SCRATCH_REG(4)),
1084 gpu_read(gpu, REG_A5XX_CP_SCRATCH_REG(5)),
1085 gpu_read(gpu, REG_A5XX_CP_SCRATCH_REG(6)),
1086 gpu_read(gpu, REG_A5XX_CP_SCRATCH_REG(7)));
1091 static void a5xx_cp_err_irq(struct msm_gpu *gpu)
1093 u32 status = gpu_read(gpu, REG_A5XX_CP_INTERRUPT_STATUS);
1095 if (status & A5XX_CP_INT_CP_OPCODE_ERROR) {
1098 gpu_write(gpu, REG_A5XX_CP_PFP_STAT_ADDR, 0);
1101 * REG_A5XX_CP_PFP_STAT_DATA is indexed, and we want index 1 so
1105 gpu_read(gpu, REG_A5XX_CP_PFP_STAT_DATA);
1106 val = gpu_read(gpu, REG_A5XX_CP_PFP_STAT_DATA);
1108 dev_err_ratelimited(gpu->dev->dev, "CP | opcode error | possible opcode=0x%8.8X\n",
1112 if (status & A5XX_CP_INT_CP_HW_FAULT_ERROR)
1113 dev_err_ratelimited(gpu->dev->dev, "CP | HW fault | status=0x%8.8X\n",
1114 gpu_read(gpu, REG_A5XX_CP_HW_FAULT));
1116 if (status & A5XX_CP_INT_CP_DMA_ERROR)
1117 dev_err_ratelimited(gpu->dev->dev, "CP | DMA error\n");
1119 if (status & A5XX_CP_INT_CP_REGISTER_PROTECTION_ERROR) {
1120 u32 val = gpu_read(gpu, REG_A5XX_CP_PROTECT_STATUS);
1122 dev_err_ratelimited(gpu->dev->dev,
1123 "CP | protected mode error | %s | addr=0x%8.8X | status=0x%8.8X\n",
1124 val & (1 << 24) ? "WRITE" : "READ",
1125 (val & 0xFFFFF) >> 2, val);
1128 if (status & A5XX_CP_INT_CP_AHB_ERROR) {
1129 u32 status = gpu_read(gpu, REG_A5XX_CP_AHB_FAULT);
1130 const char *access[16] = { "reserved", "reserved",
1131 "timestamp lo", "timestamp hi", "pfp read", "pfp write",
1132 "", "", "me read", "me write", "", "", "crashdump read",
1133 "crashdump write" };
1135 dev_err_ratelimited(gpu->dev->dev,
1136 "CP | AHB error | addr=%X access=%s error=%d | status=0x%8.8X\n",
1137 status & 0xFFFFF, access[(status >> 24) & 0xF],
1138 (status & (1 << 31)), status);
1142 static void a5xx_rbbm_err_irq(struct msm_gpu *gpu, u32 status)
1144 if (status & A5XX_RBBM_INT_0_MASK_RBBM_AHB_ERROR) {
1145 u32 val = gpu_read(gpu, REG_A5XX_RBBM_AHB_ERROR_STATUS);
1147 dev_err_ratelimited(gpu->dev->dev,
1148 "RBBM | AHB bus error | %s | addr=0x%X | ports=0x%X:0x%X\n",
1149 val & (1 << 28) ? "WRITE" : "READ",
1150 (val & 0xFFFFF) >> 2, (val >> 20) & 0x3,
1153 /* Clear the error */
1154 gpu_write(gpu, REG_A5XX_RBBM_AHB_CMD, (1 << 4));
1156 /* Clear the interrupt */
1157 gpu_write(gpu, REG_A5XX_RBBM_INT_CLEAR_CMD,
1158 A5XX_RBBM_INT_0_MASK_RBBM_AHB_ERROR);
1161 if (status & A5XX_RBBM_INT_0_MASK_RBBM_TRANSFER_TIMEOUT)
1162 dev_err_ratelimited(gpu->dev->dev, "RBBM | AHB transfer timeout\n");
1164 if (status & A5XX_RBBM_INT_0_MASK_RBBM_ME_MS_TIMEOUT)
1165 dev_err_ratelimited(gpu->dev->dev, "RBBM | ME master split | status=0x%X\n",
1166 gpu_read(gpu, REG_A5XX_RBBM_AHB_ME_SPLIT_STATUS));
1168 if (status & A5XX_RBBM_INT_0_MASK_RBBM_PFP_MS_TIMEOUT)
1169 dev_err_ratelimited(gpu->dev->dev, "RBBM | PFP master split | status=0x%X\n",
1170 gpu_read(gpu, REG_A5XX_RBBM_AHB_PFP_SPLIT_STATUS));
1172 if (status & A5XX_RBBM_INT_0_MASK_RBBM_ETS_MS_TIMEOUT)
1173 dev_err_ratelimited(gpu->dev->dev, "RBBM | ETS master split | status=0x%X\n",
1174 gpu_read(gpu, REG_A5XX_RBBM_AHB_ETS_SPLIT_STATUS));
1176 if (status & A5XX_RBBM_INT_0_MASK_RBBM_ATB_ASYNC_OVERFLOW)
1177 dev_err_ratelimited(gpu->dev->dev, "RBBM | ATB ASYNC overflow\n");
1179 if (status & A5XX_RBBM_INT_0_MASK_RBBM_ATB_BUS_OVERFLOW)
1180 dev_err_ratelimited(gpu->dev->dev, "RBBM | ATB bus overflow\n");
1183 static void a5xx_uche_err_irq(struct msm_gpu *gpu)
1185 uint64_t addr = (uint64_t) gpu_read(gpu, REG_A5XX_UCHE_TRAP_LOG_HI);
1187 addr |= gpu_read(gpu, REG_A5XX_UCHE_TRAP_LOG_LO);
1189 dev_err_ratelimited(gpu->dev->dev, "UCHE | Out of bounds access | addr=0x%llX\n",
1193 static void a5xx_gpmu_err_irq(struct msm_gpu *gpu)
1195 dev_err_ratelimited(gpu->dev->dev, "GPMU | voltage droop\n");
1198 static void a5xx_fault_detect_irq(struct msm_gpu *gpu)
1200 struct drm_device *dev = gpu->dev;
1201 struct msm_ringbuffer *ring = gpu->funcs->active_ring(gpu);
1204 * If stalled on SMMU fault, we could trip the GPU's hang detection,
1205 * but the fault handler will trigger the devcore dump, and we want
1206 * to otherwise resume normally rather than killing the submit, so
1209 if (gpu_read(gpu, REG_A5XX_RBBM_STATUS3) & BIT(24))
1212 DRM_DEV_ERROR(dev->dev, "gpu fault ring %d fence %x status %8.8X rb %4.4x/%4.4x ib1 %16.16llX/%4.4x ib2 %16.16llX/%4.4x\n",
1213 ring ? ring->id : -1, ring ? ring->seqno : 0,
1214 gpu_read(gpu, REG_A5XX_RBBM_STATUS),
1215 gpu_read(gpu, REG_A5XX_CP_RB_RPTR),
1216 gpu_read(gpu, REG_A5XX_CP_RB_WPTR),
1217 gpu_read64(gpu, REG_A5XX_CP_IB1_BASE, REG_A5XX_CP_IB1_BASE_HI),
1218 gpu_read(gpu, REG_A5XX_CP_IB1_BUFSZ),
1219 gpu_read64(gpu, REG_A5XX_CP_IB2_BASE, REG_A5XX_CP_IB2_BASE_HI),
1220 gpu_read(gpu, REG_A5XX_CP_IB2_BUFSZ));
1222 /* Turn off the hangcheck timer to keep it from bothering us */
1223 del_timer(&gpu->hangcheck_timer);
1225 kthread_queue_work(gpu->worker, &gpu->recover_work);
1228 #define RBBM_ERROR_MASK \
1229 (A5XX_RBBM_INT_0_MASK_RBBM_AHB_ERROR | \
1230 A5XX_RBBM_INT_0_MASK_RBBM_TRANSFER_TIMEOUT | \
1231 A5XX_RBBM_INT_0_MASK_RBBM_ME_MS_TIMEOUT | \
1232 A5XX_RBBM_INT_0_MASK_RBBM_PFP_MS_TIMEOUT | \
1233 A5XX_RBBM_INT_0_MASK_RBBM_ETS_MS_TIMEOUT | \
1234 A5XX_RBBM_INT_0_MASK_RBBM_ATB_ASYNC_OVERFLOW)
1236 static irqreturn_t a5xx_irq(struct msm_gpu *gpu)
1238 u32 status = gpu_read(gpu, REG_A5XX_RBBM_INT_0_STATUS);
1241 * Clear all the interrupts except RBBM_AHB_ERROR - if we clear it
1242 * before the source is cleared the interrupt will storm.
1244 gpu_write(gpu, REG_A5XX_RBBM_INT_CLEAR_CMD,
1245 status & ~A5XX_RBBM_INT_0_MASK_RBBM_AHB_ERROR);
1247 /* Pass status to a5xx_rbbm_err_irq because we've already cleared it */
1248 if (status & RBBM_ERROR_MASK)
1249 a5xx_rbbm_err_irq(gpu, status);
1251 if (status & A5XX_RBBM_INT_0_MASK_CP_HW_ERROR)
1252 a5xx_cp_err_irq(gpu);
1254 if (status & A5XX_RBBM_INT_0_MASK_MISC_HANG_DETECT)
1255 a5xx_fault_detect_irq(gpu);
1257 if (status & A5XX_RBBM_INT_0_MASK_UCHE_OOB_ACCESS)
1258 a5xx_uche_err_irq(gpu);
1260 if (status & A5XX_RBBM_INT_0_MASK_GPMU_VOLTAGE_DROOP)
1261 a5xx_gpmu_err_irq(gpu);
1263 if (status & A5XX_RBBM_INT_0_MASK_CP_CACHE_FLUSH_TS) {
1264 a5xx_preempt_trigger(gpu);
1265 msm_gpu_retire(gpu);
1268 if (status & A5XX_RBBM_INT_0_MASK_CP_SW)
1269 a5xx_preempt_irq(gpu);
1274 static const u32 a5xx_registers[] = {
1275 0x0000, 0x0002, 0x0004, 0x0020, 0x0022, 0x0026, 0x0029, 0x002B,
1276 0x002E, 0x0035, 0x0038, 0x0042, 0x0044, 0x0044, 0x0047, 0x0095,
1277 0x0097, 0x00BB, 0x03A0, 0x0464, 0x0469, 0x046F, 0x04D2, 0x04D3,
1278 0x04E0, 0x0533, 0x0540, 0x0555, 0x0800, 0x081A, 0x081F, 0x0841,
1279 0x0860, 0x0860, 0x0880, 0x08A0, 0x0B00, 0x0B12, 0x0B15, 0x0B28,
1280 0x0B78, 0x0B7F, 0x0BB0, 0x0BBD, 0x0BC0, 0x0BC6, 0x0BD0, 0x0C53,
1281 0x0C60, 0x0C61, 0x0C80, 0x0C82, 0x0C84, 0x0C85, 0x0C90, 0x0C98,
1282 0x0CA0, 0x0CA0, 0x0CB0, 0x0CB2, 0x2180, 0x2185, 0x2580, 0x2585,
1283 0x0CC1, 0x0CC1, 0x0CC4, 0x0CC7, 0x0CCC, 0x0CCC, 0x0CD0, 0x0CD8,
1284 0x0CE0, 0x0CE5, 0x0CE8, 0x0CE8, 0x0CEC, 0x0CF1, 0x0CFB, 0x0D0E,
1285 0x2100, 0x211E, 0x2140, 0x2145, 0x2500, 0x251E, 0x2540, 0x2545,
1286 0x0D10, 0x0D17, 0x0D20, 0x0D23, 0x0D30, 0x0D30, 0x20C0, 0x20C0,
1287 0x24C0, 0x24C0, 0x0E40, 0x0E43, 0x0E4A, 0x0E4A, 0x0E50, 0x0E57,
1288 0x0E60, 0x0E7C, 0x0E80, 0x0E8E, 0x0E90, 0x0E96, 0x0EA0, 0x0EA8,
1289 0x0EB0, 0x0EB2, 0xE140, 0xE147, 0xE150, 0xE187, 0xE1A0, 0xE1A9,
1290 0xE1B0, 0xE1B6, 0xE1C0, 0xE1C7, 0xE1D0, 0xE1D1, 0xE200, 0xE201,
1291 0xE210, 0xE21C, 0xE240, 0xE268, 0xE000, 0xE006, 0xE010, 0xE09A,
1292 0xE0A0, 0xE0A4, 0xE0AA, 0xE0EB, 0xE100, 0xE105, 0xE380, 0xE38F,
1293 0xE3B0, 0xE3B0, 0xE400, 0xE405, 0xE408, 0xE4E9, 0xE4F0, 0xE4F0,
1294 0xE280, 0xE280, 0xE282, 0xE2A3, 0xE2A5, 0xE2C2, 0xE940, 0xE947,
1295 0xE950, 0xE987, 0xE9A0, 0xE9A9, 0xE9B0, 0xE9B6, 0xE9C0, 0xE9C7,
1296 0xE9D0, 0xE9D1, 0xEA00, 0xEA01, 0xEA10, 0xEA1C, 0xEA40, 0xEA68,
1297 0xE800, 0xE806, 0xE810, 0xE89A, 0xE8A0, 0xE8A4, 0xE8AA, 0xE8EB,
1298 0xE900, 0xE905, 0xEB80, 0xEB8F, 0xEBB0, 0xEBB0, 0xEC00, 0xEC05,
1299 0xEC08, 0xECE9, 0xECF0, 0xECF0, 0xEA80, 0xEA80, 0xEA82, 0xEAA3,
1300 0xEAA5, 0xEAC2, 0xA800, 0xA800, 0xA820, 0xA828, 0xA840, 0xA87D,
1301 0XA880, 0xA88D, 0xA890, 0xA8A3, 0xA8D0, 0xA8D8, 0xA8E0, 0xA8F5,
1305 static void a5xx_dump(struct msm_gpu *gpu)
1307 DRM_DEV_INFO(gpu->dev->dev, "status: %08x\n",
1308 gpu_read(gpu, REG_A5XX_RBBM_STATUS));
1312 static int a5xx_pm_resume(struct msm_gpu *gpu)
1314 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
1317 /* Turn on the core power */
1318 ret = msm_gpu_pm_resume(gpu);
1322 /* Adreno 508, 509, 510, 512 needs manual RBBM sus/res control */
1323 if (!(adreno_is_a530(adreno_gpu) || adreno_is_a540(adreno_gpu))) {
1324 /* Halt the sp_input_clk at HM level */
1325 gpu_write(gpu, REG_A5XX_RBBM_CLOCK_CNTL, 0x00000055);
1326 a5xx_set_hwcg(gpu, true);
1327 /* Turn on sp_input_clk at HM level */
1328 gpu_rmw(gpu, REG_A5XX_RBBM_CLOCK_CNTL, 0xff, 0);
1332 /* Turn the RBCCU domain first to limit the chances of voltage droop */
1333 gpu_write(gpu, REG_A5XX_GPMU_RBCCU_POWER_CNTL, 0x778000);
1335 /* Wait 3 usecs before polling */
1338 ret = spin_usecs(gpu, 20, REG_A5XX_GPMU_RBCCU_PWR_CLK_STATUS,
1339 (1 << 20), (1 << 20));
1341 DRM_ERROR("%s: timeout waiting for RBCCU GDSC enable: %X\n",
1343 gpu_read(gpu, REG_A5XX_GPMU_RBCCU_PWR_CLK_STATUS));
1347 /* Turn on the SP domain */
1348 gpu_write(gpu, REG_A5XX_GPMU_SP_POWER_CNTL, 0x778000);
1349 ret = spin_usecs(gpu, 20, REG_A5XX_GPMU_SP_PWR_CLK_STATUS,
1350 (1 << 20), (1 << 20));
1352 DRM_ERROR("%s: timeout waiting for SP GDSC enable\n",
1358 static int a5xx_pm_suspend(struct msm_gpu *gpu)
1360 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
1361 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
1365 /* A508, A510 have 3 XIN ports in VBIF */
1366 if (adreno_is_a508(adreno_gpu) || adreno_is_a510(adreno_gpu))
1369 /* Clear the VBIF pipe before shutting down */
1370 gpu_write(gpu, REG_A5XX_VBIF_XIN_HALT_CTRL0, mask);
1371 spin_until((gpu_read(gpu, REG_A5XX_VBIF_XIN_HALT_CTRL1) &
1374 gpu_write(gpu, REG_A5XX_VBIF_XIN_HALT_CTRL0, 0);
1377 * Reset the VBIF before power collapse to avoid issue with FIFO
1378 * entries on Adreno A510 and A530 (the others will tend to lock up)
1380 if (adreno_is_a510(adreno_gpu) || adreno_is_a530(adreno_gpu)) {
1381 gpu_write(gpu, REG_A5XX_RBBM_BLOCK_SW_RESET_CMD, 0x003C0000);
1382 gpu_write(gpu, REG_A5XX_RBBM_BLOCK_SW_RESET_CMD, 0x00000000);
1385 ret = msm_gpu_pm_suspend(gpu);
1389 if (a5xx_gpu->has_whereami)
1390 for (i = 0; i < gpu->nr_rings; i++)
1391 a5xx_gpu->shadow[i] = 0;
1396 static int a5xx_get_timestamp(struct msm_gpu *gpu, uint64_t *value)
1398 *value = gpu_read64(gpu, REG_A5XX_RBBM_ALWAYSON_COUNTER_LO,
1399 REG_A5XX_RBBM_ALWAYSON_COUNTER_HI);
1404 struct a5xx_crashdumper {
1406 struct drm_gem_object *bo;
1410 struct a5xx_gpu_state {
1411 struct msm_gpu_state base;
1415 static int a5xx_crashdumper_init(struct msm_gpu *gpu,
1416 struct a5xx_crashdumper *dumper)
1418 dumper->ptr = msm_gem_kernel_new_locked(gpu->dev,
1419 SZ_1M, MSM_BO_WC, gpu->aspace,
1420 &dumper->bo, &dumper->iova);
1422 if (!IS_ERR(dumper->ptr))
1423 msm_gem_object_set_name(dumper->bo, "crashdump");
1425 return PTR_ERR_OR_ZERO(dumper->ptr);
1428 static int a5xx_crashdumper_run(struct msm_gpu *gpu,
1429 struct a5xx_crashdumper *dumper)
1433 if (IS_ERR_OR_NULL(dumper->ptr))
1436 gpu_write64(gpu, REG_A5XX_CP_CRASH_SCRIPT_BASE_LO,
1437 REG_A5XX_CP_CRASH_SCRIPT_BASE_HI, dumper->iova);
1439 gpu_write(gpu, REG_A5XX_CP_CRASH_DUMP_CNTL, 1);
1441 return gpu_poll_timeout(gpu, REG_A5XX_CP_CRASH_DUMP_CNTL, val,
1442 val & 0x04, 100, 10000);
1446 * These are a list of the registers that need to be read through the HLSQ
1447 * aperture through the crashdumper. These are not nominally accessible from
1448 * the CPU on a secure platform.
1450 static const struct {
1454 } a5xx_hlsq_aperture_regs[] = {
1455 { 0x35, 0xe00, 0x32 }, /* HSLQ non-context */
1456 { 0x31, 0x2080, 0x1 }, /* HLSQ 2D context 0 */
1457 { 0x33, 0x2480, 0x1 }, /* HLSQ 2D context 1 */
1458 { 0x32, 0xe780, 0x62 }, /* HLSQ 3D context 0 */
1459 { 0x34, 0xef80, 0x62 }, /* HLSQ 3D context 1 */
1460 { 0x3f, 0x0ec0, 0x40 }, /* SP non-context */
1461 { 0x3d, 0x2040, 0x1 }, /* SP 2D context 0 */
1462 { 0x3b, 0x2440, 0x1 }, /* SP 2D context 1 */
1463 { 0x3e, 0xe580, 0x170 }, /* SP 3D context 0 */
1464 { 0x3c, 0xed80, 0x170 }, /* SP 3D context 1 */
1465 { 0x3a, 0x0f00, 0x1c }, /* TP non-context */
1466 { 0x38, 0x2000, 0xa }, /* TP 2D context 0 */
1467 { 0x36, 0x2400, 0xa }, /* TP 2D context 1 */
1468 { 0x39, 0xe700, 0x80 }, /* TP 3D context 0 */
1469 { 0x37, 0xef00, 0x80 }, /* TP 3D context 1 */
1472 static void a5xx_gpu_state_get_hlsq_regs(struct msm_gpu *gpu,
1473 struct a5xx_gpu_state *a5xx_state)
1475 struct a5xx_crashdumper dumper = { 0 };
1476 u32 offset, count = 0;
1480 if (a5xx_crashdumper_init(gpu, &dumper))
1483 /* The script will be written at offset 0 */
1486 /* Start writing the data at offset 256k */
1487 offset = dumper.iova + (256 * SZ_1K);
1489 /* Count how many additional registers to get from the HLSQ aperture */
1490 for (i = 0; i < ARRAY_SIZE(a5xx_hlsq_aperture_regs); i++)
1491 count += a5xx_hlsq_aperture_regs[i].count;
1493 a5xx_state->hlsqregs = kcalloc(count, sizeof(u32), GFP_KERNEL);
1494 if (!a5xx_state->hlsqregs)
1497 /* Build the crashdump script */
1498 for (i = 0; i < ARRAY_SIZE(a5xx_hlsq_aperture_regs); i++) {
1499 u32 type = a5xx_hlsq_aperture_regs[i].type;
1500 u32 c = a5xx_hlsq_aperture_regs[i].count;
1502 /* Write the register to select the desired bank */
1503 *ptr++ = ((u64) type << 8);
1504 *ptr++ = (((u64) REG_A5XX_HLSQ_DBG_READ_SEL) << 44) |
1508 *ptr++ = (((u64) REG_A5XX_HLSQ_DBG_AHB_READ_APERTURE) << 44)
1511 offset += c * sizeof(u32);
1514 /* Write two zeros to close off the script */
1518 if (a5xx_crashdumper_run(gpu, &dumper)) {
1519 kfree(a5xx_state->hlsqregs);
1520 msm_gem_kernel_put(dumper.bo, gpu->aspace, true);
1524 /* Copy the data from the crashdumper to the state */
1525 memcpy(a5xx_state->hlsqregs, dumper.ptr + (256 * SZ_1K),
1526 count * sizeof(u32));
1528 msm_gem_kernel_put(dumper.bo, gpu->aspace, true);
1531 static struct msm_gpu_state *a5xx_gpu_state_get(struct msm_gpu *gpu)
1533 struct a5xx_gpu_state *a5xx_state = kzalloc(sizeof(*a5xx_state),
1535 bool stalled = !!(gpu_read(gpu, REG_A5XX_RBBM_STATUS3) & BIT(24));
1538 return ERR_PTR(-ENOMEM);
1540 /* Temporarily disable hardware clock gating before reading the hw */
1541 a5xx_set_hwcg(gpu, false);
1543 /* First get the generic state from the adreno core */
1544 adreno_gpu_state_get(gpu, &(a5xx_state->base));
1546 a5xx_state->base.rbbm_status = gpu_read(gpu, REG_A5XX_RBBM_STATUS);
1549 * Get the HLSQ regs with the help of the crashdumper, but only if
1550 * we are not stalled in an iommu fault (in which case the crashdumper
1551 * would not have access to memory)
1554 a5xx_gpu_state_get_hlsq_regs(gpu, a5xx_state);
1556 a5xx_set_hwcg(gpu, true);
1558 return &a5xx_state->base;
1561 static void a5xx_gpu_state_destroy(struct kref *kref)
1563 struct msm_gpu_state *state = container_of(kref,
1564 struct msm_gpu_state, ref);
1565 struct a5xx_gpu_state *a5xx_state = container_of(state,
1566 struct a5xx_gpu_state, base);
1568 kfree(a5xx_state->hlsqregs);
1570 adreno_gpu_state_destroy(state);
1574 static int a5xx_gpu_state_put(struct msm_gpu_state *state)
1576 if (IS_ERR_OR_NULL(state))
1579 return kref_put(&state->ref, a5xx_gpu_state_destroy);
1583 #if defined(CONFIG_DEBUG_FS) || defined(CONFIG_DEV_COREDUMP)
1584 static void a5xx_show(struct msm_gpu *gpu, struct msm_gpu_state *state,
1585 struct drm_printer *p)
1589 struct a5xx_gpu_state *a5xx_state = container_of(state,
1590 struct a5xx_gpu_state, base);
1592 if (IS_ERR_OR_NULL(state))
1595 adreno_show(gpu, state, p);
1597 /* Dump the additional a5xx HLSQ registers */
1598 if (!a5xx_state->hlsqregs)
1601 drm_printf(p, "registers-hlsq:\n");
1603 for (i = 0; i < ARRAY_SIZE(a5xx_hlsq_aperture_regs); i++) {
1604 u32 o = a5xx_hlsq_aperture_regs[i].regoffset;
1605 u32 c = a5xx_hlsq_aperture_regs[i].count;
1607 for (j = 0; j < c; j++, pos++, o++) {
1609 * To keep the crashdump simple we pull the entire range
1610 * for each register type but not all of the registers
1611 * in the range are valid. Fortunately invalid registers
1612 * stick out like a sore thumb with a value of
1615 if (a5xx_state->hlsqregs[pos] == 0xdeadbeef)
1618 drm_printf(p, " - { offset: 0x%04x, value: 0x%08x }\n",
1619 o << 2, a5xx_state->hlsqregs[pos]);
1625 static struct msm_ringbuffer *a5xx_active_ring(struct msm_gpu *gpu)
1627 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
1628 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
1630 return a5xx_gpu->cur_ring;
1633 static unsigned long a5xx_gpu_busy(struct msm_gpu *gpu)
1635 u64 busy_cycles, busy_time;
1637 /* Only read the gpu busy if the hardware is already active */
1638 if (pm_runtime_get_if_in_use(&gpu->pdev->dev) == 0)
1641 busy_cycles = gpu_read64(gpu, REG_A5XX_RBBM_PERFCTR_RBBM_0_LO,
1642 REG_A5XX_RBBM_PERFCTR_RBBM_0_HI);
1644 busy_time = busy_cycles - gpu->devfreq.busy_cycles;
1645 do_div(busy_time, clk_get_rate(gpu->core_clk) / 1000000);
1647 gpu->devfreq.busy_cycles = busy_cycles;
1649 pm_runtime_put(&gpu->pdev->dev);
1651 if (WARN_ON(busy_time > ~0LU))
1654 return (unsigned long)busy_time;
1657 static uint32_t a5xx_get_rptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring)
1659 struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
1660 struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
1662 if (a5xx_gpu->has_whereami)
1663 return a5xx_gpu->shadow[ring->id];
1665 return ring->memptrs->rptr = gpu_read(gpu, REG_A5XX_CP_RB_RPTR);
1668 static const struct adreno_gpu_funcs funcs = {
1670 .get_param = adreno_get_param,
1671 .hw_init = a5xx_hw_init,
1672 .pm_suspend = a5xx_pm_suspend,
1673 .pm_resume = a5xx_pm_resume,
1674 .recover = a5xx_recover,
1675 .submit = a5xx_submit,
1676 .active_ring = a5xx_active_ring,
1678 .destroy = a5xx_destroy,
1679 #if defined(CONFIG_DEBUG_FS) || defined(CONFIG_DEV_COREDUMP)
1682 #if defined(CONFIG_DEBUG_FS)
1683 .debugfs_init = a5xx_debugfs_init,
1685 .gpu_busy = a5xx_gpu_busy,
1686 .gpu_state_get = a5xx_gpu_state_get,
1687 .gpu_state_put = a5xx_gpu_state_put,
1688 .create_address_space = adreno_iommu_create_address_space,
1689 .get_rptr = a5xx_get_rptr,
1691 .get_timestamp = a5xx_get_timestamp,
1694 static void check_speed_bin(struct device *dev)
1696 struct nvmem_cell *cell;
1700 * If the OPP table specifies a opp-supported-hw property then we have
1701 * to set something with dev_pm_opp_set_supported_hw() or the table
1702 * doesn't get populated so pick an arbitrary value that should
1703 * ensure the default frequencies are selected but not conflict with any
1708 cell = nvmem_cell_get(dev, "speed_bin");
1710 if (!IS_ERR(cell)) {
1711 void *buf = nvmem_cell_read(cell, NULL);
1714 u8 bin = *((u8 *) buf);
1720 nvmem_cell_put(cell);
1723 devm_pm_opp_set_supported_hw(dev, &val, 1);
1726 struct msm_gpu *a5xx_gpu_init(struct drm_device *dev)
1728 struct msm_drm_private *priv = dev->dev_private;
1729 struct platform_device *pdev = priv->gpu_pdev;
1730 struct a5xx_gpu *a5xx_gpu = NULL;
1731 struct adreno_gpu *adreno_gpu;
1732 struct msm_gpu *gpu;
1736 DRM_DEV_ERROR(dev->dev, "No A5XX device is defined\n");
1737 return ERR_PTR(-ENXIO);
1740 a5xx_gpu = kzalloc(sizeof(*a5xx_gpu), GFP_KERNEL);
1742 return ERR_PTR(-ENOMEM);
1744 adreno_gpu = &a5xx_gpu->base;
1745 gpu = &adreno_gpu->base;
1747 adreno_gpu->registers = a5xx_registers;
1749 a5xx_gpu->lm_leakage = 0x4E001A;
1751 check_speed_bin(&pdev->dev);
1753 ret = adreno_gpu_init(dev, pdev, adreno_gpu, &funcs, 4);
1755 a5xx_destroy(&(a5xx_gpu->base.base));
1756 return ERR_PTR(ret);
1760 msm_mmu_set_fault_handler(gpu->aspace->mmu, gpu, a5xx_fault_handler);
1762 /* Set up the preemption specific bits and pieces for each ringbuffer */
1763 a5xx_preempt_init(gpu);