GNU Linux-libre 5.10.153-gnu1
[releases.git] / drivers / gpu / drm / radeon / evergreen.c
1 /*
2  * Copyright 2010 Advanced Micro Devices, Inc.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20  * OTHER DEALINGS IN THE SOFTWARE.
21  *
22  * Authors: Alex Deucher
23  */
24
25 #include <linux/firmware.h>
26 #include <linux/pci.h>
27 #include <linux/slab.h>
28
29 #include <drm/drm_vblank.h>
30 #include <drm/radeon_drm.h>
31
32 #include "atom.h"
33 #include "avivod.h"
34 #include "evergreen_blit_shaders.h"
35 #include "evergreen_reg.h"
36 #include "evergreend.h"
37 #include "radeon.h"
38 #include "radeon_asic.h"
39 #include "radeon_audio.h"
40 #include "radeon_ucode.h"
41
42 #define DC_HPDx_CONTROL(x)        (DC_HPD1_CONTROL     + (x * 0xc))
43 #define DC_HPDx_INT_CONTROL(x)    (DC_HPD1_INT_CONTROL + (x * 0xc))
44 #define DC_HPDx_INT_STATUS_REG(x) (DC_HPD1_INT_STATUS  + (x * 0xc))
45
46 /*
47  * Indirect registers accessor
48  */
49 u32 eg_cg_rreg(struct radeon_device *rdev, u32 reg)
50 {
51         unsigned long flags;
52         u32 r;
53
54         spin_lock_irqsave(&rdev->cg_idx_lock, flags);
55         WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff));
56         r = RREG32(EVERGREEN_CG_IND_DATA);
57         spin_unlock_irqrestore(&rdev->cg_idx_lock, flags);
58         return r;
59 }
60
61 void eg_cg_wreg(struct radeon_device *rdev, u32 reg, u32 v)
62 {
63         unsigned long flags;
64
65         spin_lock_irqsave(&rdev->cg_idx_lock, flags);
66         WREG32(EVERGREEN_CG_IND_ADDR, ((reg) & 0xffff));
67         WREG32(EVERGREEN_CG_IND_DATA, (v));
68         spin_unlock_irqrestore(&rdev->cg_idx_lock, flags);
69 }
70
71 u32 eg_pif_phy0_rreg(struct radeon_device *rdev, u32 reg)
72 {
73         unsigned long flags;
74         u32 r;
75
76         spin_lock_irqsave(&rdev->pif_idx_lock, flags);
77         WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff));
78         r = RREG32(EVERGREEN_PIF_PHY0_DATA);
79         spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
80         return r;
81 }
82
83 void eg_pif_phy0_wreg(struct radeon_device *rdev, u32 reg, u32 v)
84 {
85         unsigned long flags;
86
87         spin_lock_irqsave(&rdev->pif_idx_lock, flags);
88         WREG32(EVERGREEN_PIF_PHY0_INDEX, ((reg) & 0xffff));
89         WREG32(EVERGREEN_PIF_PHY0_DATA, (v));
90         spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
91 }
92
93 u32 eg_pif_phy1_rreg(struct radeon_device *rdev, u32 reg)
94 {
95         unsigned long flags;
96         u32 r;
97
98         spin_lock_irqsave(&rdev->pif_idx_lock, flags);
99         WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff));
100         r = RREG32(EVERGREEN_PIF_PHY1_DATA);
101         spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
102         return r;
103 }
104
105 void eg_pif_phy1_wreg(struct radeon_device *rdev, u32 reg, u32 v)
106 {
107         unsigned long flags;
108
109         spin_lock_irqsave(&rdev->pif_idx_lock, flags);
110         WREG32(EVERGREEN_PIF_PHY1_INDEX, ((reg) & 0xffff));
111         WREG32(EVERGREEN_PIF_PHY1_DATA, (v));
112         spin_unlock_irqrestore(&rdev->pif_idx_lock, flags);
113 }
114
115 static const u32 crtc_offsets[6] =
116 {
117         EVERGREEN_CRTC0_REGISTER_OFFSET,
118         EVERGREEN_CRTC1_REGISTER_OFFSET,
119         EVERGREEN_CRTC2_REGISTER_OFFSET,
120         EVERGREEN_CRTC3_REGISTER_OFFSET,
121         EVERGREEN_CRTC4_REGISTER_OFFSET,
122         EVERGREEN_CRTC5_REGISTER_OFFSET
123 };
124
125 #include "clearstate_evergreen.h"
126
127 static const u32 sumo_rlc_save_restore_register_list[] =
128 {
129         0x98fc,
130         0x9830,
131         0x9834,
132         0x9838,
133         0x9870,
134         0x9874,
135         0x8a14,
136         0x8b24,
137         0x8bcc,
138         0x8b10,
139         0x8d00,
140         0x8d04,
141         0x8c00,
142         0x8c04,
143         0x8c08,
144         0x8c0c,
145         0x8d8c,
146         0x8c20,
147         0x8c24,
148         0x8c28,
149         0x8c18,
150         0x8c1c,
151         0x8cf0,
152         0x8e2c,
153         0x8e38,
154         0x8c30,
155         0x9508,
156         0x9688,
157         0x9608,
158         0x960c,
159         0x9610,
160         0x9614,
161         0x88c4,
162         0x88d4,
163         0xa008,
164         0x900c,
165         0x9100,
166         0x913c,
167         0x98f8,
168         0x98f4,
169         0x9b7c,
170         0x3f8c,
171         0x8950,
172         0x8954,
173         0x8a18,
174         0x8b28,
175         0x9144,
176         0x9148,
177         0x914c,
178         0x3f90,
179         0x3f94,
180         0x915c,
181         0x9160,
182         0x9178,
183         0x917c,
184         0x9180,
185         0x918c,
186         0x9190,
187         0x9194,
188         0x9198,
189         0x919c,
190         0x91a8,
191         0x91ac,
192         0x91b0,
193         0x91b4,
194         0x91b8,
195         0x91c4,
196         0x91c8,
197         0x91cc,
198         0x91d0,
199         0x91d4,
200         0x91e0,
201         0x91e4,
202         0x91ec,
203         0x91f0,
204         0x91f4,
205         0x9200,
206         0x9204,
207         0x929c,
208         0x9150,
209         0x802c,
210 };
211
212 static void evergreen_gpu_init(struct radeon_device *rdev);
213 void evergreen_fini(struct radeon_device *rdev);
214 void evergreen_pcie_gen2_enable(struct radeon_device *rdev);
215 void evergreen_program_aspm(struct radeon_device *rdev);
216 extern void cayman_cp_int_cntl_setup(struct radeon_device *rdev,
217                                      int ring, u32 cp_int_cntl);
218 extern void cayman_vm_decode_fault(struct radeon_device *rdev,
219                                    u32 status, u32 addr);
220 void cik_init_cp_pg_table(struct radeon_device *rdev);
221
222 extern u32 si_get_csb_size(struct radeon_device *rdev);
223 extern void si_get_csb_buffer(struct radeon_device *rdev, volatile u32 *buffer);
224 extern u32 cik_get_csb_size(struct radeon_device *rdev);
225 extern void cik_get_csb_buffer(struct radeon_device *rdev, volatile u32 *buffer);
226 extern void rv770_set_clk_bypass_mode(struct radeon_device *rdev);
227
228 static const u32 evergreen_golden_registers[] =
229 {
230         0x3f90, 0xffff0000, 0xff000000,
231         0x9148, 0xffff0000, 0xff000000,
232         0x3f94, 0xffff0000, 0xff000000,
233         0x914c, 0xffff0000, 0xff000000,
234         0x9b7c, 0xffffffff, 0x00000000,
235         0x8a14, 0xffffffff, 0x00000007,
236         0x8b10, 0xffffffff, 0x00000000,
237         0x960c, 0xffffffff, 0x54763210,
238         0x88c4, 0xffffffff, 0x000000c2,
239         0x88d4, 0xffffffff, 0x00000010,
240         0x8974, 0xffffffff, 0x00000000,
241         0xc78, 0x00000080, 0x00000080,
242         0x5eb4, 0xffffffff, 0x00000002,
243         0x5e78, 0xffffffff, 0x001000f0,
244         0x6104, 0x01000300, 0x00000000,
245         0x5bc0, 0x00300000, 0x00000000,
246         0x7030, 0xffffffff, 0x00000011,
247         0x7c30, 0xffffffff, 0x00000011,
248         0x10830, 0xffffffff, 0x00000011,
249         0x11430, 0xffffffff, 0x00000011,
250         0x12030, 0xffffffff, 0x00000011,
251         0x12c30, 0xffffffff, 0x00000011,
252         0xd02c, 0xffffffff, 0x08421000,
253         0x240c, 0xffffffff, 0x00000380,
254         0x8b24, 0xffffffff, 0x00ff0fff,
255         0x28a4c, 0x06000000, 0x06000000,
256         0x10c, 0x00000001, 0x00000001,
257         0x8d00, 0xffffffff, 0x100e4848,
258         0x8d04, 0xffffffff, 0x00164745,
259         0x8c00, 0xffffffff, 0xe4000003,
260         0x8c04, 0xffffffff, 0x40600060,
261         0x8c08, 0xffffffff, 0x001c001c,
262         0x8cf0, 0xffffffff, 0x08e00620,
263         0x8c20, 0xffffffff, 0x00800080,
264         0x8c24, 0xffffffff, 0x00800080,
265         0x8c18, 0xffffffff, 0x20202078,
266         0x8c1c, 0xffffffff, 0x00001010,
267         0x28350, 0xffffffff, 0x00000000,
268         0xa008, 0xffffffff, 0x00010000,
269         0x5c4, 0xffffffff, 0x00000001,
270         0x9508, 0xffffffff, 0x00000002,
271         0x913c, 0x0000000f, 0x0000000a
272 };
273
274 static const u32 evergreen_golden_registers2[] =
275 {
276         0x2f4c, 0xffffffff, 0x00000000,
277         0x54f4, 0xffffffff, 0x00000000,
278         0x54f0, 0xffffffff, 0x00000000,
279         0x5498, 0xffffffff, 0x00000000,
280         0x549c, 0xffffffff, 0x00000000,
281         0x5494, 0xffffffff, 0x00000000,
282         0x53cc, 0xffffffff, 0x00000000,
283         0x53c8, 0xffffffff, 0x00000000,
284         0x53c4, 0xffffffff, 0x00000000,
285         0x53c0, 0xffffffff, 0x00000000,
286         0x53bc, 0xffffffff, 0x00000000,
287         0x53b8, 0xffffffff, 0x00000000,
288         0x53b4, 0xffffffff, 0x00000000,
289         0x53b0, 0xffffffff, 0x00000000
290 };
291
292 static const u32 cypress_mgcg_init[] =
293 {
294         0x802c, 0xffffffff, 0xc0000000,
295         0x5448, 0xffffffff, 0x00000100,
296         0x55e4, 0xffffffff, 0x00000100,
297         0x160c, 0xffffffff, 0x00000100,
298         0x5644, 0xffffffff, 0x00000100,
299         0xc164, 0xffffffff, 0x00000100,
300         0x8a18, 0xffffffff, 0x00000100,
301         0x897c, 0xffffffff, 0x06000100,
302         0x8b28, 0xffffffff, 0x00000100,
303         0x9144, 0xffffffff, 0x00000100,
304         0x9a60, 0xffffffff, 0x00000100,
305         0x9868, 0xffffffff, 0x00000100,
306         0x8d58, 0xffffffff, 0x00000100,
307         0x9510, 0xffffffff, 0x00000100,
308         0x949c, 0xffffffff, 0x00000100,
309         0x9654, 0xffffffff, 0x00000100,
310         0x9030, 0xffffffff, 0x00000100,
311         0x9034, 0xffffffff, 0x00000100,
312         0x9038, 0xffffffff, 0x00000100,
313         0x903c, 0xffffffff, 0x00000100,
314         0x9040, 0xffffffff, 0x00000100,
315         0xa200, 0xffffffff, 0x00000100,
316         0xa204, 0xffffffff, 0x00000100,
317         0xa208, 0xffffffff, 0x00000100,
318         0xa20c, 0xffffffff, 0x00000100,
319         0x971c, 0xffffffff, 0x00000100,
320         0x977c, 0xffffffff, 0x00000100,
321         0x3f80, 0xffffffff, 0x00000100,
322         0xa210, 0xffffffff, 0x00000100,
323         0xa214, 0xffffffff, 0x00000100,
324         0x4d8, 0xffffffff, 0x00000100,
325         0x9784, 0xffffffff, 0x00000100,
326         0x9698, 0xffffffff, 0x00000100,
327         0x4d4, 0xffffffff, 0x00000200,
328         0x30cc, 0xffffffff, 0x00000100,
329         0xd0c0, 0xffffffff, 0xff000100,
330         0x802c, 0xffffffff, 0x40000000,
331         0x915c, 0xffffffff, 0x00010000,
332         0x9160, 0xffffffff, 0x00030002,
333         0x9178, 0xffffffff, 0x00070000,
334         0x917c, 0xffffffff, 0x00030002,
335         0x9180, 0xffffffff, 0x00050004,
336         0x918c, 0xffffffff, 0x00010006,
337         0x9190, 0xffffffff, 0x00090008,
338         0x9194, 0xffffffff, 0x00070000,
339         0x9198, 0xffffffff, 0x00030002,
340         0x919c, 0xffffffff, 0x00050004,
341         0x91a8, 0xffffffff, 0x00010006,
342         0x91ac, 0xffffffff, 0x00090008,
343         0x91b0, 0xffffffff, 0x00070000,
344         0x91b4, 0xffffffff, 0x00030002,
345         0x91b8, 0xffffffff, 0x00050004,
346         0x91c4, 0xffffffff, 0x00010006,
347         0x91c8, 0xffffffff, 0x00090008,
348         0x91cc, 0xffffffff, 0x00070000,
349         0x91d0, 0xffffffff, 0x00030002,
350         0x91d4, 0xffffffff, 0x00050004,
351         0x91e0, 0xffffffff, 0x00010006,
352         0x91e4, 0xffffffff, 0x00090008,
353         0x91e8, 0xffffffff, 0x00000000,
354         0x91ec, 0xffffffff, 0x00070000,
355         0x91f0, 0xffffffff, 0x00030002,
356         0x91f4, 0xffffffff, 0x00050004,
357         0x9200, 0xffffffff, 0x00010006,
358         0x9204, 0xffffffff, 0x00090008,
359         0x9208, 0xffffffff, 0x00070000,
360         0x920c, 0xffffffff, 0x00030002,
361         0x9210, 0xffffffff, 0x00050004,
362         0x921c, 0xffffffff, 0x00010006,
363         0x9220, 0xffffffff, 0x00090008,
364         0x9224, 0xffffffff, 0x00070000,
365         0x9228, 0xffffffff, 0x00030002,
366         0x922c, 0xffffffff, 0x00050004,
367         0x9238, 0xffffffff, 0x00010006,
368         0x923c, 0xffffffff, 0x00090008,
369         0x9240, 0xffffffff, 0x00070000,
370         0x9244, 0xffffffff, 0x00030002,
371         0x9248, 0xffffffff, 0x00050004,
372         0x9254, 0xffffffff, 0x00010006,
373         0x9258, 0xffffffff, 0x00090008,
374         0x925c, 0xffffffff, 0x00070000,
375         0x9260, 0xffffffff, 0x00030002,
376         0x9264, 0xffffffff, 0x00050004,
377         0x9270, 0xffffffff, 0x00010006,
378         0x9274, 0xffffffff, 0x00090008,
379         0x9278, 0xffffffff, 0x00070000,
380         0x927c, 0xffffffff, 0x00030002,
381         0x9280, 0xffffffff, 0x00050004,
382         0x928c, 0xffffffff, 0x00010006,
383         0x9290, 0xffffffff, 0x00090008,
384         0x9294, 0xffffffff, 0x00000000,
385         0x929c, 0xffffffff, 0x00000001,
386         0x802c, 0xffffffff, 0x40010000,
387         0x915c, 0xffffffff, 0x00010000,
388         0x9160, 0xffffffff, 0x00030002,
389         0x9178, 0xffffffff, 0x00070000,
390         0x917c, 0xffffffff, 0x00030002,
391         0x9180, 0xffffffff, 0x00050004,
392         0x918c, 0xffffffff, 0x00010006,
393         0x9190, 0xffffffff, 0x00090008,
394         0x9194, 0xffffffff, 0x00070000,
395         0x9198, 0xffffffff, 0x00030002,
396         0x919c, 0xffffffff, 0x00050004,
397         0x91a8, 0xffffffff, 0x00010006,
398         0x91ac, 0xffffffff, 0x00090008,
399         0x91b0, 0xffffffff, 0x00070000,
400         0x91b4, 0xffffffff, 0x00030002,
401         0x91b8, 0xffffffff, 0x00050004,
402         0x91c4, 0xffffffff, 0x00010006,
403         0x91c8, 0xffffffff, 0x00090008,
404         0x91cc, 0xffffffff, 0x00070000,
405         0x91d0, 0xffffffff, 0x00030002,
406         0x91d4, 0xffffffff, 0x00050004,
407         0x91e0, 0xffffffff, 0x00010006,
408         0x91e4, 0xffffffff, 0x00090008,
409         0x91e8, 0xffffffff, 0x00000000,
410         0x91ec, 0xffffffff, 0x00070000,
411         0x91f0, 0xffffffff, 0x00030002,
412         0x91f4, 0xffffffff, 0x00050004,
413         0x9200, 0xffffffff, 0x00010006,
414         0x9204, 0xffffffff, 0x00090008,
415         0x9208, 0xffffffff, 0x00070000,
416         0x920c, 0xffffffff, 0x00030002,
417         0x9210, 0xffffffff, 0x00050004,
418         0x921c, 0xffffffff, 0x00010006,
419         0x9220, 0xffffffff, 0x00090008,
420         0x9224, 0xffffffff, 0x00070000,
421         0x9228, 0xffffffff, 0x00030002,
422         0x922c, 0xffffffff, 0x00050004,
423         0x9238, 0xffffffff, 0x00010006,
424         0x923c, 0xffffffff, 0x00090008,
425         0x9240, 0xffffffff, 0x00070000,
426         0x9244, 0xffffffff, 0x00030002,
427         0x9248, 0xffffffff, 0x00050004,
428         0x9254, 0xffffffff, 0x00010006,
429         0x9258, 0xffffffff, 0x00090008,
430         0x925c, 0xffffffff, 0x00070000,
431         0x9260, 0xffffffff, 0x00030002,
432         0x9264, 0xffffffff, 0x00050004,
433         0x9270, 0xffffffff, 0x00010006,
434         0x9274, 0xffffffff, 0x00090008,
435         0x9278, 0xffffffff, 0x00070000,
436         0x927c, 0xffffffff, 0x00030002,
437         0x9280, 0xffffffff, 0x00050004,
438         0x928c, 0xffffffff, 0x00010006,
439         0x9290, 0xffffffff, 0x00090008,
440         0x9294, 0xffffffff, 0x00000000,
441         0x929c, 0xffffffff, 0x00000001,
442         0x802c, 0xffffffff, 0xc0000000
443 };
444
445 static const u32 redwood_mgcg_init[] =
446 {
447         0x802c, 0xffffffff, 0xc0000000,
448         0x5448, 0xffffffff, 0x00000100,
449         0x55e4, 0xffffffff, 0x00000100,
450         0x160c, 0xffffffff, 0x00000100,
451         0x5644, 0xffffffff, 0x00000100,
452         0xc164, 0xffffffff, 0x00000100,
453         0x8a18, 0xffffffff, 0x00000100,
454         0x897c, 0xffffffff, 0x06000100,
455         0x8b28, 0xffffffff, 0x00000100,
456         0x9144, 0xffffffff, 0x00000100,
457         0x9a60, 0xffffffff, 0x00000100,
458         0x9868, 0xffffffff, 0x00000100,
459         0x8d58, 0xffffffff, 0x00000100,
460         0x9510, 0xffffffff, 0x00000100,
461         0x949c, 0xffffffff, 0x00000100,
462         0x9654, 0xffffffff, 0x00000100,
463         0x9030, 0xffffffff, 0x00000100,
464         0x9034, 0xffffffff, 0x00000100,
465         0x9038, 0xffffffff, 0x00000100,
466         0x903c, 0xffffffff, 0x00000100,
467         0x9040, 0xffffffff, 0x00000100,
468         0xa200, 0xffffffff, 0x00000100,
469         0xa204, 0xffffffff, 0x00000100,
470         0xa208, 0xffffffff, 0x00000100,
471         0xa20c, 0xffffffff, 0x00000100,
472         0x971c, 0xffffffff, 0x00000100,
473         0x977c, 0xffffffff, 0x00000100,
474         0x3f80, 0xffffffff, 0x00000100,
475         0xa210, 0xffffffff, 0x00000100,
476         0xa214, 0xffffffff, 0x00000100,
477         0x4d8, 0xffffffff, 0x00000100,
478         0x9784, 0xffffffff, 0x00000100,
479         0x9698, 0xffffffff, 0x00000100,
480         0x4d4, 0xffffffff, 0x00000200,
481         0x30cc, 0xffffffff, 0x00000100,
482         0xd0c0, 0xffffffff, 0xff000100,
483         0x802c, 0xffffffff, 0x40000000,
484         0x915c, 0xffffffff, 0x00010000,
485         0x9160, 0xffffffff, 0x00030002,
486         0x9178, 0xffffffff, 0x00070000,
487         0x917c, 0xffffffff, 0x00030002,
488         0x9180, 0xffffffff, 0x00050004,
489         0x918c, 0xffffffff, 0x00010006,
490         0x9190, 0xffffffff, 0x00090008,
491         0x9194, 0xffffffff, 0x00070000,
492         0x9198, 0xffffffff, 0x00030002,
493         0x919c, 0xffffffff, 0x00050004,
494         0x91a8, 0xffffffff, 0x00010006,
495         0x91ac, 0xffffffff, 0x00090008,
496         0x91b0, 0xffffffff, 0x00070000,
497         0x91b4, 0xffffffff, 0x00030002,
498         0x91b8, 0xffffffff, 0x00050004,
499         0x91c4, 0xffffffff, 0x00010006,
500         0x91c8, 0xffffffff, 0x00090008,
501         0x91cc, 0xffffffff, 0x00070000,
502         0x91d0, 0xffffffff, 0x00030002,
503         0x91d4, 0xffffffff, 0x00050004,
504         0x91e0, 0xffffffff, 0x00010006,
505         0x91e4, 0xffffffff, 0x00090008,
506         0x91e8, 0xffffffff, 0x00000000,
507         0x91ec, 0xffffffff, 0x00070000,
508         0x91f0, 0xffffffff, 0x00030002,
509         0x91f4, 0xffffffff, 0x00050004,
510         0x9200, 0xffffffff, 0x00010006,
511         0x9204, 0xffffffff, 0x00090008,
512         0x9294, 0xffffffff, 0x00000000,
513         0x929c, 0xffffffff, 0x00000001,
514         0x802c, 0xffffffff, 0xc0000000
515 };
516
517 static const u32 cedar_golden_registers[] =
518 {
519         0x3f90, 0xffff0000, 0xff000000,
520         0x9148, 0xffff0000, 0xff000000,
521         0x3f94, 0xffff0000, 0xff000000,
522         0x914c, 0xffff0000, 0xff000000,
523         0x9b7c, 0xffffffff, 0x00000000,
524         0x8a14, 0xffffffff, 0x00000007,
525         0x8b10, 0xffffffff, 0x00000000,
526         0x960c, 0xffffffff, 0x54763210,
527         0x88c4, 0xffffffff, 0x000000c2,
528         0x88d4, 0xffffffff, 0x00000000,
529         0x8974, 0xffffffff, 0x00000000,
530         0xc78, 0x00000080, 0x00000080,
531         0x5eb4, 0xffffffff, 0x00000002,
532         0x5e78, 0xffffffff, 0x001000f0,
533         0x6104, 0x01000300, 0x00000000,
534         0x5bc0, 0x00300000, 0x00000000,
535         0x7030, 0xffffffff, 0x00000011,
536         0x7c30, 0xffffffff, 0x00000011,
537         0x10830, 0xffffffff, 0x00000011,
538         0x11430, 0xffffffff, 0x00000011,
539         0xd02c, 0xffffffff, 0x08421000,
540         0x240c, 0xffffffff, 0x00000380,
541         0x8b24, 0xffffffff, 0x00ff0fff,
542         0x28a4c, 0x06000000, 0x06000000,
543         0x10c, 0x00000001, 0x00000001,
544         0x8d00, 0xffffffff, 0x100e4848,
545         0x8d04, 0xffffffff, 0x00164745,
546         0x8c00, 0xffffffff, 0xe4000003,
547         0x8c04, 0xffffffff, 0x40600060,
548         0x8c08, 0xffffffff, 0x001c001c,
549         0x8cf0, 0xffffffff, 0x08e00410,
550         0x8c20, 0xffffffff, 0x00800080,
551         0x8c24, 0xffffffff, 0x00800080,
552         0x8c18, 0xffffffff, 0x20202078,
553         0x8c1c, 0xffffffff, 0x00001010,
554         0x28350, 0xffffffff, 0x00000000,
555         0xa008, 0xffffffff, 0x00010000,
556         0x5c4, 0xffffffff, 0x00000001,
557         0x9508, 0xffffffff, 0x00000002
558 };
559
560 static const u32 cedar_mgcg_init[] =
561 {
562         0x802c, 0xffffffff, 0xc0000000,
563         0x5448, 0xffffffff, 0x00000100,
564         0x55e4, 0xffffffff, 0x00000100,
565         0x160c, 0xffffffff, 0x00000100,
566         0x5644, 0xffffffff, 0x00000100,
567         0xc164, 0xffffffff, 0x00000100,
568         0x8a18, 0xffffffff, 0x00000100,
569         0x897c, 0xffffffff, 0x06000100,
570         0x8b28, 0xffffffff, 0x00000100,
571         0x9144, 0xffffffff, 0x00000100,
572         0x9a60, 0xffffffff, 0x00000100,
573         0x9868, 0xffffffff, 0x00000100,
574         0x8d58, 0xffffffff, 0x00000100,
575         0x9510, 0xffffffff, 0x00000100,
576         0x949c, 0xffffffff, 0x00000100,
577         0x9654, 0xffffffff, 0x00000100,
578         0x9030, 0xffffffff, 0x00000100,
579         0x9034, 0xffffffff, 0x00000100,
580         0x9038, 0xffffffff, 0x00000100,
581         0x903c, 0xffffffff, 0x00000100,
582         0x9040, 0xffffffff, 0x00000100,
583         0xa200, 0xffffffff, 0x00000100,
584         0xa204, 0xffffffff, 0x00000100,
585         0xa208, 0xffffffff, 0x00000100,
586         0xa20c, 0xffffffff, 0x00000100,
587         0x971c, 0xffffffff, 0x00000100,
588         0x977c, 0xffffffff, 0x00000100,
589         0x3f80, 0xffffffff, 0x00000100,
590         0xa210, 0xffffffff, 0x00000100,
591         0xa214, 0xffffffff, 0x00000100,
592         0x4d8, 0xffffffff, 0x00000100,
593         0x9784, 0xffffffff, 0x00000100,
594         0x9698, 0xffffffff, 0x00000100,
595         0x4d4, 0xffffffff, 0x00000200,
596         0x30cc, 0xffffffff, 0x00000100,
597         0xd0c0, 0xffffffff, 0xff000100,
598         0x802c, 0xffffffff, 0x40000000,
599         0x915c, 0xffffffff, 0x00010000,
600         0x9178, 0xffffffff, 0x00050000,
601         0x917c, 0xffffffff, 0x00030002,
602         0x918c, 0xffffffff, 0x00010004,
603         0x9190, 0xffffffff, 0x00070006,
604         0x9194, 0xffffffff, 0x00050000,
605         0x9198, 0xffffffff, 0x00030002,
606         0x91a8, 0xffffffff, 0x00010004,
607         0x91ac, 0xffffffff, 0x00070006,
608         0x91e8, 0xffffffff, 0x00000000,
609         0x9294, 0xffffffff, 0x00000000,
610         0x929c, 0xffffffff, 0x00000001,
611         0x802c, 0xffffffff, 0xc0000000
612 };
613
614 static const u32 juniper_mgcg_init[] =
615 {
616         0x802c, 0xffffffff, 0xc0000000,
617         0x5448, 0xffffffff, 0x00000100,
618         0x55e4, 0xffffffff, 0x00000100,
619         0x160c, 0xffffffff, 0x00000100,
620         0x5644, 0xffffffff, 0x00000100,
621         0xc164, 0xffffffff, 0x00000100,
622         0x8a18, 0xffffffff, 0x00000100,
623         0x897c, 0xffffffff, 0x06000100,
624         0x8b28, 0xffffffff, 0x00000100,
625         0x9144, 0xffffffff, 0x00000100,
626         0x9a60, 0xffffffff, 0x00000100,
627         0x9868, 0xffffffff, 0x00000100,
628         0x8d58, 0xffffffff, 0x00000100,
629         0x9510, 0xffffffff, 0x00000100,
630         0x949c, 0xffffffff, 0x00000100,
631         0x9654, 0xffffffff, 0x00000100,
632         0x9030, 0xffffffff, 0x00000100,
633         0x9034, 0xffffffff, 0x00000100,
634         0x9038, 0xffffffff, 0x00000100,
635         0x903c, 0xffffffff, 0x00000100,
636         0x9040, 0xffffffff, 0x00000100,
637         0xa200, 0xffffffff, 0x00000100,
638         0xa204, 0xffffffff, 0x00000100,
639         0xa208, 0xffffffff, 0x00000100,
640         0xa20c, 0xffffffff, 0x00000100,
641         0x971c, 0xffffffff, 0x00000100,
642         0xd0c0, 0xffffffff, 0xff000100,
643         0x802c, 0xffffffff, 0x40000000,
644         0x915c, 0xffffffff, 0x00010000,
645         0x9160, 0xffffffff, 0x00030002,
646         0x9178, 0xffffffff, 0x00070000,
647         0x917c, 0xffffffff, 0x00030002,
648         0x9180, 0xffffffff, 0x00050004,
649         0x918c, 0xffffffff, 0x00010006,
650         0x9190, 0xffffffff, 0x00090008,
651         0x9194, 0xffffffff, 0x00070000,
652         0x9198, 0xffffffff, 0x00030002,
653         0x919c, 0xffffffff, 0x00050004,
654         0x91a8, 0xffffffff, 0x00010006,
655         0x91ac, 0xffffffff, 0x00090008,
656         0x91b0, 0xffffffff, 0x00070000,
657         0x91b4, 0xffffffff, 0x00030002,
658         0x91b8, 0xffffffff, 0x00050004,
659         0x91c4, 0xffffffff, 0x00010006,
660         0x91c8, 0xffffffff, 0x00090008,
661         0x91cc, 0xffffffff, 0x00070000,
662         0x91d0, 0xffffffff, 0x00030002,
663         0x91d4, 0xffffffff, 0x00050004,
664         0x91e0, 0xffffffff, 0x00010006,
665         0x91e4, 0xffffffff, 0x00090008,
666         0x91e8, 0xffffffff, 0x00000000,
667         0x91ec, 0xffffffff, 0x00070000,
668         0x91f0, 0xffffffff, 0x00030002,
669         0x91f4, 0xffffffff, 0x00050004,
670         0x9200, 0xffffffff, 0x00010006,
671         0x9204, 0xffffffff, 0x00090008,
672         0x9208, 0xffffffff, 0x00070000,
673         0x920c, 0xffffffff, 0x00030002,
674         0x9210, 0xffffffff, 0x00050004,
675         0x921c, 0xffffffff, 0x00010006,
676         0x9220, 0xffffffff, 0x00090008,
677         0x9224, 0xffffffff, 0x00070000,
678         0x9228, 0xffffffff, 0x00030002,
679         0x922c, 0xffffffff, 0x00050004,
680         0x9238, 0xffffffff, 0x00010006,
681         0x923c, 0xffffffff, 0x00090008,
682         0x9240, 0xffffffff, 0x00070000,
683         0x9244, 0xffffffff, 0x00030002,
684         0x9248, 0xffffffff, 0x00050004,
685         0x9254, 0xffffffff, 0x00010006,
686         0x9258, 0xffffffff, 0x00090008,
687         0x925c, 0xffffffff, 0x00070000,
688         0x9260, 0xffffffff, 0x00030002,
689         0x9264, 0xffffffff, 0x00050004,
690         0x9270, 0xffffffff, 0x00010006,
691         0x9274, 0xffffffff, 0x00090008,
692         0x9278, 0xffffffff, 0x00070000,
693         0x927c, 0xffffffff, 0x00030002,
694         0x9280, 0xffffffff, 0x00050004,
695         0x928c, 0xffffffff, 0x00010006,
696         0x9290, 0xffffffff, 0x00090008,
697         0x9294, 0xffffffff, 0x00000000,
698         0x929c, 0xffffffff, 0x00000001,
699         0x802c, 0xffffffff, 0xc0000000,
700         0x977c, 0xffffffff, 0x00000100,
701         0x3f80, 0xffffffff, 0x00000100,
702         0xa210, 0xffffffff, 0x00000100,
703         0xa214, 0xffffffff, 0x00000100,
704         0x4d8, 0xffffffff, 0x00000100,
705         0x9784, 0xffffffff, 0x00000100,
706         0x9698, 0xffffffff, 0x00000100,
707         0x4d4, 0xffffffff, 0x00000200,
708         0x30cc, 0xffffffff, 0x00000100,
709         0x802c, 0xffffffff, 0xc0000000
710 };
711
712 static const u32 supersumo_golden_registers[] =
713 {
714         0x5eb4, 0xffffffff, 0x00000002,
715         0x5c4, 0xffffffff, 0x00000001,
716         0x7030, 0xffffffff, 0x00000011,
717         0x7c30, 0xffffffff, 0x00000011,
718         0x6104, 0x01000300, 0x00000000,
719         0x5bc0, 0x00300000, 0x00000000,
720         0x8c04, 0xffffffff, 0x40600060,
721         0x8c08, 0xffffffff, 0x001c001c,
722         0x8c20, 0xffffffff, 0x00800080,
723         0x8c24, 0xffffffff, 0x00800080,
724         0x8c18, 0xffffffff, 0x20202078,
725         0x8c1c, 0xffffffff, 0x00001010,
726         0x918c, 0xffffffff, 0x00010006,
727         0x91a8, 0xffffffff, 0x00010006,
728         0x91c4, 0xffffffff, 0x00010006,
729         0x91e0, 0xffffffff, 0x00010006,
730         0x9200, 0xffffffff, 0x00010006,
731         0x9150, 0xffffffff, 0x6e944040,
732         0x917c, 0xffffffff, 0x00030002,
733         0x9180, 0xffffffff, 0x00050004,
734         0x9198, 0xffffffff, 0x00030002,
735         0x919c, 0xffffffff, 0x00050004,
736         0x91b4, 0xffffffff, 0x00030002,
737         0x91b8, 0xffffffff, 0x00050004,
738         0x91d0, 0xffffffff, 0x00030002,
739         0x91d4, 0xffffffff, 0x00050004,
740         0x91f0, 0xffffffff, 0x00030002,
741         0x91f4, 0xffffffff, 0x00050004,
742         0x915c, 0xffffffff, 0x00010000,
743         0x9160, 0xffffffff, 0x00030002,
744         0x3f90, 0xffff0000, 0xff000000,
745         0x9178, 0xffffffff, 0x00070000,
746         0x9194, 0xffffffff, 0x00070000,
747         0x91b0, 0xffffffff, 0x00070000,
748         0x91cc, 0xffffffff, 0x00070000,
749         0x91ec, 0xffffffff, 0x00070000,
750         0x9148, 0xffff0000, 0xff000000,
751         0x9190, 0xffffffff, 0x00090008,
752         0x91ac, 0xffffffff, 0x00090008,
753         0x91c8, 0xffffffff, 0x00090008,
754         0x91e4, 0xffffffff, 0x00090008,
755         0x9204, 0xffffffff, 0x00090008,
756         0x3f94, 0xffff0000, 0xff000000,
757         0x914c, 0xffff0000, 0xff000000,
758         0x929c, 0xffffffff, 0x00000001,
759         0x8a18, 0xffffffff, 0x00000100,
760         0x8b28, 0xffffffff, 0x00000100,
761         0x9144, 0xffffffff, 0x00000100,
762         0x5644, 0xffffffff, 0x00000100,
763         0x9b7c, 0xffffffff, 0x00000000,
764         0x8030, 0xffffffff, 0x0000100a,
765         0x8a14, 0xffffffff, 0x00000007,
766         0x8b24, 0xffffffff, 0x00ff0fff,
767         0x8b10, 0xffffffff, 0x00000000,
768         0x28a4c, 0x06000000, 0x06000000,
769         0x4d8, 0xffffffff, 0x00000100,
770         0x913c, 0xffff000f, 0x0100000a,
771         0x960c, 0xffffffff, 0x54763210,
772         0x88c4, 0xffffffff, 0x000000c2,
773         0x88d4, 0xffffffff, 0x00000010,
774         0x8974, 0xffffffff, 0x00000000,
775         0xc78, 0x00000080, 0x00000080,
776         0x5e78, 0xffffffff, 0x001000f0,
777         0xd02c, 0xffffffff, 0x08421000,
778         0xa008, 0xffffffff, 0x00010000,
779         0x8d00, 0xffffffff, 0x100e4848,
780         0x8d04, 0xffffffff, 0x00164745,
781         0x8c00, 0xffffffff, 0xe4000003,
782         0x8cf0, 0x1fffffff, 0x08e00620,
783         0x28350, 0xffffffff, 0x00000000,
784         0x9508, 0xffffffff, 0x00000002
785 };
786
787 static const u32 sumo_golden_registers[] =
788 {
789         0x900c, 0x00ffffff, 0x0017071f,
790         0x8c18, 0xffffffff, 0x10101060,
791         0x8c1c, 0xffffffff, 0x00001010,
792         0x8c30, 0x0000000f, 0x00000005,
793         0x9688, 0x0000000f, 0x00000007
794 };
795
796 static const u32 wrestler_golden_registers[] =
797 {
798         0x5eb4, 0xffffffff, 0x00000002,
799         0x5c4, 0xffffffff, 0x00000001,
800         0x7030, 0xffffffff, 0x00000011,
801         0x7c30, 0xffffffff, 0x00000011,
802         0x6104, 0x01000300, 0x00000000,
803         0x5bc0, 0x00300000, 0x00000000,
804         0x918c, 0xffffffff, 0x00010006,
805         0x91a8, 0xffffffff, 0x00010006,
806         0x9150, 0xffffffff, 0x6e944040,
807         0x917c, 0xffffffff, 0x00030002,
808         0x9198, 0xffffffff, 0x00030002,
809         0x915c, 0xffffffff, 0x00010000,
810         0x3f90, 0xffff0000, 0xff000000,
811         0x9178, 0xffffffff, 0x00070000,
812         0x9194, 0xffffffff, 0x00070000,
813         0x9148, 0xffff0000, 0xff000000,
814         0x9190, 0xffffffff, 0x00090008,
815         0x91ac, 0xffffffff, 0x00090008,
816         0x3f94, 0xffff0000, 0xff000000,
817         0x914c, 0xffff0000, 0xff000000,
818         0x929c, 0xffffffff, 0x00000001,
819         0x8a18, 0xffffffff, 0x00000100,
820         0x8b28, 0xffffffff, 0x00000100,
821         0x9144, 0xffffffff, 0x00000100,
822         0x9b7c, 0xffffffff, 0x00000000,
823         0x8030, 0xffffffff, 0x0000100a,
824         0x8a14, 0xffffffff, 0x00000001,
825         0x8b24, 0xffffffff, 0x00ff0fff,
826         0x8b10, 0xffffffff, 0x00000000,
827         0x28a4c, 0x06000000, 0x06000000,
828         0x4d8, 0xffffffff, 0x00000100,
829         0x913c, 0xffff000f, 0x0100000a,
830         0x960c, 0xffffffff, 0x54763210,
831         0x88c4, 0xffffffff, 0x000000c2,
832         0x88d4, 0xffffffff, 0x00000010,
833         0x8974, 0xffffffff, 0x00000000,
834         0xc78, 0x00000080, 0x00000080,
835         0x5e78, 0xffffffff, 0x001000f0,
836         0xd02c, 0xffffffff, 0x08421000,
837         0xa008, 0xffffffff, 0x00010000,
838         0x8d00, 0xffffffff, 0x100e4848,
839         0x8d04, 0xffffffff, 0x00164745,
840         0x8c00, 0xffffffff, 0xe4000003,
841         0x8cf0, 0x1fffffff, 0x08e00410,
842         0x28350, 0xffffffff, 0x00000000,
843         0x9508, 0xffffffff, 0x00000002,
844         0x900c, 0xffffffff, 0x0017071f,
845         0x8c18, 0xffffffff, 0x10101060,
846         0x8c1c, 0xffffffff, 0x00001010
847 };
848
849 static const u32 barts_golden_registers[] =
850 {
851         0x5eb4, 0xffffffff, 0x00000002,
852         0x5e78, 0x8f311ff1, 0x001000f0,
853         0x3f90, 0xffff0000, 0xff000000,
854         0x9148, 0xffff0000, 0xff000000,
855         0x3f94, 0xffff0000, 0xff000000,
856         0x914c, 0xffff0000, 0xff000000,
857         0xc78, 0x00000080, 0x00000080,
858         0xbd4, 0x70073777, 0x00010001,
859         0xd02c, 0xbfffff1f, 0x08421000,
860         0xd0b8, 0x03773777, 0x02011003,
861         0x5bc0, 0x00200000, 0x50100000,
862         0x98f8, 0x33773777, 0x02011003,
863         0x98fc, 0xffffffff, 0x76543210,
864         0x7030, 0x31000311, 0x00000011,
865         0x2f48, 0x00000007, 0x02011003,
866         0x6b28, 0x00000010, 0x00000012,
867         0x7728, 0x00000010, 0x00000012,
868         0x10328, 0x00000010, 0x00000012,
869         0x10f28, 0x00000010, 0x00000012,
870         0x11b28, 0x00000010, 0x00000012,
871         0x12728, 0x00000010, 0x00000012,
872         0x240c, 0x000007ff, 0x00000380,
873         0x8a14, 0xf000001f, 0x00000007,
874         0x8b24, 0x3fff3fff, 0x00ff0fff,
875         0x8b10, 0x0000ff0f, 0x00000000,
876         0x28a4c, 0x07ffffff, 0x06000000,
877         0x10c, 0x00000001, 0x00010003,
878         0xa02c, 0xffffffff, 0x0000009b,
879         0x913c, 0x0000000f, 0x0100000a,
880         0x8d00, 0xffff7f7f, 0x100e4848,
881         0x8d04, 0x00ffffff, 0x00164745,
882         0x8c00, 0xfffc0003, 0xe4000003,
883         0x8c04, 0xf8ff00ff, 0x40600060,
884         0x8c08, 0x00ff00ff, 0x001c001c,
885         0x8cf0, 0x1fff1fff, 0x08e00620,
886         0x8c20, 0x0fff0fff, 0x00800080,
887         0x8c24, 0x0fff0fff, 0x00800080,
888         0x8c18, 0xffffffff, 0x20202078,
889         0x8c1c, 0x0000ffff, 0x00001010,
890         0x28350, 0x00000f01, 0x00000000,
891         0x9508, 0x3700001f, 0x00000002,
892         0x960c, 0xffffffff, 0x54763210,
893         0x88c4, 0x001f3ae3, 0x000000c2,
894         0x88d4, 0x0000001f, 0x00000010,
895         0x8974, 0xffffffff, 0x00000000
896 };
897
898 static const u32 turks_golden_registers[] =
899 {
900         0x5eb4, 0xffffffff, 0x00000002,
901         0x5e78, 0x8f311ff1, 0x001000f0,
902         0x8c8, 0x00003000, 0x00001070,
903         0x8cc, 0x000fffff, 0x00040035,
904         0x3f90, 0xffff0000, 0xfff00000,
905         0x9148, 0xffff0000, 0xfff00000,
906         0x3f94, 0xffff0000, 0xfff00000,
907         0x914c, 0xffff0000, 0xfff00000,
908         0xc78, 0x00000080, 0x00000080,
909         0xbd4, 0x00073007, 0x00010002,
910         0xd02c, 0xbfffff1f, 0x08421000,
911         0xd0b8, 0x03773777, 0x02010002,
912         0x5bc0, 0x00200000, 0x50100000,
913         0x98f8, 0x33773777, 0x00010002,
914         0x98fc, 0xffffffff, 0x33221100,
915         0x7030, 0x31000311, 0x00000011,
916         0x2f48, 0x33773777, 0x00010002,
917         0x6b28, 0x00000010, 0x00000012,
918         0x7728, 0x00000010, 0x00000012,
919         0x10328, 0x00000010, 0x00000012,
920         0x10f28, 0x00000010, 0x00000012,
921         0x11b28, 0x00000010, 0x00000012,
922         0x12728, 0x00000010, 0x00000012,
923         0x240c, 0x000007ff, 0x00000380,
924         0x8a14, 0xf000001f, 0x00000007,
925         0x8b24, 0x3fff3fff, 0x00ff0fff,
926         0x8b10, 0x0000ff0f, 0x00000000,
927         0x28a4c, 0x07ffffff, 0x06000000,
928         0x10c, 0x00000001, 0x00010003,
929         0xa02c, 0xffffffff, 0x0000009b,
930         0x913c, 0x0000000f, 0x0100000a,
931         0x8d00, 0xffff7f7f, 0x100e4848,
932         0x8d04, 0x00ffffff, 0x00164745,
933         0x8c00, 0xfffc0003, 0xe4000003,
934         0x8c04, 0xf8ff00ff, 0x40600060,
935         0x8c08, 0x00ff00ff, 0x001c001c,
936         0x8cf0, 0x1fff1fff, 0x08e00410,
937         0x8c20, 0x0fff0fff, 0x00800080,
938         0x8c24, 0x0fff0fff, 0x00800080,
939         0x8c18, 0xffffffff, 0x20202078,
940         0x8c1c, 0x0000ffff, 0x00001010,
941         0x28350, 0x00000f01, 0x00000000,
942         0x9508, 0x3700001f, 0x00000002,
943         0x960c, 0xffffffff, 0x54763210,
944         0x88c4, 0x001f3ae3, 0x000000c2,
945         0x88d4, 0x0000001f, 0x00000010,
946         0x8974, 0xffffffff, 0x00000000
947 };
948
949 static const u32 caicos_golden_registers[] =
950 {
951         0x5eb4, 0xffffffff, 0x00000002,
952         0x5e78, 0x8f311ff1, 0x001000f0,
953         0x8c8, 0x00003420, 0x00001450,
954         0x8cc, 0x000fffff, 0x00040035,
955         0x3f90, 0xffff0000, 0xfffc0000,
956         0x9148, 0xffff0000, 0xfffc0000,
957         0x3f94, 0xffff0000, 0xfffc0000,
958         0x914c, 0xffff0000, 0xfffc0000,
959         0xc78, 0x00000080, 0x00000080,
960         0xbd4, 0x00073007, 0x00010001,
961         0xd02c, 0xbfffff1f, 0x08421000,
962         0xd0b8, 0x03773777, 0x02010001,
963         0x5bc0, 0x00200000, 0x50100000,
964         0x98f8, 0x33773777, 0x02010001,
965         0x98fc, 0xffffffff, 0x33221100,
966         0x7030, 0x31000311, 0x00000011,
967         0x2f48, 0x33773777, 0x02010001,
968         0x6b28, 0x00000010, 0x00000012,
969         0x7728, 0x00000010, 0x00000012,
970         0x10328, 0x00000010, 0x00000012,
971         0x10f28, 0x00000010, 0x00000012,
972         0x11b28, 0x00000010, 0x00000012,
973         0x12728, 0x00000010, 0x00000012,
974         0x240c, 0x000007ff, 0x00000380,
975         0x8a14, 0xf000001f, 0x00000001,
976         0x8b24, 0x3fff3fff, 0x00ff0fff,
977         0x8b10, 0x0000ff0f, 0x00000000,
978         0x28a4c, 0x07ffffff, 0x06000000,
979         0x10c, 0x00000001, 0x00010003,
980         0xa02c, 0xffffffff, 0x0000009b,
981         0x913c, 0x0000000f, 0x0100000a,
982         0x8d00, 0xffff7f7f, 0x100e4848,
983         0x8d04, 0x00ffffff, 0x00164745,
984         0x8c00, 0xfffc0003, 0xe4000003,
985         0x8c04, 0xf8ff00ff, 0x40600060,
986         0x8c08, 0x00ff00ff, 0x001c001c,
987         0x8cf0, 0x1fff1fff, 0x08e00410,
988         0x8c20, 0x0fff0fff, 0x00800080,
989         0x8c24, 0x0fff0fff, 0x00800080,
990         0x8c18, 0xffffffff, 0x20202078,
991         0x8c1c, 0x0000ffff, 0x00001010,
992         0x28350, 0x00000f01, 0x00000000,
993         0x9508, 0x3700001f, 0x00000002,
994         0x960c, 0xffffffff, 0x54763210,
995         0x88c4, 0x001f3ae3, 0x000000c2,
996         0x88d4, 0x0000001f, 0x00000010,
997         0x8974, 0xffffffff, 0x00000000
998 };
999
1000 static void evergreen_init_golden_registers(struct radeon_device *rdev)
1001 {
1002         switch (rdev->family) {
1003         case CHIP_CYPRESS:
1004         case CHIP_HEMLOCK:
1005                 radeon_program_register_sequence(rdev,
1006                                                  evergreen_golden_registers,
1007                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers));
1008                 radeon_program_register_sequence(rdev,
1009                                                  evergreen_golden_registers2,
1010                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers2));
1011                 radeon_program_register_sequence(rdev,
1012                                                  cypress_mgcg_init,
1013                                                  (const u32)ARRAY_SIZE(cypress_mgcg_init));
1014                 break;
1015         case CHIP_JUNIPER:
1016                 radeon_program_register_sequence(rdev,
1017                                                  evergreen_golden_registers,
1018                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers));
1019                 radeon_program_register_sequence(rdev,
1020                                                  evergreen_golden_registers2,
1021                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers2));
1022                 radeon_program_register_sequence(rdev,
1023                                                  juniper_mgcg_init,
1024                                                  (const u32)ARRAY_SIZE(juniper_mgcg_init));
1025                 break;
1026         case CHIP_REDWOOD:
1027                 radeon_program_register_sequence(rdev,
1028                                                  evergreen_golden_registers,
1029                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers));
1030                 radeon_program_register_sequence(rdev,
1031                                                  evergreen_golden_registers2,
1032                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers2));
1033                 radeon_program_register_sequence(rdev,
1034                                                  redwood_mgcg_init,
1035                                                  (const u32)ARRAY_SIZE(redwood_mgcg_init));
1036                 break;
1037         case CHIP_CEDAR:
1038                 radeon_program_register_sequence(rdev,
1039                                                  cedar_golden_registers,
1040                                                  (const u32)ARRAY_SIZE(cedar_golden_registers));
1041                 radeon_program_register_sequence(rdev,
1042                                                  evergreen_golden_registers2,
1043                                                  (const u32)ARRAY_SIZE(evergreen_golden_registers2));
1044                 radeon_program_register_sequence(rdev,
1045                                                  cedar_mgcg_init,
1046                                                  (const u32)ARRAY_SIZE(cedar_mgcg_init));
1047                 break;
1048         case CHIP_PALM:
1049                 radeon_program_register_sequence(rdev,
1050                                                  wrestler_golden_registers,
1051                                                  (const u32)ARRAY_SIZE(wrestler_golden_registers));
1052                 break;
1053         case CHIP_SUMO:
1054                 radeon_program_register_sequence(rdev,
1055                                                  supersumo_golden_registers,
1056                                                  (const u32)ARRAY_SIZE(supersumo_golden_registers));
1057                 break;
1058         case CHIP_SUMO2:
1059                 radeon_program_register_sequence(rdev,
1060                                                  supersumo_golden_registers,
1061                                                  (const u32)ARRAY_SIZE(supersumo_golden_registers));
1062                 radeon_program_register_sequence(rdev,
1063                                                  sumo_golden_registers,
1064                                                  (const u32)ARRAY_SIZE(sumo_golden_registers));
1065                 break;
1066         case CHIP_BARTS:
1067                 radeon_program_register_sequence(rdev,
1068                                                  barts_golden_registers,
1069                                                  (const u32)ARRAY_SIZE(barts_golden_registers));
1070                 break;
1071         case CHIP_TURKS:
1072                 radeon_program_register_sequence(rdev,
1073                                                  turks_golden_registers,
1074                                                  (const u32)ARRAY_SIZE(turks_golden_registers));
1075                 break;
1076         case CHIP_CAICOS:
1077                 radeon_program_register_sequence(rdev,
1078                                                  caicos_golden_registers,
1079                                                  (const u32)ARRAY_SIZE(caicos_golden_registers));
1080                 break;
1081         default:
1082                 break;
1083         }
1084 }
1085
1086 /**
1087  * evergreen_get_allowed_info_register - fetch the register for the info ioctl
1088  *
1089  * @rdev: radeon_device pointer
1090  * @reg: register offset in bytes
1091  * @val: register value
1092  *
1093  * Returns 0 for success or -EINVAL for an invalid register
1094  *
1095  */
1096 int evergreen_get_allowed_info_register(struct radeon_device *rdev,
1097                                         u32 reg, u32 *val)
1098 {
1099         switch (reg) {
1100         case GRBM_STATUS:
1101         case GRBM_STATUS_SE0:
1102         case GRBM_STATUS_SE1:
1103         case SRBM_STATUS:
1104         case SRBM_STATUS2:
1105         case DMA_STATUS_REG:
1106         case UVD_STATUS:
1107                 *val = RREG32(reg);
1108                 return 0;
1109         default:
1110                 return -EINVAL;
1111         }
1112 }
1113
1114 void evergreen_tiling_fields(unsigned tiling_flags, unsigned *bankw,
1115                              unsigned *bankh, unsigned *mtaspect,
1116                              unsigned *tile_split)
1117 {
1118         *bankw = (tiling_flags >> RADEON_TILING_EG_BANKW_SHIFT) & RADEON_TILING_EG_BANKW_MASK;
1119         *bankh = (tiling_flags >> RADEON_TILING_EG_BANKH_SHIFT) & RADEON_TILING_EG_BANKH_MASK;
1120         *mtaspect = (tiling_flags >> RADEON_TILING_EG_MACRO_TILE_ASPECT_SHIFT) & RADEON_TILING_EG_MACRO_TILE_ASPECT_MASK;
1121         *tile_split = (tiling_flags >> RADEON_TILING_EG_TILE_SPLIT_SHIFT) & RADEON_TILING_EG_TILE_SPLIT_MASK;
1122         switch (*bankw) {
1123         default:
1124         case 1: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_1; break;
1125         case 2: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_2; break;
1126         case 4: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_4; break;
1127         case 8: *bankw = EVERGREEN_ADDR_SURF_BANK_WIDTH_8; break;
1128         }
1129         switch (*bankh) {
1130         default:
1131         case 1: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_1; break;
1132         case 2: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_2; break;
1133         case 4: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_4; break;
1134         case 8: *bankh = EVERGREEN_ADDR_SURF_BANK_HEIGHT_8; break;
1135         }
1136         switch (*mtaspect) {
1137         default:
1138         case 1: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_1; break;
1139         case 2: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_2; break;
1140         case 4: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_4; break;
1141         case 8: *mtaspect = EVERGREEN_ADDR_SURF_MACRO_TILE_ASPECT_8; break;
1142         }
1143 }
1144
1145 static int sumo_set_uvd_clock(struct radeon_device *rdev, u32 clock,
1146                               u32 cntl_reg, u32 status_reg)
1147 {
1148         int r, i;
1149         struct atom_clock_dividers dividers;
1150
1151         r = radeon_atom_get_clock_dividers(rdev, COMPUTE_ENGINE_PLL_PARAM,
1152                                            clock, false, &dividers);
1153         if (r)
1154                 return r;
1155
1156         WREG32_P(cntl_reg, dividers.post_div, ~(DCLK_DIR_CNTL_EN|DCLK_DIVIDER_MASK));
1157
1158         for (i = 0; i < 100; i++) {
1159                 if (RREG32(status_reg) & DCLK_STATUS)
1160                         break;
1161                 mdelay(10);
1162         }
1163         if (i == 100)
1164                 return -ETIMEDOUT;
1165
1166         return 0;
1167 }
1168
1169 int sumo_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk)
1170 {
1171         int r = 0;
1172         u32 cg_scratch = RREG32(CG_SCRATCH1);
1173
1174         r = sumo_set_uvd_clock(rdev, vclk, CG_VCLK_CNTL, CG_VCLK_STATUS);
1175         if (r)
1176                 goto done;
1177         cg_scratch &= 0xffff0000;
1178         cg_scratch |= vclk / 100; /* Mhz */
1179
1180         r = sumo_set_uvd_clock(rdev, dclk, CG_DCLK_CNTL, CG_DCLK_STATUS);
1181         if (r)
1182                 goto done;
1183         cg_scratch &= 0x0000ffff;
1184         cg_scratch |= (dclk / 100) << 16; /* Mhz */
1185
1186 done:
1187         WREG32(CG_SCRATCH1, cg_scratch);
1188
1189         return r;
1190 }
1191
1192 int evergreen_set_uvd_clocks(struct radeon_device *rdev, u32 vclk, u32 dclk)
1193 {
1194         /* start off with something large */
1195         unsigned fb_div = 0, vclk_div = 0, dclk_div = 0;
1196         int r;
1197
1198         /* bypass vclk and dclk with bclk */
1199         WREG32_P(CG_UPLL_FUNC_CNTL_2,
1200                 VCLK_SRC_SEL(1) | DCLK_SRC_SEL(1),
1201                 ~(VCLK_SRC_SEL_MASK | DCLK_SRC_SEL_MASK));
1202
1203         /* put PLL in bypass mode */
1204         WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_BYPASS_EN_MASK, ~UPLL_BYPASS_EN_MASK);
1205
1206         if (!vclk || !dclk) {
1207                 /* keep the Bypass mode, put PLL to sleep */
1208                 WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_SLEEP_MASK, ~UPLL_SLEEP_MASK);
1209                 return 0;
1210         }
1211
1212         r = radeon_uvd_calc_upll_dividers(rdev, vclk, dclk, 125000, 250000,
1213                                           16384, 0x03FFFFFF, 0, 128, 5,
1214                                           &fb_div, &vclk_div, &dclk_div);
1215         if (r)
1216                 return r;
1217
1218         /* set VCO_MODE to 1 */
1219         WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_VCO_MODE_MASK, ~UPLL_VCO_MODE_MASK);
1220
1221         /* toggle UPLL_SLEEP to 1 then back to 0 */
1222         WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_SLEEP_MASK, ~UPLL_SLEEP_MASK);
1223         WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_SLEEP_MASK);
1224
1225         /* deassert UPLL_RESET */
1226         WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_RESET_MASK);
1227
1228         mdelay(1);
1229
1230         r = radeon_uvd_send_upll_ctlreq(rdev, CG_UPLL_FUNC_CNTL);
1231         if (r)
1232                 return r;
1233
1234         /* assert UPLL_RESET again */
1235         WREG32_P(CG_UPLL_FUNC_CNTL, UPLL_RESET_MASK, ~UPLL_RESET_MASK);
1236
1237         /* disable spread spectrum. */
1238         WREG32_P(CG_UPLL_SPREAD_SPECTRUM, 0, ~SSEN_MASK);
1239
1240         /* set feedback divider */
1241         WREG32_P(CG_UPLL_FUNC_CNTL_3, UPLL_FB_DIV(fb_div), ~UPLL_FB_DIV_MASK);
1242
1243         /* set ref divider to 0 */
1244         WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_REF_DIV_MASK);
1245
1246         if (fb_div < 307200)
1247                 WREG32_P(CG_UPLL_FUNC_CNTL_4, 0, ~UPLL_SPARE_ISPARE9);
1248         else
1249                 WREG32_P(CG_UPLL_FUNC_CNTL_4, UPLL_SPARE_ISPARE9, ~UPLL_SPARE_ISPARE9);
1250
1251         /* set PDIV_A and PDIV_B */
1252         WREG32_P(CG_UPLL_FUNC_CNTL_2,
1253                 UPLL_PDIV_A(vclk_div) | UPLL_PDIV_B(dclk_div),
1254                 ~(UPLL_PDIV_A_MASK | UPLL_PDIV_B_MASK));
1255
1256         /* give the PLL some time to settle */
1257         mdelay(15);
1258
1259         /* deassert PLL_RESET */
1260         WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_RESET_MASK);
1261
1262         mdelay(15);
1263
1264         /* switch from bypass mode to normal mode */
1265         WREG32_P(CG_UPLL_FUNC_CNTL, 0, ~UPLL_BYPASS_EN_MASK);
1266
1267         r = radeon_uvd_send_upll_ctlreq(rdev, CG_UPLL_FUNC_CNTL);
1268         if (r)
1269                 return r;
1270
1271         /* switch VCLK and DCLK selection */
1272         WREG32_P(CG_UPLL_FUNC_CNTL_2,
1273                 VCLK_SRC_SEL(2) | DCLK_SRC_SEL(2),
1274                 ~(VCLK_SRC_SEL_MASK | DCLK_SRC_SEL_MASK));
1275
1276         mdelay(100);
1277
1278         return 0;
1279 }
1280
1281 void evergreen_fix_pci_max_read_req_size(struct radeon_device *rdev)
1282 {
1283         int readrq;
1284         u16 v;
1285
1286         readrq = pcie_get_readrq(rdev->pdev);
1287         v = ffs(readrq) - 8;
1288         /* if bios or OS sets MAX_READ_REQUEST_SIZE to an invalid value, fix it
1289          * to avoid hangs or perfomance issues
1290          */
1291         if ((v == 0) || (v == 6) || (v == 7))
1292                 pcie_set_readrq(rdev->pdev, 512);
1293 }
1294
1295 void dce4_program_fmt(struct drm_encoder *encoder)
1296 {
1297         struct drm_device *dev = encoder->dev;
1298         struct radeon_device *rdev = dev->dev_private;
1299         struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder);
1300         struct radeon_crtc *radeon_crtc = to_radeon_crtc(encoder->crtc);
1301         struct drm_connector *connector = radeon_get_connector_for_encoder(encoder);
1302         int bpc = 0;
1303         u32 tmp = 0;
1304         enum radeon_connector_dither dither = RADEON_FMT_DITHER_DISABLE;
1305
1306         if (connector) {
1307                 struct radeon_connector *radeon_connector = to_radeon_connector(connector);
1308                 bpc = radeon_get_monitor_bpc(connector);
1309                 dither = radeon_connector->dither;
1310         }
1311
1312         /* LVDS/eDP FMT is set up by atom */
1313         if (radeon_encoder->devices & ATOM_DEVICE_LCD_SUPPORT)
1314                 return;
1315
1316         /* not needed for analog */
1317         if ((radeon_encoder->encoder_id == ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC1) ||
1318             (radeon_encoder->encoder_id == ENCODER_OBJECT_ID_INTERNAL_KLDSCP_DAC2))
1319                 return;
1320
1321         if (bpc == 0)
1322                 return;
1323
1324         switch (bpc) {
1325         case 6:
1326                 if (dither == RADEON_FMT_DITHER_ENABLE)
1327                         /* XXX sort out optimal dither settings */
1328                         tmp |= (FMT_FRAME_RANDOM_ENABLE | FMT_HIGHPASS_RANDOM_ENABLE |
1329                                 FMT_SPATIAL_DITHER_EN);
1330                 else
1331                         tmp |= FMT_TRUNCATE_EN;
1332                 break;
1333         case 8:
1334                 if (dither == RADEON_FMT_DITHER_ENABLE)
1335                         /* XXX sort out optimal dither settings */
1336                         tmp |= (FMT_FRAME_RANDOM_ENABLE | FMT_HIGHPASS_RANDOM_ENABLE |
1337                                 FMT_RGB_RANDOM_ENABLE |
1338                                 FMT_SPATIAL_DITHER_EN | FMT_SPATIAL_DITHER_DEPTH);
1339                 else
1340                         tmp |= (FMT_TRUNCATE_EN | FMT_TRUNCATE_DEPTH);
1341                 break;
1342         case 10:
1343         default:
1344                 /* not needed */
1345                 break;
1346         }
1347
1348         WREG32(FMT_BIT_DEPTH_CONTROL + radeon_crtc->crtc_offset, tmp);
1349 }
1350
1351 static bool dce4_is_in_vblank(struct radeon_device *rdev, int crtc)
1352 {
1353         if (RREG32(EVERGREEN_CRTC_STATUS + crtc_offsets[crtc]) & EVERGREEN_CRTC_V_BLANK)
1354                 return true;
1355         else
1356                 return false;
1357 }
1358
1359 static bool dce4_is_counter_moving(struct radeon_device *rdev, int crtc)
1360 {
1361         u32 pos1, pos2;
1362
1363         pos1 = RREG32(EVERGREEN_CRTC_STATUS_POSITION + crtc_offsets[crtc]);
1364         pos2 = RREG32(EVERGREEN_CRTC_STATUS_POSITION + crtc_offsets[crtc]);
1365
1366         if (pos1 != pos2)
1367                 return true;
1368         else
1369                 return false;
1370 }
1371
1372 /**
1373  * dce4_wait_for_vblank - vblank wait asic callback.
1374  *
1375  * @rdev: radeon_device pointer
1376  * @crtc: crtc to wait for vblank on
1377  *
1378  * Wait for vblank on the requested crtc (evergreen+).
1379  */
1380 void dce4_wait_for_vblank(struct radeon_device *rdev, int crtc)
1381 {
1382         unsigned i = 0;
1383
1384         if (crtc >= rdev->num_crtc)
1385                 return;
1386
1387         if (!(RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[crtc]) & EVERGREEN_CRTC_MASTER_EN))
1388                 return;
1389
1390         /* depending on when we hit vblank, we may be close to active; if so,
1391          * wait for another frame.
1392          */
1393         while (dce4_is_in_vblank(rdev, crtc)) {
1394                 if (i++ % 100 == 0) {
1395                         if (!dce4_is_counter_moving(rdev, crtc))
1396                                 break;
1397                 }
1398         }
1399
1400         while (!dce4_is_in_vblank(rdev, crtc)) {
1401                 if (i++ % 100 == 0) {
1402                         if (!dce4_is_counter_moving(rdev, crtc))
1403                                 break;
1404                 }
1405         }
1406 }
1407
1408 /**
1409  * evergreen_page_flip - pageflip callback.
1410  *
1411  * @rdev: radeon_device pointer
1412  * @crtc_id: crtc to cleanup pageflip on
1413  * @crtc_base: new address of the crtc (GPU MC address)
1414  *
1415  * Triggers the actual pageflip by updating the primary
1416  * surface base address (evergreen+).
1417  */
1418 void evergreen_page_flip(struct radeon_device *rdev, int crtc_id, u64 crtc_base,
1419                          bool async)
1420 {
1421         struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
1422
1423         /* update the scanout addresses */
1424         WREG32(EVERGREEN_GRPH_FLIP_CONTROL + radeon_crtc->crtc_offset,
1425                async ? EVERGREEN_GRPH_SURFACE_UPDATE_H_RETRACE_EN : 0);
1426         WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS_HIGH + radeon_crtc->crtc_offset,
1427                upper_32_bits(crtc_base));
1428         WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset,
1429                (u32)crtc_base);
1430         /* post the write */
1431         RREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS + radeon_crtc->crtc_offset);
1432 }
1433
1434 /**
1435  * evergreen_page_flip_pending - check if page flip is still pending
1436  *
1437  * @rdev: radeon_device pointer
1438  * @crtc_id: crtc to check
1439  *
1440  * Returns the current update pending status.
1441  */
1442 bool evergreen_page_flip_pending(struct radeon_device *rdev, int crtc_id)
1443 {
1444         struct radeon_crtc *radeon_crtc = rdev->mode_info.crtcs[crtc_id];
1445
1446         /* Return current update_pending status: */
1447         return !!(RREG32(EVERGREEN_GRPH_UPDATE + radeon_crtc->crtc_offset) &
1448                 EVERGREEN_GRPH_SURFACE_UPDATE_PENDING);
1449 }
1450
1451 /* get temperature in millidegrees */
1452 int evergreen_get_temp(struct radeon_device *rdev)
1453 {
1454         u32 temp, toffset;
1455         int actual_temp = 0;
1456
1457         if (rdev->family == CHIP_JUNIPER) {
1458                 toffset = (RREG32(CG_THERMAL_CTRL) & TOFFSET_MASK) >>
1459                         TOFFSET_SHIFT;
1460                 temp = (RREG32(CG_TS0_STATUS) & TS0_ADC_DOUT_MASK) >>
1461                         TS0_ADC_DOUT_SHIFT;
1462
1463                 if (toffset & 0x100)
1464                         actual_temp = temp / 2 - (0x200 - toffset);
1465                 else
1466                         actual_temp = temp / 2 + toffset;
1467
1468                 actual_temp = actual_temp * 1000;
1469
1470         } else {
1471                 temp = (RREG32(CG_MULT_THERMAL_STATUS) & ASIC_T_MASK) >>
1472                         ASIC_T_SHIFT;
1473
1474                 if (temp & 0x400)
1475                         actual_temp = -256;
1476                 else if (temp & 0x200)
1477                         actual_temp = 255;
1478                 else if (temp & 0x100) {
1479                         actual_temp = temp & 0x1ff;
1480                         actual_temp |= ~0x1ff;
1481                 } else
1482                         actual_temp = temp & 0xff;
1483
1484                 actual_temp = (actual_temp * 1000) / 2;
1485         }
1486
1487         return actual_temp;
1488 }
1489
1490 int sumo_get_temp(struct radeon_device *rdev)
1491 {
1492         u32 temp = RREG32(CG_THERMAL_STATUS) & 0xff;
1493         int actual_temp = temp - 49;
1494
1495         return actual_temp * 1000;
1496 }
1497
1498 /**
1499  * sumo_pm_init_profile - Initialize power profiles callback.
1500  *
1501  * @rdev: radeon_device pointer
1502  *
1503  * Initialize the power states used in profile mode
1504  * (sumo, trinity, SI).
1505  * Used for profile mode only.
1506  */
1507 void sumo_pm_init_profile(struct radeon_device *rdev)
1508 {
1509         int idx;
1510
1511         /* default */
1512         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_ps_idx = rdev->pm.default_power_state_index;
1513         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_ps_idx = rdev->pm.default_power_state_index;
1514         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_cm_idx = 0;
1515         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_cm_idx = 0;
1516
1517         /* low,mid sh/mh */
1518         if (rdev->flags & RADEON_IS_MOBILITY)
1519                 idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0);
1520         else
1521                 idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0);
1522
1523         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx;
1524         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx;
1525         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_cm_idx = 0;
1526         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_cm_idx = 0;
1527
1528         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx;
1529         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx;
1530         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_cm_idx = 0;
1531         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_cm_idx = 0;
1532
1533         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx;
1534         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx;
1535         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_cm_idx = 0;
1536         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_cm_idx = 0;
1537
1538         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx;
1539         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx;
1540         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_cm_idx = 0;
1541         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_cm_idx = 0;
1542
1543         /* high sh/mh */
1544         idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0);
1545         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx;
1546         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx;
1547         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_cm_idx = 0;
1548         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_cm_idx =
1549                 rdev->pm.power_state[idx].num_clock_modes - 1;
1550
1551         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx;
1552         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx;
1553         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_cm_idx = 0;
1554         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_cm_idx =
1555                 rdev->pm.power_state[idx].num_clock_modes - 1;
1556 }
1557
1558 /**
1559  * btc_pm_init_profile - Initialize power profiles callback.
1560  *
1561  * @rdev: radeon_device pointer
1562  *
1563  * Initialize the power states used in profile mode
1564  * (BTC, cayman).
1565  * Used for profile mode only.
1566  */
1567 void btc_pm_init_profile(struct radeon_device *rdev)
1568 {
1569         int idx;
1570
1571         /* default */
1572         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_ps_idx = rdev->pm.default_power_state_index;
1573         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_ps_idx = rdev->pm.default_power_state_index;
1574         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_off_cm_idx = 0;
1575         rdev->pm.profiles[PM_PROFILE_DEFAULT_IDX].dpms_on_cm_idx = 2;
1576         /* starting with BTC, there is one state that is used for both
1577          * MH and SH.  Difference is that we always use the high clock index for
1578          * mclk.
1579          */
1580         if (rdev->flags & RADEON_IS_MOBILITY)
1581                 idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0);
1582         else
1583                 idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0);
1584         /* low sh */
1585         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx;
1586         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx;
1587         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_cm_idx = 0;
1588         rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_cm_idx = 0;
1589         /* mid sh */
1590         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx;
1591         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx;
1592         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_cm_idx = 0;
1593         rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_cm_idx = 1;
1594         /* high sh */
1595         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx;
1596         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx;
1597         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_cm_idx = 0;
1598         rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_cm_idx = 2;
1599         /* low mh */
1600         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx;
1601         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx;
1602         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_cm_idx = 0;
1603         rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_cm_idx = 0;
1604         /* mid mh */
1605         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx;
1606         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx;
1607         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_cm_idx = 0;
1608         rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_cm_idx = 1;
1609         /* high mh */
1610         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx;
1611         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx;
1612         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_cm_idx = 0;
1613         rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_cm_idx = 2;
1614 }
1615
1616 /**
1617  * evergreen_pm_misc - set additional pm hw parameters callback.
1618  *
1619  * @rdev: radeon_device pointer
1620  *
1621  * Set non-clock parameters associated with a power state
1622  * (voltage, etc.) (evergreen+).
1623  */
1624 void evergreen_pm_misc(struct radeon_device *rdev)
1625 {
1626         int req_ps_idx = rdev->pm.requested_power_state_index;
1627         int req_cm_idx = rdev->pm.requested_clock_mode_index;
1628         struct radeon_power_state *ps = &rdev->pm.power_state[req_ps_idx];
1629         struct radeon_voltage *voltage = &ps->clock_info[req_cm_idx].voltage;
1630
1631         if (voltage->type == VOLTAGE_SW) {
1632                 /* 0xff0x are flags rather then an actual voltage */
1633                 if ((voltage->voltage & 0xff00) == 0xff00)
1634                         return;
1635                 if (voltage->voltage && (voltage->voltage != rdev->pm.current_vddc)) {
1636                         radeon_atom_set_voltage(rdev, voltage->voltage, SET_VOLTAGE_TYPE_ASIC_VDDC);
1637                         rdev->pm.current_vddc = voltage->voltage;
1638                         DRM_DEBUG("Setting: vddc: %d\n", voltage->voltage);
1639                 }
1640
1641                 /* starting with BTC, there is one state that is used for both
1642                  * MH and SH.  Difference is that we always use the high clock index for
1643                  * mclk and vddci.
1644                  */
1645                 if ((rdev->pm.pm_method == PM_METHOD_PROFILE) &&
1646                     (rdev->family >= CHIP_BARTS) &&
1647                     rdev->pm.active_crtc_count &&
1648                     ((rdev->pm.profile_index == PM_PROFILE_MID_MH_IDX) ||
1649                      (rdev->pm.profile_index == PM_PROFILE_LOW_MH_IDX)))
1650                         voltage = &rdev->pm.power_state[req_ps_idx].
1651                                 clock_info[rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_cm_idx].voltage;
1652
1653                 /* 0xff0x are flags rather then an actual voltage */
1654                 if ((voltage->vddci & 0xff00) == 0xff00)
1655                         return;
1656                 if (voltage->vddci && (voltage->vddci != rdev->pm.current_vddci)) {
1657                         radeon_atom_set_voltage(rdev, voltage->vddci, SET_VOLTAGE_TYPE_ASIC_VDDCI);
1658                         rdev->pm.current_vddci = voltage->vddci;
1659                         DRM_DEBUG("Setting: vddci: %d\n", voltage->vddci);
1660                 }
1661         }
1662 }
1663
1664 /**
1665  * evergreen_pm_prepare - pre-power state change callback.
1666  *
1667  * @rdev: radeon_device pointer
1668  *
1669  * Prepare for a power state change (evergreen+).
1670  */
1671 void evergreen_pm_prepare(struct radeon_device *rdev)
1672 {
1673         struct drm_device *ddev = rdev->ddev;
1674         struct drm_crtc *crtc;
1675         struct radeon_crtc *radeon_crtc;
1676         u32 tmp;
1677
1678         /* disable any active CRTCs */
1679         list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
1680                 radeon_crtc = to_radeon_crtc(crtc);
1681                 if (radeon_crtc->enabled) {
1682                         tmp = RREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset);
1683                         tmp |= EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
1684                         WREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset, tmp);
1685                 }
1686         }
1687 }
1688
1689 /**
1690  * evergreen_pm_finish - post-power state change callback.
1691  *
1692  * @rdev: radeon_device pointer
1693  *
1694  * Clean up after a power state change (evergreen+).
1695  */
1696 void evergreen_pm_finish(struct radeon_device *rdev)
1697 {
1698         struct drm_device *ddev = rdev->ddev;
1699         struct drm_crtc *crtc;
1700         struct radeon_crtc *radeon_crtc;
1701         u32 tmp;
1702
1703         /* enable any active CRTCs */
1704         list_for_each_entry(crtc, &ddev->mode_config.crtc_list, head) {
1705                 radeon_crtc = to_radeon_crtc(crtc);
1706                 if (radeon_crtc->enabled) {
1707                         tmp = RREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset);
1708                         tmp &= ~EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
1709                         WREG32(EVERGREEN_CRTC_CONTROL + radeon_crtc->crtc_offset, tmp);
1710                 }
1711         }
1712 }
1713
1714 /**
1715  * evergreen_hpd_sense - hpd sense callback.
1716  *
1717  * @rdev: radeon_device pointer
1718  * @hpd: hpd (hotplug detect) pin
1719  *
1720  * Checks if a digital monitor is connected (evergreen+).
1721  * Returns true if connected, false if not connected.
1722  */
1723 bool evergreen_hpd_sense(struct radeon_device *rdev, enum radeon_hpd_id hpd)
1724 {
1725         if (hpd == RADEON_HPD_NONE)
1726                 return false;
1727
1728         return !!(RREG32(DC_HPDx_INT_STATUS_REG(hpd)) & DC_HPDx_SENSE);
1729 }
1730
1731 /**
1732  * evergreen_hpd_set_polarity - hpd set polarity callback.
1733  *
1734  * @rdev: radeon_device pointer
1735  * @hpd: hpd (hotplug detect) pin
1736  *
1737  * Set the polarity of the hpd pin (evergreen+).
1738  */
1739 void evergreen_hpd_set_polarity(struct radeon_device *rdev,
1740                                 enum radeon_hpd_id hpd)
1741 {
1742         bool connected = evergreen_hpd_sense(rdev, hpd);
1743
1744         if (hpd == RADEON_HPD_NONE)
1745                 return;
1746
1747         if (connected)
1748                 WREG32_AND(DC_HPDx_INT_CONTROL(hpd), ~DC_HPDx_INT_POLARITY);
1749         else
1750                 WREG32_OR(DC_HPDx_INT_CONTROL(hpd), DC_HPDx_INT_POLARITY);
1751 }
1752
1753 /**
1754  * evergreen_hpd_init - hpd setup callback.
1755  *
1756  * @rdev: radeon_device pointer
1757  *
1758  * Setup the hpd pins used by the card (evergreen+).
1759  * Enable the pin, set the polarity, and enable the hpd interrupts.
1760  */
1761 void evergreen_hpd_init(struct radeon_device *rdev)
1762 {
1763         struct drm_device *dev = rdev->ddev;
1764         struct drm_connector *connector;
1765         unsigned enabled = 0;
1766         u32 tmp = DC_HPDx_CONNECTION_TIMER(0x9c4) |
1767                 DC_HPDx_RX_INT_TIMER(0xfa) | DC_HPDx_EN;
1768
1769         list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
1770                 enum radeon_hpd_id hpd =
1771                         to_radeon_connector(connector)->hpd.hpd;
1772
1773                 if (connector->connector_type == DRM_MODE_CONNECTOR_eDP ||
1774                     connector->connector_type == DRM_MODE_CONNECTOR_LVDS) {
1775                         /* don't try to enable hpd on eDP or LVDS avoid breaking the
1776                          * aux dp channel on imac and help (but not completely fix)
1777                          * https://bugzilla.redhat.com/show_bug.cgi?id=726143
1778                          * also avoid interrupt storms during dpms.
1779                          */
1780                         continue;
1781                 }
1782
1783                 if (hpd == RADEON_HPD_NONE)
1784                         continue;
1785
1786                 WREG32(DC_HPDx_CONTROL(hpd), tmp);
1787                 enabled |= 1 << hpd;
1788
1789                 radeon_hpd_set_polarity(rdev, hpd);
1790         }
1791         radeon_irq_kms_enable_hpd(rdev, enabled);
1792 }
1793
1794 /**
1795  * evergreen_hpd_fini - hpd tear down callback.
1796  *
1797  * @rdev: radeon_device pointer
1798  *
1799  * Tear down the hpd pins used by the card (evergreen+).
1800  * Disable the hpd interrupts.
1801  */
1802 void evergreen_hpd_fini(struct radeon_device *rdev)
1803 {
1804         struct drm_device *dev = rdev->ddev;
1805         struct drm_connector *connector;
1806         unsigned disabled = 0;
1807
1808         list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
1809                 enum radeon_hpd_id hpd =
1810                         to_radeon_connector(connector)->hpd.hpd;
1811
1812                 if (hpd == RADEON_HPD_NONE)
1813                         continue;
1814
1815                 WREG32(DC_HPDx_CONTROL(hpd), 0);
1816                 disabled |= 1 << hpd;
1817         }
1818         radeon_irq_kms_disable_hpd(rdev, disabled);
1819 }
1820
1821 /* watermark setup */
1822
1823 static u32 evergreen_line_buffer_adjust(struct radeon_device *rdev,
1824                                         struct radeon_crtc *radeon_crtc,
1825                                         struct drm_display_mode *mode,
1826                                         struct drm_display_mode *other_mode)
1827 {
1828         u32 tmp, buffer_alloc, i;
1829         u32 pipe_offset = radeon_crtc->crtc_id * 0x20;
1830         /*
1831          * Line Buffer Setup
1832          * There are 3 line buffers, each one shared by 2 display controllers.
1833          * DC_LB_MEMORY_SPLIT controls how that line buffer is shared between
1834          * the display controllers.  The paritioning is done via one of four
1835          * preset allocations specified in bits 2:0:
1836          * first display controller
1837          *  0 - first half of lb (3840 * 2)
1838          *  1 - first 3/4 of lb (5760 * 2)
1839          *  2 - whole lb (7680 * 2), other crtc must be disabled
1840          *  3 - first 1/4 of lb (1920 * 2)
1841          * second display controller
1842          *  4 - second half of lb (3840 * 2)
1843          *  5 - second 3/4 of lb (5760 * 2)
1844          *  6 - whole lb (7680 * 2), other crtc must be disabled
1845          *  7 - last 1/4 of lb (1920 * 2)
1846          */
1847         /* this can get tricky if we have two large displays on a paired group
1848          * of crtcs.  Ideally for multiple large displays we'd assign them to
1849          * non-linked crtcs for maximum line buffer allocation.
1850          */
1851         if (radeon_crtc->base.enabled && mode) {
1852                 if (other_mode) {
1853                         tmp = 0; /* 1/2 */
1854                         buffer_alloc = 1;
1855                 } else {
1856                         tmp = 2; /* whole */
1857                         buffer_alloc = 2;
1858                 }
1859         } else {
1860                 tmp = 0;
1861                 buffer_alloc = 0;
1862         }
1863
1864         /* second controller of the pair uses second half of the lb */
1865         if (radeon_crtc->crtc_id % 2)
1866                 tmp += 4;
1867         WREG32(DC_LB_MEMORY_SPLIT + radeon_crtc->crtc_offset, tmp);
1868
1869         if (ASIC_IS_DCE41(rdev) || ASIC_IS_DCE5(rdev)) {
1870                 WREG32(PIPE0_DMIF_BUFFER_CONTROL + pipe_offset,
1871                        DMIF_BUFFERS_ALLOCATED(buffer_alloc));
1872                 for (i = 0; i < rdev->usec_timeout; i++) {
1873                         if (RREG32(PIPE0_DMIF_BUFFER_CONTROL + pipe_offset) &
1874                             DMIF_BUFFERS_ALLOCATED_COMPLETED)
1875                                 break;
1876                         udelay(1);
1877                 }
1878         }
1879
1880         if (radeon_crtc->base.enabled && mode) {
1881                 switch (tmp) {
1882                 case 0:
1883                 case 4:
1884                 default:
1885                         if (ASIC_IS_DCE5(rdev))
1886                                 return 4096 * 2;
1887                         else
1888                                 return 3840 * 2;
1889                 case 1:
1890                 case 5:
1891                         if (ASIC_IS_DCE5(rdev))
1892                                 return 6144 * 2;
1893                         else
1894                                 return 5760 * 2;
1895                 case 2:
1896                 case 6:
1897                         if (ASIC_IS_DCE5(rdev))
1898                                 return 8192 * 2;
1899                         else
1900                                 return 7680 * 2;
1901                 case 3:
1902                 case 7:
1903                         if (ASIC_IS_DCE5(rdev))
1904                                 return 2048 * 2;
1905                         else
1906                                 return 1920 * 2;
1907                 }
1908         }
1909
1910         /* controller not enabled, so no lb used */
1911         return 0;
1912 }
1913
1914 u32 evergreen_get_number_of_dram_channels(struct radeon_device *rdev)
1915 {
1916         u32 tmp = RREG32(MC_SHARED_CHMAP);
1917
1918         switch ((tmp & NOOFCHAN_MASK) >> NOOFCHAN_SHIFT) {
1919         case 0:
1920         default:
1921                 return 1;
1922         case 1:
1923                 return 2;
1924         case 2:
1925                 return 4;
1926         case 3:
1927                 return 8;
1928         }
1929 }
1930
1931 struct evergreen_wm_params {
1932         u32 dram_channels; /* number of dram channels */
1933         u32 yclk;          /* bandwidth per dram data pin in kHz */
1934         u32 sclk;          /* engine clock in kHz */
1935         u32 disp_clk;      /* display clock in kHz */
1936         u32 src_width;     /* viewport width */
1937         u32 active_time;   /* active display time in ns */
1938         u32 blank_time;    /* blank time in ns */
1939         bool interlaced;    /* mode is interlaced */
1940         fixed20_12 vsc;    /* vertical scale ratio */
1941         u32 num_heads;     /* number of active crtcs */
1942         u32 bytes_per_pixel; /* bytes per pixel display + overlay */
1943         u32 lb_size;       /* line buffer allocated to pipe */
1944         u32 vtaps;         /* vertical scaler taps */
1945 };
1946
1947 static u32 evergreen_dram_bandwidth(struct evergreen_wm_params *wm)
1948 {
1949         /* Calculate DRAM Bandwidth and the part allocated to display. */
1950         fixed20_12 dram_efficiency; /* 0.7 */
1951         fixed20_12 yclk, dram_channels, bandwidth;
1952         fixed20_12 a;
1953
1954         a.full = dfixed_const(1000);
1955         yclk.full = dfixed_const(wm->yclk);
1956         yclk.full = dfixed_div(yclk, a);
1957         dram_channels.full = dfixed_const(wm->dram_channels * 4);
1958         a.full = dfixed_const(10);
1959         dram_efficiency.full = dfixed_const(7);
1960         dram_efficiency.full = dfixed_div(dram_efficiency, a);
1961         bandwidth.full = dfixed_mul(dram_channels, yclk);
1962         bandwidth.full = dfixed_mul(bandwidth, dram_efficiency);
1963
1964         return dfixed_trunc(bandwidth);
1965 }
1966
1967 static u32 evergreen_dram_bandwidth_for_display(struct evergreen_wm_params *wm)
1968 {
1969         /* Calculate DRAM Bandwidth and the part allocated to display. */
1970         fixed20_12 disp_dram_allocation; /* 0.3 to 0.7 */
1971         fixed20_12 yclk, dram_channels, bandwidth;
1972         fixed20_12 a;
1973
1974         a.full = dfixed_const(1000);
1975         yclk.full = dfixed_const(wm->yclk);
1976         yclk.full = dfixed_div(yclk, a);
1977         dram_channels.full = dfixed_const(wm->dram_channels * 4);
1978         a.full = dfixed_const(10);
1979         disp_dram_allocation.full = dfixed_const(3); /* XXX worse case value 0.3 */
1980         disp_dram_allocation.full = dfixed_div(disp_dram_allocation, a);
1981         bandwidth.full = dfixed_mul(dram_channels, yclk);
1982         bandwidth.full = dfixed_mul(bandwidth, disp_dram_allocation);
1983
1984         return dfixed_trunc(bandwidth);
1985 }
1986
1987 static u32 evergreen_data_return_bandwidth(struct evergreen_wm_params *wm)
1988 {
1989         /* Calculate the display Data return Bandwidth */
1990         fixed20_12 return_efficiency; /* 0.8 */
1991         fixed20_12 sclk, bandwidth;
1992         fixed20_12 a;
1993
1994         a.full = dfixed_const(1000);
1995         sclk.full = dfixed_const(wm->sclk);
1996         sclk.full = dfixed_div(sclk, a);
1997         a.full = dfixed_const(10);
1998         return_efficiency.full = dfixed_const(8);
1999         return_efficiency.full = dfixed_div(return_efficiency, a);
2000         a.full = dfixed_const(32);
2001         bandwidth.full = dfixed_mul(a, sclk);
2002         bandwidth.full = dfixed_mul(bandwidth, return_efficiency);
2003
2004         return dfixed_trunc(bandwidth);
2005 }
2006
2007 static u32 evergreen_dmif_request_bandwidth(struct evergreen_wm_params *wm)
2008 {
2009         /* Calculate the DMIF Request Bandwidth */
2010         fixed20_12 disp_clk_request_efficiency; /* 0.8 */
2011         fixed20_12 disp_clk, bandwidth;
2012         fixed20_12 a;
2013
2014         a.full = dfixed_const(1000);
2015         disp_clk.full = dfixed_const(wm->disp_clk);
2016         disp_clk.full = dfixed_div(disp_clk, a);
2017         a.full = dfixed_const(10);
2018         disp_clk_request_efficiency.full = dfixed_const(8);
2019         disp_clk_request_efficiency.full = dfixed_div(disp_clk_request_efficiency, a);
2020         a.full = dfixed_const(32);
2021         bandwidth.full = dfixed_mul(a, disp_clk);
2022         bandwidth.full = dfixed_mul(bandwidth, disp_clk_request_efficiency);
2023
2024         return dfixed_trunc(bandwidth);
2025 }
2026
2027 static u32 evergreen_available_bandwidth(struct evergreen_wm_params *wm)
2028 {
2029         /* Calculate the Available bandwidth. Display can use this temporarily but not in average. */
2030         u32 dram_bandwidth = evergreen_dram_bandwidth(wm);
2031         u32 data_return_bandwidth = evergreen_data_return_bandwidth(wm);
2032         u32 dmif_req_bandwidth = evergreen_dmif_request_bandwidth(wm);
2033
2034         return min(dram_bandwidth, min(data_return_bandwidth, dmif_req_bandwidth));
2035 }
2036
2037 static u32 evergreen_average_bandwidth(struct evergreen_wm_params *wm)
2038 {
2039         /* Calculate the display mode Average Bandwidth
2040          * DisplayMode should contain the source and destination dimensions,
2041          * timing, etc.
2042          */
2043         fixed20_12 bpp;
2044         fixed20_12 line_time;
2045         fixed20_12 src_width;
2046         fixed20_12 bandwidth;
2047         fixed20_12 a;
2048
2049         a.full = dfixed_const(1000);
2050         line_time.full = dfixed_const(wm->active_time + wm->blank_time);
2051         line_time.full = dfixed_div(line_time, a);
2052         bpp.full = dfixed_const(wm->bytes_per_pixel);
2053         src_width.full = dfixed_const(wm->src_width);
2054         bandwidth.full = dfixed_mul(src_width, bpp);
2055         bandwidth.full = dfixed_mul(bandwidth, wm->vsc);
2056         bandwidth.full = dfixed_div(bandwidth, line_time);
2057
2058         return dfixed_trunc(bandwidth);
2059 }
2060
2061 static u32 evergreen_latency_watermark(struct evergreen_wm_params *wm)
2062 {
2063         /* First calcualte the latency in ns */
2064         u32 mc_latency = 2000; /* 2000 ns. */
2065         u32 available_bandwidth = evergreen_available_bandwidth(wm);
2066         u32 worst_chunk_return_time = (512 * 8 * 1000) / available_bandwidth;
2067         u32 cursor_line_pair_return_time = (128 * 4 * 1000) / available_bandwidth;
2068         u32 dc_latency = 40000000 / wm->disp_clk; /* dc pipe latency */
2069         u32 other_heads_data_return_time = ((wm->num_heads + 1) * worst_chunk_return_time) +
2070                 (wm->num_heads * cursor_line_pair_return_time);
2071         u32 latency = mc_latency + other_heads_data_return_time + dc_latency;
2072         u32 max_src_lines_per_dst_line, lb_fill_bw, line_fill_time;
2073         fixed20_12 a, b, c;
2074
2075         if (wm->num_heads == 0)
2076                 return 0;
2077
2078         a.full = dfixed_const(2);
2079         b.full = dfixed_const(1);
2080         if ((wm->vsc.full > a.full) ||
2081             ((wm->vsc.full > b.full) && (wm->vtaps >= 3)) ||
2082             (wm->vtaps >= 5) ||
2083             ((wm->vsc.full >= a.full) && wm->interlaced))
2084                 max_src_lines_per_dst_line = 4;
2085         else
2086                 max_src_lines_per_dst_line = 2;
2087
2088         a.full = dfixed_const(available_bandwidth);
2089         b.full = dfixed_const(wm->num_heads);
2090         a.full = dfixed_div(a, b);
2091
2092         lb_fill_bw = min(dfixed_trunc(a), wm->disp_clk * wm->bytes_per_pixel / 1000);
2093
2094         a.full = dfixed_const(max_src_lines_per_dst_line * wm->src_width * wm->bytes_per_pixel);
2095         b.full = dfixed_const(1000);
2096         c.full = dfixed_const(lb_fill_bw);
2097         b.full = dfixed_div(c, b);
2098         a.full = dfixed_div(a, b);
2099         line_fill_time = dfixed_trunc(a);
2100
2101         if (line_fill_time < wm->active_time)
2102                 return latency;
2103         else
2104                 return latency + (line_fill_time - wm->active_time);
2105
2106 }
2107
2108 static bool evergreen_average_bandwidth_vs_dram_bandwidth_for_display(struct evergreen_wm_params *wm)
2109 {
2110         if (evergreen_average_bandwidth(wm) <=
2111             (evergreen_dram_bandwidth_for_display(wm) / wm->num_heads))
2112                 return true;
2113         else
2114                 return false;
2115 };
2116
2117 static bool evergreen_average_bandwidth_vs_available_bandwidth(struct evergreen_wm_params *wm)
2118 {
2119         if (evergreen_average_bandwidth(wm) <=
2120             (evergreen_available_bandwidth(wm) / wm->num_heads))
2121                 return true;
2122         else
2123                 return false;
2124 };
2125
2126 static bool evergreen_check_latency_hiding(struct evergreen_wm_params *wm)
2127 {
2128         u32 lb_partitions = wm->lb_size / wm->src_width;
2129         u32 line_time = wm->active_time + wm->blank_time;
2130         u32 latency_tolerant_lines;
2131         u32 latency_hiding;
2132         fixed20_12 a;
2133
2134         a.full = dfixed_const(1);
2135         if (wm->vsc.full > a.full)
2136                 latency_tolerant_lines = 1;
2137         else {
2138                 if (lb_partitions <= (wm->vtaps + 1))
2139                         latency_tolerant_lines = 1;
2140                 else
2141                         latency_tolerant_lines = 2;
2142         }
2143
2144         latency_hiding = (latency_tolerant_lines * line_time + wm->blank_time);
2145
2146         if (evergreen_latency_watermark(wm) <= latency_hiding)
2147                 return true;
2148         else
2149                 return false;
2150 }
2151
2152 static void evergreen_program_watermarks(struct radeon_device *rdev,
2153                                          struct radeon_crtc *radeon_crtc,
2154                                          u32 lb_size, u32 num_heads)
2155 {
2156         struct drm_display_mode *mode = &radeon_crtc->base.mode;
2157         struct evergreen_wm_params wm_low, wm_high;
2158         u32 dram_channels;
2159         u32 active_time;
2160         u32 line_time = 0;
2161         u32 latency_watermark_a = 0, latency_watermark_b = 0;
2162         u32 priority_a_mark = 0, priority_b_mark = 0;
2163         u32 priority_a_cnt = PRIORITY_OFF;
2164         u32 priority_b_cnt = PRIORITY_OFF;
2165         u32 pipe_offset = radeon_crtc->crtc_id * 16;
2166         u32 tmp, arb_control3;
2167         fixed20_12 a, b, c;
2168
2169         if (radeon_crtc->base.enabled && num_heads && mode) {
2170                 active_time = (u32) div_u64((u64)mode->crtc_hdisplay * 1000000,
2171                                             (u32)mode->clock);
2172                 line_time = (u32) div_u64((u64)mode->crtc_htotal * 1000000,
2173                                           (u32)mode->clock);
2174                 line_time = min(line_time, (u32)65535);
2175                 priority_a_cnt = 0;
2176                 priority_b_cnt = 0;
2177                 dram_channels = evergreen_get_number_of_dram_channels(rdev);
2178
2179                 /* watermark for high clocks */
2180                 if ((rdev->pm.pm_method == PM_METHOD_DPM) && rdev->pm.dpm_enabled) {
2181                         wm_high.yclk =
2182                                 radeon_dpm_get_mclk(rdev, false) * 10;
2183                         wm_high.sclk =
2184                                 radeon_dpm_get_sclk(rdev, false) * 10;
2185                 } else {
2186                         wm_high.yclk = rdev->pm.current_mclk * 10;
2187                         wm_high.sclk = rdev->pm.current_sclk * 10;
2188                 }
2189
2190                 wm_high.disp_clk = mode->clock;
2191                 wm_high.src_width = mode->crtc_hdisplay;
2192                 wm_high.active_time = active_time;
2193                 wm_high.blank_time = line_time - wm_high.active_time;
2194                 wm_high.interlaced = false;
2195                 if (mode->flags & DRM_MODE_FLAG_INTERLACE)
2196                         wm_high.interlaced = true;
2197                 wm_high.vsc = radeon_crtc->vsc;
2198                 wm_high.vtaps = 1;
2199                 if (radeon_crtc->rmx_type != RMX_OFF)
2200                         wm_high.vtaps = 2;
2201                 wm_high.bytes_per_pixel = 4; /* XXX: get this from fb config */
2202                 wm_high.lb_size = lb_size;
2203                 wm_high.dram_channels = dram_channels;
2204                 wm_high.num_heads = num_heads;
2205
2206                 /* watermark for low clocks */
2207                 if ((rdev->pm.pm_method == PM_METHOD_DPM) && rdev->pm.dpm_enabled) {
2208                         wm_low.yclk =
2209                                 radeon_dpm_get_mclk(rdev, true) * 10;
2210                         wm_low.sclk =
2211                                 radeon_dpm_get_sclk(rdev, true) * 10;
2212                 } else {
2213                         wm_low.yclk = rdev->pm.current_mclk * 10;
2214                         wm_low.sclk = rdev->pm.current_sclk * 10;
2215                 }
2216
2217                 wm_low.disp_clk = mode->clock;
2218                 wm_low.src_width = mode->crtc_hdisplay;
2219                 wm_low.active_time = active_time;
2220                 wm_low.blank_time = line_time - wm_low.active_time;
2221                 wm_low.interlaced = false;
2222                 if (mode->flags & DRM_MODE_FLAG_INTERLACE)
2223                         wm_low.interlaced = true;
2224                 wm_low.vsc = radeon_crtc->vsc;
2225                 wm_low.vtaps = 1;
2226                 if (radeon_crtc->rmx_type != RMX_OFF)
2227                         wm_low.vtaps = 2;
2228                 wm_low.bytes_per_pixel = 4; /* XXX: get this from fb config */
2229                 wm_low.lb_size = lb_size;
2230                 wm_low.dram_channels = dram_channels;
2231                 wm_low.num_heads = num_heads;
2232
2233                 /* set for high clocks */
2234                 latency_watermark_a = min(evergreen_latency_watermark(&wm_high), (u32)65535);
2235                 /* set for low clocks */
2236                 latency_watermark_b = min(evergreen_latency_watermark(&wm_low), (u32)65535);
2237
2238                 /* possibly force display priority to high */
2239                 /* should really do this at mode validation time... */
2240                 if (!evergreen_average_bandwidth_vs_dram_bandwidth_for_display(&wm_high) ||
2241                     !evergreen_average_bandwidth_vs_available_bandwidth(&wm_high) ||
2242                     !evergreen_check_latency_hiding(&wm_high) ||
2243                     (rdev->disp_priority == 2)) {
2244                         DRM_DEBUG_KMS("force priority a to high\n");
2245                         priority_a_cnt |= PRIORITY_ALWAYS_ON;
2246                 }
2247                 if (!evergreen_average_bandwidth_vs_dram_bandwidth_for_display(&wm_low) ||
2248                     !evergreen_average_bandwidth_vs_available_bandwidth(&wm_low) ||
2249                     !evergreen_check_latency_hiding(&wm_low) ||
2250                     (rdev->disp_priority == 2)) {
2251                         DRM_DEBUG_KMS("force priority b to high\n");
2252                         priority_b_cnt |= PRIORITY_ALWAYS_ON;
2253                 }
2254
2255                 a.full = dfixed_const(1000);
2256                 b.full = dfixed_const(mode->clock);
2257                 b.full = dfixed_div(b, a);
2258                 c.full = dfixed_const(latency_watermark_a);
2259                 c.full = dfixed_mul(c, b);
2260                 c.full = dfixed_mul(c, radeon_crtc->hsc);
2261                 c.full = dfixed_div(c, a);
2262                 a.full = dfixed_const(16);
2263                 c.full = dfixed_div(c, a);
2264                 priority_a_mark = dfixed_trunc(c);
2265                 priority_a_cnt |= priority_a_mark & PRIORITY_MARK_MASK;
2266
2267                 a.full = dfixed_const(1000);
2268                 b.full = dfixed_const(mode->clock);
2269                 b.full = dfixed_div(b, a);
2270                 c.full = dfixed_const(latency_watermark_b);
2271                 c.full = dfixed_mul(c, b);
2272                 c.full = dfixed_mul(c, radeon_crtc->hsc);
2273                 c.full = dfixed_div(c, a);
2274                 a.full = dfixed_const(16);
2275                 c.full = dfixed_div(c, a);
2276                 priority_b_mark = dfixed_trunc(c);
2277                 priority_b_cnt |= priority_b_mark & PRIORITY_MARK_MASK;
2278
2279                 /* Save number of lines the linebuffer leads before the scanout */
2280                 radeon_crtc->lb_vblank_lead_lines = DIV_ROUND_UP(lb_size, mode->crtc_hdisplay);
2281         }
2282
2283         /* select wm A */
2284         arb_control3 = RREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset);
2285         tmp = arb_control3;
2286         tmp &= ~LATENCY_WATERMARK_MASK(3);
2287         tmp |= LATENCY_WATERMARK_MASK(1);
2288         WREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset, tmp);
2289         WREG32(PIPE0_LATENCY_CONTROL + pipe_offset,
2290                (LATENCY_LOW_WATERMARK(latency_watermark_a) |
2291                 LATENCY_HIGH_WATERMARK(line_time)));
2292         /* select wm B */
2293         tmp = RREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset);
2294         tmp &= ~LATENCY_WATERMARK_MASK(3);
2295         tmp |= LATENCY_WATERMARK_MASK(2);
2296         WREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset, tmp);
2297         WREG32(PIPE0_LATENCY_CONTROL + pipe_offset,
2298                (LATENCY_LOW_WATERMARK(latency_watermark_b) |
2299                 LATENCY_HIGH_WATERMARK(line_time)));
2300         /* restore original selection */
2301         WREG32(PIPE0_ARBITRATION_CONTROL3 + pipe_offset, arb_control3);
2302
2303         /* write the priority marks */
2304         WREG32(PRIORITY_A_CNT + radeon_crtc->crtc_offset, priority_a_cnt);
2305         WREG32(PRIORITY_B_CNT + radeon_crtc->crtc_offset, priority_b_cnt);
2306
2307         /* save values for DPM */
2308         radeon_crtc->line_time = line_time;
2309         radeon_crtc->wm_high = latency_watermark_a;
2310         radeon_crtc->wm_low = latency_watermark_b;
2311 }
2312
2313 /**
2314  * evergreen_bandwidth_update - update display watermarks callback.
2315  *
2316  * @rdev: radeon_device pointer
2317  *
2318  * Update the display watermarks based on the requested mode(s)
2319  * (evergreen+).
2320  */
2321 void evergreen_bandwidth_update(struct radeon_device *rdev)
2322 {
2323         struct drm_display_mode *mode0 = NULL;
2324         struct drm_display_mode *mode1 = NULL;
2325         u32 num_heads = 0, lb_size;
2326         int i;
2327
2328         if (!rdev->mode_info.mode_config_initialized)
2329                 return;
2330
2331         radeon_update_display_priority(rdev);
2332
2333         for (i = 0; i < rdev->num_crtc; i++) {
2334                 if (rdev->mode_info.crtcs[i]->base.enabled)
2335                         num_heads++;
2336         }
2337         for (i = 0; i < rdev->num_crtc; i += 2) {
2338                 mode0 = &rdev->mode_info.crtcs[i]->base.mode;
2339                 mode1 = &rdev->mode_info.crtcs[i+1]->base.mode;
2340                 lb_size = evergreen_line_buffer_adjust(rdev, rdev->mode_info.crtcs[i], mode0, mode1);
2341                 evergreen_program_watermarks(rdev, rdev->mode_info.crtcs[i], lb_size, num_heads);
2342                 lb_size = evergreen_line_buffer_adjust(rdev, rdev->mode_info.crtcs[i+1], mode1, mode0);
2343                 evergreen_program_watermarks(rdev, rdev->mode_info.crtcs[i+1], lb_size, num_heads);
2344         }
2345 }
2346
2347 /**
2348  * evergreen_mc_wait_for_idle - wait for MC idle callback.
2349  *
2350  * @rdev: radeon_device pointer
2351  *
2352  * Wait for the MC (memory controller) to be idle.
2353  * (evergreen+).
2354  * Returns 0 if the MC is idle, -1 if not.
2355  */
2356 int evergreen_mc_wait_for_idle(struct radeon_device *rdev)
2357 {
2358         unsigned i;
2359         u32 tmp;
2360
2361         for (i = 0; i < rdev->usec_timeout; i++) {
2362                 /* read MC_STATUS */
2363                 tmp = RREG32(SRBM_STATUS) & 0x1F00;
2364                 if (!tmp)
2365                         return 0;
2366                 udelay(1);
2367         }
2368         return -1;
2369 }
2370
2371 /*
2372  * GART
2373  */
2374 void evergreen_pcie_gart_tlb_flush(struct radeon_device *rdev)
2375 {
2376         unsigned i;
2377         u32 tmp;
2378
2379         WREG32(HDP_MEM_COHERENCY_FLUSH_CNTL, 0x1);
2380
2381         WREG32(VM_CONTEXT0_REQUEST_RESPONSE, REQUEST_TYPE(1));
2382         for (i = 0; i < rdev->usec_timeout; i++) {
2383                 /* read MC_STATUS */
2384                 tmp = RREG32(VM_CONTEXT0_REQUEST_RESPONSE);
2385                 tmp = (tmp & RESPONSE_TYPE_MASK) >> RESPONSE_TYPE_SHIFT;
2386                 if (tmp == 2) {
2387                         pr_warn("[drm] r600 flush TLB failed\n");
2388                         return;
2389                 }
2390                 if (tmp) {
2391                         return;
2392                 }
2393                 udelay(1);
2394         }
2395 }
2396
2397 static int evergreen_pcie_gart_enable(struct radeon_device *rdev)
2398 {
2399         u32 tmp;
2400         int r;
2401
2402         if (rdev->gart.robj == NULL) {
2403                 dev_err(rdev->dev, "No VRAM object for PCIE GART.\n");
2404                 return -EINVAL;
2405         }
2406         r = radeon_gart_table_vram_pin(rdev);
2407         if (r)
2408                 return r;
2409         /* Setup L2 cache */
2410         WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
2411                                 ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
2412                                 EFFECTIVE_L2_QUEUE_SIZE(7));
2413         WREG32(VM_L2_CNTL2, 0);
2414         WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
2415         /* Setup TLB control */
2416         tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
2417                 SYSTEM_ACCESS_MODE_NOT_IN_SYS |
2418                 SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
2419                 EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
2420         if (rdev->flags & RADEON_IS_IGP) {
2421                 WREG32(FUS_MC_VM_MD_L1_TLB0_CNTL, tmp);
2422                 WREG32(FUS_MC_VM_MD_L1_TLB1_CNTL, tmp);
2423                 WREG32(FUS_MC_VM_MD_L1_TLB2_CNTL, tmp);
2424         } else {
2425                 WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
2426                 WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
2427                 WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
2428                 if ((rdev->family == CHIP_JUNIPER) ||
2429                     (rdev->family == CHIP_CYPRESS) ||
2430                     (rdev->family == CHIP_HEMLOCK) ||
2431                     (rdev->family == CHIP_BARTS))
2432                         WREG32(MC_VM_MD_L1_TLB3_CNTL, tmp);
2433         }
2434         WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
2435         WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
2436         WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
2437         WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
2438         WREG32(VM_CONTEXT0_PAGE_TABLE_START_ADDR, rdev->mc.gtt_start >> 12);
2439         WREG32(VM_CONTEXT0_PAGE_TABLE_END_ADDR, rdev->mc.gtt_end >> 12);
2440         WREG32(VM_CONTEXT0_PAGE_TABLE_BASE_ADDR, rdev->gart.table_addr >> 12);
2441         WREG32(VM_CONTEXT0_CNTL, ENABLE_CONTEXT | PAGE_TABLE_DEPTH(0) |
2442                                 RANGE_PROTECTION_FAULT_ENABLE_DEFAULT);
2443         WREG32(VM_CONTEXT0_PROTECTION_FAULT_DEFAULT_ADDR,
2444                         (u32)(rdev->dummy_page.addr >> 12));
2445         WREG32(VM_CONTEXT1_CNTL, 0);
2446
2447         evergreen_pcie_gart_tlb_flush(rdev);
2448         DRM_INFO("PCIE GART of %uM enabled (table at 0x%016llX).\n",
2449                  (unsigned)(rdev->mc.gtt_size >> 20),
2450                  (unsigned long long)rdev->gart.table_addr);
2451         rdev->gart.ready = true;
2452         return 0;
2453 }
2454
2455 static void evergreen_pcie_gart_disable(struct radeon_device *rdev)
2456 {
2457         u32 tmp;
2458
2459         /* Disable all tables */
2460         WREG32(VM_CONTEXT0_CNTL, 0);
2461         WREG32(VM_CONTEXT1_CNTL, 0);
2462
2463         /* Setup L2 cache */
2464         WREG32(VM_L2_CNTL, ENABLE_L2_FRAGMENT_PROCESSING |
2465                                 EFFECTIVE_L2_QUEUE_SIZE(7));
2466         WREG32(VM_L2_CNTL2, 0);
2467         WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
2468         /* Setup TLB control */
2469         tmp = EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
2470         WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
2471         WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
2472         WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
2473         WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
2474         WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
2475         WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
2476         WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
2477         radeon_gart_table_vram_unpin(rdev);
2478 }
2479
2480 static void evergreen_pcie_gart_fini(struct radeon_device *rdev)
2481 {
2482         evergreen_pcie_gart_disable(rdev);
2483         radeon_gart_table_vram_free(rdev);
2484         radeon_gart_fini(rdev);
2485 }
2486
2487
2488 static void evergreen_agp_enable(struct radeon_device *rdev)
2489 {
2490         u32 tmp;
2491
2492         /* Setup L2 cache */
2493         WREG32(VM_L2_CNTL, ENABLE_L2_CACHE | ENABLE_L2_FRAGMENT_PROCESSING |
2494                                 ENABLE_L2_PTE_CACHE_LRU_UPDATE_BY_WRITE |
2495                                 EFFECTIVE_L2_QUEUE_SIZE(7));
2496         WREG32(VM_L2_CNTL2, 0);
2497         WREG32(VM_L2_CNTL3, BANK_SELECT(0) | CACHE_UPDATE_MODE(2));
2498         /* Setup TLB control */
2499         tmp = ENABLE_L1_TLB | ENABLE_L1_FRAGMENT_PROCESSING |
2500                 SYSTEM_ACCESS_MODE_NOT_IN_SYS |
2501                 SYSTEM_APERTURE_UNMAPPED_ACCESS_PASS_THRU |
2502                 EFFECTIVE_L1_TLB_SIZE(5) | EFFECTIVE_L1_QUEUE_SIZE(5);
2503         WREG32(MC_VM_MD_L1_TLB0_CNTL, tmp);
2504         WREG32(MC_VM_MD_L1_TLB1_CNTL, tmp);
2505         WREG32(MC_VM_MD_L1_TLB2_CNTL, tmp);
2506         WREG32(MC_VM_MB_L1_TLB0_CNTL, tmp);
2507         WREG32(MC_VM_MB_L1_TLB1_CNTL, tmp);
2508         WREG32(MC_VM_MB_L1_TLB2_CNTL, tmp);
2509         WREG32(MC_VM_MB_L1_TLB3_CNTL, tmp);
2510         WREG32(VM_CONTEXT0_CNTL, 0);
2511         WREG32(VM_CONTEXT1_CNTL, 0);
2512 }
2513
2514 static const unsigned ni_dig_offsets[] =
2515 {
2516         NI_DIG0_REGISTER_OFFSET,
2517         NI_DIG1_REGISTER_OFFSET,
2518         NI_DIG2_REGISTER_OFFSET,
2519         NI_DIG3_REGISTER_OFFSET,
2520         NI_DIG4_REGISTER_OFFSET,
2521         NI_DIG5_REGISTER_OFFSET
2522 };
2523
2524 static const unsigned ni_tx_offsets[] =
2525 {
2526         NI_DCIO_UNIPHY0_UNIPHY_TX_CONTROL1,
2527         NI_DCIO_UNIPHY1_UNIPHY_TX_CONTROL1,
2528         NI_DCIO_UNIPHY2_UNIPHY_TX_CONTROL1,
2529         NI_DCIO_UNIPHY3_UNIPHY_TX_CONTROL1,
2530         NI_DCIO_UNIPHY4_UNIPHY_TX_CONTROL1,
2531         NI_DCIO_UNIPHY5_UNIPHY_TX_CONTROL1
2532 };
2533
2534 static const unsigned evergreen_dp_offsets[] =
2535 {
2536         EVERGREEN_DP0_REGISTER_OFFSET,
2537         EVERGREEN_DP1_REGISTER_OFFSET,
2538         EVERGREEN_DP2_REGISTER_OFFSET,
2539         EVERGREEN_DP3_REGISTER_OFFSET,
2540         EVERGREEN_DP4_REGISTER_OFFSET,
2541         EVERGREEN_DP5_REGISTER_OFFSET
2542 };
2543
2544 static const unsigned evergreen_disp_int_status[] =
2545 {
2546         DISP_INTERRUPT_STATUS,
2547         DISP_INTERRUPT_STATUS_CONTINUE,
2548         DISP_INTERRUPT_STATUS_CONTINUE2,
2549         DISP_INTERRUPT_STATUS_CONTINUE3,
2550         DISP_INTERRUPT_STATUS_CONTINUE4,
2551         DISP_INTERRUPT_STATUS_CONTINUE5
2552 };
2553
2554 /*
2555  * Assumption is that EVERGREEN_CRTC_MASTER_EN enable for requested crtc
2556  * We go from crtc to connector and it is not relible  since it
2557  * should be an opposite direction .If crtc is enable then
2558  * find the dig_fe which selects this crtc and insure that it enable.
2559  * if such dig_fe is found then find dig_be which selects found dig_be and
2560  * insure that it enable and in DP_SST mode.
2561  * if UNIPHY_PLL_CONTROL1.enable then we should disconnect timing
2562  * from dp symbols clocks .
2563  */
2564 static bool evergreen_is_dp_sst_stream_enabled(struct radeon_device *rdev,
2565                                                unsigned crtc_id, unsigned *ret_dig_fe)
2566 {
2567         unsigned i;
2568         unsigned dig_fe;
2569         unsigned dig_be;
2570         unsigned dig_en_be;
2571         unsigned uniphy_pll;
2572         unsigned digs_fe_selected;
2573         unsigned dig_be_mode;
2574         unsigned dig_fe_mask;
2575         bool is_enabled = false;
2576         bool found_crtc = false;
2577
2578         /* loop through all running dig_fe to find selected crtc */
2579         for (i = 0; i < ARRAY_SIZE(ni_dig_offsets); i++) {
2580                 dig_fe = RREG32(NI_DIG_FE_CNTL + ni_dig_offsets[i]);
2581                 if (dig_fe & NI_DIG_FE_CNTL_SYMCLK_FE_ON &&
2582                     crtc_id == NI_DIG_FE_CNTL_SOURCE_SELECT(dig_fe)) {
2583                         /* found running pipe */
2584                         found_crtc = true;
2585                         dig_fe_mask = 1 << i;
2586                         dig_fe = i;
2587                         break;
2588                 }
2589         }
2590
2591         if (found_crtc) {
2592                 /* loop through all running dig_be to find selected dig_fe */
2593                 for (i = 0; i < ARRAY_SIZE(ni_dig_offsets); i++) {
2594                         dig_be = RREG32(NI_DIG_BE_CNTL + ni_dig_offsets[i]);
2595                         /* if dig_fe_selected by dig_be? */
2596                         digs_fe_selected = NI_DIG_BE_CNTL_FE_SOURCE_SELECT(dig_be);
2597                         dig_be_mode = NI_DIG_FE_CNTL_MODE(dig_be);
2598                         if (dig_fe_mask &  digs_fe_selected &&
2599                             /* if dig_be in sst mode? */
2600                             dig_be_mode == NI_DIG_BE_DPSST) {
2601                                 dig_en_be = RREG32(NI_DIG_BE_EN_CNTL +
2602                                                    ni_dig_offsets[i]);
2603                                 uniphy_pll = RREG32(NI_DCIO_UNIPHY0_PLL_CONTROL1 +
2604                                                     ni_tx_offsets[i]);
2605                                 /* dig_be enable and tx is running */
2606                                 if (dig_en_be & NI_DIG_BE_EN_CNTL_ENABLE &&
2607                                     dig_en_be & NI_DIG_BE_EN_CNTL_SYMBCLK_ON &&
2608                                     uniphy_pll & NI_DCIO_UNIPHY0_PLL_CONTROL1_ENABLE) {
2609                                         is_enabled = true;
2610                                         *ret_dig_fe = dig_fe;
2611                                         break;
2612                                 }
2613                         }
2614                 }
2615         }
2616
2617         return is_enabled;
2618 }
2619
2620 /*
2621  * Blank dig when in dp sst mode
2622  * Dig ignores crtc timing
2623  */
2624 static void evergreen_blank_dp_output(struct radeon_device *rdev,
2625                                       unsigned dig_fe)
2626 {
2627         unsigned stream_ctrl;
2628         unsigned fifo_ctrl;
2629         unsigned counter = 0;
2630
2631         if (dig_fe >= ARRAY_SIZE(evergreen_dp_offsets)) {
2632                 DRM_ERROR("invalid dig_fe %d\n", dig_fe);
2633                 return;
2634         }
2635
2636         stream_ctrl = RREG32(EVERGREEN_DP_VID_STREAM_CNTL +
2637                              evergreen_dp_offsets[dig_fe]);
2638         if (!(stream_ctrl & EVERGREEN_DP_VID_STREAM_CNTL_ENABLE)) {
2639                 DRM_ERROR("dig %d , should be enable\n", dig_fe);
2640                 return;
2641         }
2642
2643         stream_ctrl &=~EVERGREEN_DP_VID_STREAM_CNTL_ENABLE;
2644         WREG32(EVERGREEN_DP_VID_STREAM_CNTL +
2645                evergreen_dp_offsets[dig_fe], stream_ctrl);
2646
2647         stream_ctrl = RREG32(EVERGREEN_DP_VID_STREAM_CNTL +
2648                              evergreen_dp_offsets[dig_fe]);
2649         while (counter < 32 && stream_ctrl & EVERGREEN_DP_VID_STREAM_STATUS) {
2650                 msleep(1);
2651                 counter++;
2652                 stream_ctrl = RREG32(EVERGREEN_DP_VID_STREAM_CNTL +
2653                                      evergreen_dp_offsets[dig_fe]);
2654         }
2655         if (counter >= 32 )
2656                 DRM_ERROR("counter exceeds %d\n", counter);
2657
2658         fifo_ctrl = RREG32(EVERGREEN_DP_STEER_FIFO + evergreen_dp_offsets[dig_fe]);
2659         fifo_ctrl |= EVERGREEN_DP_STEER_FIFO_RESET;
2660         WREG32(EVERGREEN_DP_STEER_FIFO + evergreen_dp_offsets[dig_fe], fifo_ctrl);
2661
2662 }
2663
2664 void evergreen_mc_stop(struct radeon_device *rdev, struct evergreen_mc_save *save)
2665 {
2666         u32 crtc_enabled, tmp, frame_count, blackout;
2667         int i, j;
2668         unsigned dig_fe;
2669
2670         if (!ASIC_IS_NODCE(rdev)) {
2671                 save->vga_render_control = RREG32(VGA_RENDER_CONTROL);
2672                 save->vga_hdp_control = RREG32(VGA_HDP_CONTROL);
2673
2674                 /* disable VGA render */
2675                 WREG32(VGA_RENDER_CONTROL, 0);
2676         }
2677         /* blank the display controllers */
2678         for (i = 0; i < rdev->num_crtc; i++) {
2679                 crtc_enabled = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]) & EVERGREEN_CRTC_MASTER_EN;
2680                 if (crtc_enabled) {
2681                         save->crtc_enabled[i] = true;
2682                         if (ASIC_IS_DCE6(rdev)) {
2683                                 tmp = RREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i]);
2684                                 if (!(tmp & EVERGREEN_CRTC_BLANK_DATA_EN)) {
2685                                         radeon_wait_for_vblank(rdev, i);
2686                                         WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
2687                                         tmp |= EVERGREEN_CRTC_BLANK_DATA_EN;
2688                                         WREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i], tmp);
2689                                         WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
2690                                 }
2691                         } else {
2692                                 tmp = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]);
2693                                 if (!(tmp & EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE)) {
2694                                         radeon_wait_for_vblank(rdev, i);
2695                                         WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
2696                                         tmp |= EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
2697                                         WREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i], tmp);
2698                                         WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
2699                                 }
2700                         }
2701                         /* wait for the next frame */
2702                         frame_count = radeon_get_vblank_counter(rdev, i);
2703                         for (j = 0; j < rdev->usec_timeout; j++) {
2704                                 if (radeon_get_vblank_counter(rdev, i) != frame_count)
2705                                         break;
2706                                 udelay(1);
2707                         }
2708                         /*we should disable dig if it drives dp sst*/
2709                         /*but we are in radeon_device_init and the topology is unknown*/
2710                         /*and it is available after radeon_modeset_init*/
2711                         /*the following method radeon_atom_encoder_dpms_dig*/
2712                         /*does the job if we initialize it properly*/
2713                         /*for now we do it this manually*/
2714                         /**/
2715                         if (ASIC_IS_DCE5(rdev) &&
2716                             evergreen_is_dp_sst_stream_enabled(rdev, i ,&dig_fe))
2717                                 evergreen_blank_dp_output(rdev, dig_fe);
2718                         /*we could remove 6 lines below*/
2719                         /* XXX this is a hack to avoid strange behavior with EFI on certain systems */
2720                         WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
2721                         tmp = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]);
2722                         tmp &= ~EVERGREEN_CRTC_MASTER_EN;
2723                         WREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i], tmp);
2724                         WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
2725                         save->crtc_enabled[i] = false;
2726                         /* ***** */
2727                 } else {
2728                         save->crtc_enabled[i] = false;
2729                 }
2730         }
2731
2732         radeon_mc_wait_for_idle(rdev);
2733
2734         blackout = RREG32(MC_SHARED_BLACKOUT_CNTL);
2735         if ((blackout & BLACKOUT_MODE_MASK) != 1) {
2736                 /* Block CPU access */
2737                 WREG32(BIF_FB_EN, 0);
2738                 /* blackout the MC */
2739                 blackout &= ~BLACKOUT_MODE_MASK;
2740                 WREG32(MC_SHARED_BLACKOUT_CNTL, blackout | 1);
2741         }
2742         /* wait for the MC to settle */
2743         udelay(100);
2744
2745         /* lock double buffered regs */
2746         for (i = 0; i < rdev->num_crtc; i++) {
2747                 if (save->crtc_enabled[i]) {
2748                         tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
2749                         if (!(tmp & EVERGREEN_GRPH_UPDATE_LOCK)) {
2750                                 tmp |= EVERGREEN_GRPH_UPDATE_LOCK;
2751                                 WREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i], tmp);
2752                         }
2753                         tmp = RREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i]);
2754                         if (!(tmp & 1)) {
2755                                 tmp |= 1;
2756                                 WREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i], tmp);
2757                         }
2758                 }
2759         }
2760 }
2761
2762 void evergreen_mc_resume(struct radeon_device *rdev, struct evergreen_mc_save *save)
2763 {
2764         u32 tmp, frame_count;
2765         int i, j;
2766
2767         /* update crtc base addresses */
2768         for (i = 0; i < rdev->num_crtc; i++) {
2769                 WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS_HIGH + crtc_offsets[i],
2770                        upper_32_bits(rdev->mc.vram_start));
2771                 WREG32(EVERGREEN_GRPH_SECONDARY_SURFACE_ADDRESS_HIGH + crtc_offsets[i],
2772                        upper_32_bits(rdev->mc.vram_start));
2773                 WREG32(EVERGREEN_GRPH_PRIMARY_SURFACE_ADDRESS + crtc_offsets[i],
2774                        (u32)rdev->mc.vram_start);
2775                 WREG32(EVERGREEN_GRPH_SECONDARY_SURFACE_ADDRESS + crtc_offsets[i],
2776                        (u32)rdev->mc.vram_start);
2777         }
2778
2779         if (!ASIC_IS_NODCE(rdev)) {
2780                 WREG32(EVERGREEN_VGA_MEMORY_BASE_ADDRESS_HIGH, upper_32_bits(rdev->mc.vram_start));
2781                 WREG32(EVERGREEN_VGA_MEMORY_BASE_ADDRESS, (u32)rdev->mc.vram_start);
2782         }
2783
2784         /* unlock regs and wait for update */
2785         for (i = 0; i < rdev->num_crtc; i++) {
2786                 if (save->crtc_enabled[i]) {
2787                         tmp = RREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i]);
2788                         if ((tmp & 0x7) != 0) {
2789                                 tmp &= ~0x7;
2790                                 WREG32(EVERGREEN_MASTER_UPDATE_MODE + crtc_offsets[i], tmp);
2791                         }
2792                         tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
2793                         if (tmp & EVERGREEN_GRPH_UPDATE_LOCK) {
2794                                 tmp &= ~EVERGREEN_GRPH_UPDATE_LOCK;
2795                                 WREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i], tmp);
2796                         }
2797                         tmp = RREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i]);
2798                         if (tmp & 1) {
2799                                 tmp &= ~1;
2800                                 WREG32(EVERGREEN_MASTER_UPDATE_LOCK + crtc_offsets[i], tmp);
2801                         }
2802                         for (j = 0; j < rdev->usec_timeout; j++) {
2803                                 tmp = RREG32(EVERGREEN_GRPH_UPDATE + crtc_offsets[i]);
2804                                 if ((tmp & EVERGREEN_GRPH_SURFACE_UPDATE_PENDING) == 0)
2805                                         break;
2806                                 udelay(1);
2807                         }
2808                 }
2809         }
2810
2811         /* unblackout the MC */
2812         tmp = RREG32(MC_SHARED_BLACKOUT_CNTL);
2813         tmp &= ~BLACKOUT_MODE_MASK;
2814         WREG32(MC_SHARED_BLACKOUT_CNTL, tmp);
2815         /* allow CPU access */
2816         WREG32(BIF_FB_EN, FB_READ_EN | FB_WRITE_EN);
2817
2818         for (i = 0; i < rdev->num_crtc; i++) {
2819                 if (save->crtc_enabled[i]) {
2820                         if (ASIC_IS_DCE6(rdev)) {
2821                                 tmp = RREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i]);
2822                                 tmp &= ~EVERGREEN_CRTC_BLANK_DATA_EN;
2823                                 WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
2824                                 WREG32(EVERGREEN_CRTC_BLANK_CONTROL + crtc_offsets[i], tmp);
2825                                 WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
2826                         } else {
2827                                 tmp = RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]);
2828                                 tmp &= ~EVERGREEN_CRTC_DISP_READ_REQUEST_DISABLE;
2829                                 WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 1);
2830                                 WREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i], tmp);
2831                                 WREG32(EVERGREEN_CRTC_UPDATE_LOCK + crtc_offsets[i], 0);
2832                         }
2833                         /* wait for the next frame */
2834                         frame_count = radeon_get_vblank_counter(rdev, i);
2835                         for (j = 0; j < rdev->usec_timeout; j++) {
2836                                 if (radeon_get_vblank_counter(rdev, i) != frame_count)
2837                                         break;
2838                                 udelay(1);
2839                         }
2840                 }
2841         }
2842         if (!ASIC_IS_NODCE(rdev)) {
2843                 /* Unlock vga access */
2844                 WREG32(VGA_HDP_CONTROL, save->vga_hdp_control);
2845                 mdelay(1);
2846                 WREG32(VGA_RENDER_CONTROL, save->vga_render_control);
2847         }
2848 }
2849
2850 void evergreen_mc_program(struct radeon_device *rdev)
2851 {
2852         struct evergreen_mc_save save;
2853         u32 tmp;
2854         int i, j;
2855
2856         /* Initialize HDP */
2857         for (i = 0, j = 0; i < 32; i++, j += 0x18) {
2858                 WREG32((0x2c14 + j), 0x00000000);
2859                 WREG32((0x2c18 + j), 0x00000000);
2860                 WREG32((0x2c1c + j), 0x00000000);
2861                 WREG32((0x2c20 + j), 0x00000000);
2862                 WREG32((0x2c24 + j), 0x00000000);
2863         }
2864         WREG32(HDP_REG_COHERENCY_FLUSH_CNTL, 0);
2865
2866         evergreen_mc_stop(rdev, &save);
2867         if (evergreen_mc_wait_for_idle(rdev)) {
2868                 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
2869         }
2870         /* Lockout access through VGA aperture*/
2871         WREG32(VGA_HDP_CONTROL, VGA_MEMORY_DISABLE);
2872         /* Update configuration */
2873         if (rdev->flags & RADEON_IS_AGP) {
2874                 if (rdev->mc.vram_start < rdev->mc.gtt_start) {
2875                         /* VRAM before AGP */
2876                         WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
2877                                 rdev->mc.vram_start >> 12);
2878                         WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
2879                                 rdev->mc.gtt_end >> 12);
2880                 } else {
2881                         /* VRAM after AGP */
2882                         WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
2883                                 rdev->mc.gtt_start >> 12);
2884                         WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
2885                                 rdev->mc.vram_end >> 12);
2886                 }
2887         } else {
2888                 WREG32(MC_VM_SYSTEM_APERTURE_LOW_ADDR,
2889                         rdev->mc.vram_start >> 12);
2890                 WREG32(MC_VM_SYSTEM_APERTURE_HIGH_ADDR,
2891                         rdev->mc.vram_end >> 12);
2892         }
2893         WREG32(MC_VM_SYSTEM_APERTURE_DEFAULT_ADDR, rdev->vram_scratch.gpu_addr >> 12);
2894         /* llano/ontario only */
2895         if ((rdev->family == CHIP_PALM) ||
2896             (rdev->family == CHIP_SUMO) ||
2897             (rdev->family == CHIP_SUMO2)) {
2898                 tmp = RREG32(MC_FUS_VM_FB_OFFSET) & 0x000FFFFF;
2899                 tmp |= ((rdev->mc.vram_end >> 20) & 0xF) << 24;
2900                 tmp |= ((rdev->mc.vram_start >> 20) & 0xF) << 20;
2901                 WREG32(MC_FUS_VM_FB_OFFSET, tmp);
2902         }
2903         tmp = ((rdev->mc.vram_end >> 24) & 0xFFFF) << 16;
2904         tmp |= ((rdev->mc.vram_start >> 24) & 0xFFFF);
2905         WREG32(MC_VM_FB_LOCATION, tmp);
2906         WREG32(HDP_NONSURFACE_BASE, (rdev->mc.vram_start >> 8));
2907         WREG32(HDP_NONSURFACE_INFO, (2 << 7) | (1 << 30));
2908         WREG32(HDP_NONSURFACE_SIZE, 0x3FFFFFFF);
2909         if (rdev->flags & RADEON_IS_AGP) {
2910                 WREG32(MC_VM_AGP_TOP, rdev->mc.gtt_end >> 16);
2911                 WREG32(MC_VM_AGP_BOT, rdev->mc.gtt_start >> 16);
2912                 WREG32(MC_VM_AGP_BASE, rdev->mc.agp_base >> 22);
2913         } else {
2914                 WREG32(MC_VM_AGP_BASE, 0);
2915                 WREG32(MC_VM_AGP_TOP, 0x0FFFFFFF);
2916                 WREG32(MC_VM_AGP_BOT, 0x0FFFFFFF);
2917         }
2918         if (evergreen_mc_wait_for_idle(rdev)) {
2919                 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
2920         }
2921         evergreen_mc_resume(rdev, &save);
2922         /* we need to own VRAM, so turn off the VGA renderer here
2923          * to stop it overwriting our objects */
2924         rv515_vga_render_disable(rdev);
2925 }
2926
2927 /*
2928  * CP.
2929  */
2930 void evergreen_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib)
2931 {
2932         struct radeon_ring *ring = &rdev->ring[ib->ring];
2933         u32 next_rptr;
2934
2935         /* set to DX10/11 mode */
2936         radeon_ring_write(ring, PACKET3(PACKET3_MODE_CONTROL, 0));
2937         radeon_ring_write(ring, 1);
2938
2939         if (ring->rptr_save_reg) {
2940                 next_rptr = ring->wptr + 3 + 4;
2941                 radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1));
2942                 radeon_ring_write(ring, ((ring->rptr_save_reg - 
2943                                           PACKET3_SET_CONFIG_REG_START) >> 2));
2944                 radeon_ring_write(ring, next_rptr);
2945         } else if (rdev->wb.enabled) {
2946                 next_rptr = ring->wptr + 5 + 4;
2947                 radeon_ring_write(ring, PACKET3(PACKET3_MEM_WRITE, 3));
2948                 radeon_ring_write(ring, ring->next_rptr_gpu_addr & 0xfffffffc);
2949                 radeon_ring_write(ring, (upper_32_bits(ring->next_rptr_gpu_addr) & 0xff) | (1 << 18));
2950                 radeon_ring_write(ring, next_rptr);
2951                 radeon_ring_write(ring, 0);
2952         }
2953
2954         radeon_ring_write(ring, PACKET3(PACKET3_INDIRECT_BUFFER, 2));
2955         radeon_ring_write(ring,
2956 #ifdef __BIG_ENDIAN
2957                           (2 << 0) |
2958 #endif
2959                           (ib->gpu_addr & 0xFFFFFFFC));
2960         radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF);
2961         radeon_ring_write(ring, ib->length_dw);
2962 }
2963
2964
2965 static int evergreen_cp_load_microcode(struct radeon_device *rdev)
2966 {
2967         const __be32 *fw_data;
2968         int i;
2969
2970         if (!rdev->me_fw || !rdev->pfp_fw)
2971                 return -EINVAL;
2972
2973         r700_cp_stop(rdev);
2974         WREG32(CP_RB_CNTL,
2975 #ifdef __BIG_ENDIAN
2976                BUF_SWAP_32BIT |
2977 #endif
2978                RB_NO_UPDATE | RB_BLKSZ(15) | RB_BUFSZ(3));
2979
2980         fw_data = (const __be32 *)rdev->pfp_fw->data;
2981         WREG32(CP_PFP_UCODE_ADDR, 0);
2982         for (i = 0; i < EVERGREEN_PFP_UCODE_SIZE; i++)
2983                 WREG32(CP_PFP_UCODE_DATA, be32_to_cpup(fw_data++));
2984         WREG32(CP_PFP_UCODE_ADDR, 0);
2985
2986         fw_data = (const __be32 *)rdev->me_fw->data;
2987         WREG32(CP_ME_RAM_WADDR, 0);
2988         for (i = 0; i < EVERGREEN_PM4_UCODE_SIZE; i++)
2989                 WREG32(CP_ME_RAM_DATA, be32_to_cpup(fw_data++));
2990
2991         WREG32(CP_PFP_UCODE_ADDR, 0);
2992         WREG32(CP_ME_RAM_WADDR, 0);
2993         WREG32(CP_ME_RAM_RADDR, 0);
2994         return 0;
2995 }
2996
2997 static int evergreen_cp_start(struct radeon_device *rdev)
2998 {
2999         struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
3000         int r, i;
3001         uint32_t cp_me;
3002
3003         r = radeon_ring_lock(rdev, ring, 7);
3004         if (r) {
3005                 DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
3006                 return r;
3007         }
3008         radeon_ring_write(ring, PACKET3(PACKET3_ME_INITIALIZE, 5));
3009         radeon_ring_write(ring, 0x1);
3010         radeon_ring_write(ring, 0x0);
3011         radeon_ring_write(ring, rdev->config.evergreen.max_hw_contexts - 1);
3012         radeon_ring_write(ring, PACKET3_ME_INITIALIZE_DEVICE_ID(1));
3013         radeon_ring_write(ring, 0);
3014         radeon_ring_write(ring, 0);
3015         radeon_ring_unlock_commit(rdev, ring, false);
3016
3017         cp_me = 0xff;
3018         WREG32(CP_ME_CNTL, cp_me);
3019
3020         r = radeon_ring_lock(rdev, ring, evergreen_default_size + 19);
3021         if (r) {
3022                 DRM_ERROR("radeon: cp failed to lock ring (%d).\n", r);
3023                 return r;
3024         }
3025
3026         /* setup clear context state */
3027         radeon_ring_write(ring, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
3028         radeon_ring_write(ring, PACKET3_PREAMBLE_BEGIN_CLEAR_STATE);
3029
3030         for (i = 0; i < evergreen_default_size; i++)
3031                 radeon_ring_write(ring, evergreen_default_state[i]);
3032
3033         radeon_ring_write(ring, PACKET3(PACKET3_PREAMBLE_CNTL, 0));
3034         radeon_ring_write(ring, PACKET3_PREAMBLE_END_CLEAR_STATE);
3035
3036         /* set clear context state */
3037         radeon_ring_write(ring, PACKET3(PACKET3_CLEAR_STATE, 0));
3038         radeon_ring_write(ring, 0);
3039
3040         /* SQ_VTX_BASE_VTX_LOC */
3041         radeon_ring_write(ring, 0xc0026f00);
3042         radeon_ring_write(ring, 0x00000000);
3043         radeon_ring_write(ring, 0x00000000);
3044         radeon_ring_write(ring, 0x00000000);
3045
3046         /* Clear consts */
3047         radeon_ring_write(ring, 0xc0036f00);
3048         radeon_ring_write(ring, 0x00000bc4);
3049         radeon_ring_write(ring, 0xffffffff);
3050         radeon_ring_write(ring, 0xffffffff);
3051         radeon_ring_write(ring, 0xffffffff);
3052
3053         radeon_ring_write(ring, 0xc0026900);
3054         radeon_ring_write(ring, 0x00000316);
3055         radeon_ring_write(ring, 0x0000000e); /* VGT_VERTEX_REUSE_BLOCK_CNTL */
3056         radeon_ring_write(ring, 0x00000010); /*  */
3057
3058         radeon_ring_unlock_commit(rdev, ring, false);
3059
3060         return 0;
3061 }
3062
3063 static int evergreen_cp_resume(struct radeon_device *rdev)
3064 {
3065         struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
3066         u32 tmp;
3067         u32 rb_bufsz;
3068         int r;
3069
3070         /* Reset cp; if cp is reset, then PA, SH, VGT also need to be reset */
3071         WREG32(GRBM_SOFT_RESET, (SOFT_RESET_CP |
3072                                  SOFT_RESET_PA |
3073                                  SOFT_RESET_SH |
3074                                  SOFT_RESET_VGT |
3075                                  SOFT_RESET_SPI |
3076                                  SOFT_RESET_SX));
3077         RREG32(GRBM_SOFT_RESET);
3078         mdelay(15);
3079         WREG32(GRBM_SOFT_RESET, 0);
3080         RREG32(GRBM_SOFT_RESET);
3081
3082         /* Set ring buffer size */
3083         rb_bufsz = order_base_2(ring->ring_size / 8);
3084         tmp = (order_base_2(RADEON_GPU_PAGE_SIZE/8) << 8) | rb_bufsz;
3085 #ifdef __BIG_ENDIAN
3086         tmp |= BUF_SWAP_32BIT;
3087 #endif
3088         WREG32(CP_RB_CNTL, tmp);
3089         WREG32(CP_SEM_WAIT_TIMER, 0x0);
3090         WREG32(CP_SEM_INCOMPLETE_TIMER_CNTL, 0x0);
3091
3092         /* Set the write pointer delay */
3093         WREG32(CP_RB_WPTR_DELAY, 0);
3094
3095         /* Initialize the ring buffer's read and write pointers */
3096         WREG32(CP_RB_CNTL, tmp | RB_RPTR_WR_ENA);
3097         WREG32(CP_RB_RPTR_WR, 0);
3098         ring->wptr = 0;
3099         WREG32(CP_RB_WPTR, ring->wptr);
3100
3101         /* set the wb address whether it's enabled or not */
3102         WREG32(CP_RB_RPTR_ADDR,
3103                ((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC));
3104         WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
3105         WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
3106
3107         if (rdev->wb.enabled)
3108                 WREG32(SCRATCH_UMSK, 0xff);
3109         else {
3110                 tmp |= RB_NO_UPDATE;
3111                 WREG32(SCRATCH_UMSK, 0);
3112         }
3113
3114         mdelay(1);
3115         WREG32(CP_RB_CNTL, tmp);
3116
3117         WREG32(CP_RB_BASE, ring->gpu_addr >> 8);
3118         WREG32(CP_DEBUG, (1 << 27) | (1 << 28));
3119
3120         evergreen_cp_start(rdev);
3121         ring->ready = true;
3122         r = radeon_ring_test(rdev, RADEON_RING_TYPE_GFX_INDEX, ring);
3123         if (r) {
3124                 ring->ready = false;
3125                 return r;
3126         }
3127         return 0;
3128 }
3129
3130 /*
3131  * Core functions
3132  */
3133 static void evergreen_gpu_init(struct radeon_device *rdev)
3134 {
3135         u32 gb_addr_config;
3136         u32 mc_shared_chmap, mc_arb_ramcfg;
3137         u32 sx_debug_1;
3138         u32 smx_dc_ctl0;
3139         u32 sq_config;
3140         u32 sq_lds_resource_mgmt;
3141         u32 sq_gpr_resource_mgmt_1;
3142         u32 sq_gpr_resource_mgmt_2;
3143         u32 sq_gpr_resource_mgmt_3;
3144         u32 sq_thread_resource_mgmt;
3145         u32 sq_thread_resource_mgmt_2;
3146         u32 sq_stack_resource_mgmt_1;
3147         u32 sq_stack_resource_mgmt_2;
3148         u32 sq_stack_resource_mgmt_3;
3149         u32 vgt_cache_invalidation;
3150         u32 hdp_host_path_cntl, tmp;
3151         u32 disabled_rb_mask;
3152         int i, j, ps_thread_count;
3153
3154         switch (rdev->family) {
3155         case CHIP_CYPRESS:
3156         case CHIP_HEMLOCK:
3157                 rdev->config.evergreen.num_ses = 2;
3158                 rdev->config.evergreen.max_pipes = 4;
3159                 rdev->config.evergreen.max_tile_pipes = 8;
3160                 rdev->config.evergreen.max_simds = 10;
3161                 rdev->config.evergreen.max_backends = 4 * rdev->config.evergreen.num_ses;
3162                 rdev->config.evergreen.max_gprs = 256;
3163                 rdev->config.evergreen.max_threads = 248;
3164                 rdev->config.evergreen.max_gs_threads = 32;
3165                 rdev->config.evergreen.max_stack_entries = 512;
3166                 rdev->config.evergreen.sx_num_of_sets = 4;
3167                 rdev->config.evergreen.sx_max_export_size = 256;
3168                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3169                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3170                 rdev->config.evergreen.max_hw_contexts = 8;
3171                 rdev->config.evergreen.sq_num_cf_insts = 2;
3172
3173                 rdev->config.evergreen.sc_prim_fifo_size = 0x100;
3174                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3175                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3176                 gb_addr_config = CYPRESS_GB_ADDR_CONFIG_GOLDEN;
3177                 break;
3178         case CHIP_JUNIPER:
3179                 rdev->config.evergreen.num_ses = 1;
3180                 rdev->config.evergreen.max_pipes = 4;
3181                 rdev->config.evergreen.max_tile_pipes = 4;
3182                 rdev->config.evergreen.max_simds = 10;
3183                 rdev->config.evergreen.max_backends = 4 * rdev->config.evergreen.num_ses;
3184                 rdev->config.evergreen.max_gprs = 256;
3185                 rdev->config.evergreen.max_threads = 248;
3186                 rdev->config.evergreen.max_gs_threads = 32;
3187                 rdev->config.evergreen.max_stack_entries = 512;
3188                 rdev->config.evergreen.sx_num_of_sets = 4;
3189                 rdev->config.evergreen.sx_max_export_size = 256;
3190                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3191                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3192                 rdev->config.evergreen.max_hw_contexts = 8;
3193                 rdev->config.evergreen.sq_num_cf_insts = 2;
3194
3195                 rdev->config.evergreen.sc_prim_fifo_size = 0x100;
3196                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3197                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3198                 gb_addr_config = JUNIPER_GB_ADDR_CONFIG_GOLDEN;
3199                 break;
3200         case CHIP_REDWOOD:
3201                 rdev->config.evergreen.num_ses = 1;
3202                 rdev->config.evergreen.max_pipes = 4;
3203                 rdev->config.evergreen.max_tile_pipes = 4;
3204                 rdev->config.evergreen.max_simds = 5;
3205                 rdev->config.evergreen.max_backends = 2 * rdev->config.evergreen.num_ses;
3206                 rdev->config.evergreen.max_gprs = 256;
3207                 rdev->config.evergreen.max_threads = 248;
3208                 rdev->config.evergreen.max_gs_threads = 32;
3209                 rdev->config.evergreen.max_stack_entries = 256;
3210                 rdev->config.evergreen.sx_num_of_sets = 4;
3211                 rdev->config.evergreen.sx_max_export_size = 256;
3212                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3213                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3214                 rdev->config.evergreen.max_hw_contexts = 8;
3215                 rdev->config.evergreen.sq_num_cf_insts = 2;
3216
3217                 rdev->config.evergreen.sc_prim_fifo_size = 0x100;
3218                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3219                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3220                 gb_addr_config = REDWOOD_GB_ADDR_CONFIG_GOLDEN;
3221                 break;
3222         case CHIP_CEDAR:
3223         default:
3224                 rdev->config.evergreen.num_ses = 1;
3225                 rdev->config.evergreen.max_pipes = 2;
3226                 rdev->config.evergreen.max_tile_pipes = 2;
3227                 rdev->config.evergreen.max_simds = 2;
3228                 rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
3229                 rdev->config.evergreen.max_gprs = 256;
3230                 rdev->config.evergreen.max_threads = 192;
3231                 rdev->config.evergreen.max_gs_threads = 16;
3232                 rdev->config.evergreen.max_stack_entries = 256;
3233                 rdev->config.evergreen.sx_num_of_sets = 4;
3234                 rdev->config.evergreen.sx_max_export_size = 128;
3235                 rdev->config.evergreen.sx_max_export_pos_size = 32;
3236                 rdev->config.evergreen.sx_max_export_smx_size = 96;
3237                 rdev->config.evergreen.max_hw_contexts = 4;
3238                 rdev->config.evergreen.sq_num_cf_insts = 1;
3239
3240                 rdev->config.evergreen.sc_prim_fifo_size = 0x40;
3241                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3242                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3243                 gb_addr_config = CEDAR_GB_ADDR_CONFIG_GOLDEN;
3244                 break;
3245         case CHIP_PALM:
3246                 rdev->config.evergreen.num_ses = 1;
3247                 rdev->config.evergreen.max_pipes = 2;
3248                 rdev->config.evergreen.max_tile_pipes = 2;
3249                 rdev->config.evergreen.max_simds = 2;
3250                 rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
3251                 rdev->config.evergreen.max_gprs = 256;
3252                 rdev->config.evergreen.max_threads = 192;
3253                 rdev->config.evergreen.max_gs_threads = 16;
3254                 rdev->config.evergreen.max_stack_entries = 256;
3255                 rdev->config.evergreen.sx_num_of_sets = 4;
3256                 rdev->config.evergreen.sx_max_export_size = 128;
3257                 rdev->config.evergreen.sx_max_export_pos_size = 32;
3258                 rdev->config.evergreen.sx_max_export_smx_size = 96;
3259                 rdev->config.evergreen.max_hw_contexts = 4;
3260                 rdev->config.evergreen.sq_num_cf_insts = 1;
3261
3262                 rdev->config.evergreen.sc_prim_fifo_size = 0x40;
3263                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3264                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3265                 gb_addr_config = CEDAR_GB_ADDR_CONFIG_GOLDEN;
3266                 break;
3267         case CHIP_SUMO:
3268                 rdev->config.evergreen.num_ses = 1;
3269                 rdev->config.evergreen.max_pipes = 4;
3270                 rdev->config.evergreen.max_tile_pipes = 4;
3271                 if (rdev->pdev->device == 0x9648)
3272                         rdev->config.evergreen.max_simds = 3;
3273                 else if ((rdev->pdev->device == 0x9647) ||
3274                          (rdev->pdev->device == 0x964a))
3275                         rdev->config.evergreen.max_simds = 4;
3276                 else
3277                         rdev->config.evergreen.max_simds = 5;
3278                 rdev->config.evergreen.max_backends = 2 * rdev->config.evergreen.num_ses;
3279                 rdev->config.evergreen.max_gprs = 256;
3280                 rdev->config.evergreen.max_threads = 248;
3281                 rdev->config.evergreen.max_gs_threads = 32;
3282                 rdev->config.evergreen.max_stack_entries = 256;
3283                 rdev->config.evergreen.sx_num_of_sets = 4;
3284                 rdev->config.evergreen.sx_max_export_size = 256;
3285                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3286                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3287                 rdev->config.evergreen.max_hw_contexts = 8;
3288                 rdev->config.evergreen.sq_num_cf_insts = 2;
3289
3290                 rdev->config.evergreen.sc_prim_fifo_size = 0x40;
3291                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3292                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3293                 gb_addr_config = SUMO_GB_ADDR_CONFIG_GOLDEN;
3294                 break;
3295         case CHIP_SUMO2:
3296                 rdev->config.evergreen.num_ses = 1;
3297                 rdev->config.evergreen.max_pipes = 4;
3298                 rdev->config.evergreen.max_tile_pipes = 4;
3299                 rdev->config.evergreen.max_simds = 2;
3300                 rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
3301                 rdev->config.evergreen.max_gprs = 256;
3302                 rdev->config.evergreen.max_threads = 248;
3303                 rdev->config.evergreen.max_gs_threads = 32;
3304                 rdev->config.evergreen.max_stack_entries = 512;
3305                 rdev->config.evergreen.sx_num_of_sets = 4;
3306                 rdev->config.evergreen.sx_max_export_size = 256;
3307                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3308                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3309                 rdev->config.evergreen.max_hw_contexts = 4;
3310                 rdev->config.evergreen.sq_num_cf_insts = 2;
3311
3312                 rdev->config.evergreen.sc_prim_fifo_size = 0x40;
3313                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3314                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3315                 gb_addr_config = SUMO2_GB_ADDR_CONFIG_GOLDEN;
3316                 break;
3317         case CHIP_BARTS:
3318                 rdev->config.evergreen.num_ses = 2;
3319                 rdev->config.evergreen.max_pipes = 4;
3320                 rdev->config.evergreen.max_tile_pipes = 8;
3321                 rdev->config.evergreen.max_simds = 7;
3322                 rdev->config.evergreen.max_backends = 4 * rdev->config.evergreen.num_ses;
3323                 rdev->config.evergreen.max_gprs = 256;
3324                 rdev->config.evergreen.max_threads = 248;
3325                 rdev->config.evergreen.max_gs_threads = 32;
3326                 rdev->config.evergreen.max_stack_entries = 512;
3327                 rdev->config.evergreen.sx_num_of_sets = 4;
3328                 rdev->config.evergreen.sx_max_export_size = 256;
3329                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3330                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3331                 rdev->config.evergreen.max_hw_contexts = 8;
3332                 rdev->config.evergreen.sq_num_cf_insts = 2;
3333
3334                 rdev->config.evergreen.sc_prim_fifo_size = 0x100;
3335                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3336                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3337                 gb_addr_config = BARTS_GB_ADDR_CONFIG_GOLDEN;
3338                 break;
3339         case CHIP_TURKS:
3340                 rdev->config.evergreen.num_ses = 1;
3341                 rdev->config.evergreen.max_pipes = 4;
3342                 rdev->config.evergreen.max_tile_pipes = 4;
3343                 rdev->config.evergreen.max_simds = 6;
3344                 rdev->config.evergreen.max_backends = 2 * rdev->config.evergreen.num_ses;
3345                 rdev->config.evergreen.max_gprs = 256;
3346                 rdev->config.evergreen.max_threads = 248;
3347                 rdev->config.evergreen.max_gs_threads = 32;
3348                 rdev->config.evergreen.max_stack_entries = 256;
3349                 rdev->config.evergreen.sx_num_of_sets = 4;
3350                 rdev->config.evergreen.sx_max_export_size = 256;
3351                 rdev->config.evergreen.sx_max_export_pos_size = 64;
3352                 rdev->config.evergreen.sx_max_export_smx_size = 192;
3353                 rdev->config.evergreen.max_hw_contexts = 8;
3354                 rdev->config.evergreen.sq_num_cf_insts = 2;
3355
3356                 rdev->config.evergreen.sc_prim_fifo_size = 0x100;
3357                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3358                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3359                 gb_addr_config = TURKS_GB_ADDR_CONFIG_GOLDEN;
3360                 break;
3361         case CHIP_CAICOS:
3362                 rdev->config.evergreen.num_ses = 1;
3363                 rdev->config.evergreen.max_pipes = 2;
3364                 rdev->config.evergreen.max_tile_pipes = 2;
3365                 rdev->config.evergreen.max_simds = 2;
3366                 rdev->config.evergreen.max_backends = 1 * rdev->config.evergreen.num_ses;
3367                 rdev->config.evergreen.max_gprs = 256;
3368                 rdev->config.evergreen.max_threads = 192;
3369                 rdev->config.evergreen.max_gs_threads = 16;
3370                 rdev->config.evergreen.max_stack_entries = 256;
3371                 rdev->config.evergreen.sx_num_of_sets = 4;
3372                 rdev->config.evergreen.sx_max_export_size = 128;
3373                 rdev->config.evergreen.sx_max_export_pos_size = 32;
3374                 rdev->config.evergreen.sx_max_export_smx_size = 96;
3375                 rdev->config.evergreen.max_hw_contexts = 4;
3376                 rdev->config.evergreen.sq_num_cf_insts = 1;
3377
3378                 rdev->config.evergreen.sc_prim_fifo_size = 0x40;
3379                 rdev->config.evergreen.sc_hiz_tile_fifo_size = 0x30;
3380                 rdev->config.evergreen.sc_earlyz_tile_fifo_size = 0x130;
3381                 gb_addr_config = CAICOS_GB_ADDR_CONFIG_GOLDEN;
3382                 break;
3383         }
3384
3385         /* Initialize HDP */
3386         for (i = 0, j = 0; i < 32; i++, j += 0x18) {
3387                 WREG32((0x2c14 + j), 0x00000000);
3388                 WREG32((0x2c18 + j), 0x00000000);
3389                 WREG32((0x2c1c + j), 0x00000000);
3390                 WREG32((0x2c20 + j), 0x00000000);
3391                 WREG32((0x2c24 + j), 0x00000000);
3392         }
3393
3394         WREG32(GRBM_CNTL, GRBM_READ_TIMEOUT(0xff));
3395         WREG32(SRBM_INT_CNTL, 0x1);
3396         WREG32(SRBM_INT_ACK, 0x1);
3397
3398         evergreen_fix_pci_max_read_req_size(rdev);
3399
3400         mc_shared_chmap = RREG32(MC_SHARED_CHMAP);
3401         if ((rdev->family == CHIP_PALM) ||
3402             (rdev->family == CHIP_SUMO) ||
3403             (rdev->family == CHIP_SUMO2))
3404                 mc_arb_ramcfg = RREG32(FUS_MC_ARB_RAMCFG);
3405         else
3406                 mc_arb_ramcfg = RREG32(MC_ARB_RAMCFG);
3407
3408         /* setup tiling info dword.  gb_addr_config is not adequate since it does
3409          * not have bank info, so create a custom tiling dword.
3410          * bits 3:0   num_pipes
3411          * bits 7:4   num_banks
3412          * bits 11:8  group_size
3413          * bits 15:12 row_size
3414          */
3415         rdev->config.evergreen.tile_config = 0;
3416         switch (rdev->config.evergreen.max_tile_pipes) {
3417         case 1:
3418         default:
3419                 rdev->config.evergreen.tile_config |= (0 << 0);
3420                 break;
3421         case 2:
3422                 rdev->config.evergreen.tile_config |= (1 << 0);
3423                 break;
3424         case 4:
3425                 rdev->config.evergreen.tile_config |= (2 << 0);
3426                 break;
3427         case 8:
3428                 rdev->config.evergreen.tile_config |= (3 << 0);
3429                 break;
3430         }
3431         /* num banks is 8 on all fusion asics. 0 = 4, 1 = 8, 2 = 16 */
3432         if (rdev->flags & RADEON_IS_IGP)
3433                 rdev->config.evergreen.tile_config |= 1 << 4;
3434         else {
3435                 switch ((mc_arb_ramcfg & NOOFBANK_MASK) >> NOOFBANK_SHIFT) {
3436                 case 0: /* four banks */
3437                         rdev->config.evergreen.tile_config |= 0 << 4;
3438                         break;
3439                 case 1: /* eight banks */
3440                         rdev->config.evergreen.tile_config |= 1 << 4;
3441                         break;
3442                 case 2: /* sixteen banks */
3443                 default:
3444                         rdev->config.evergreen.tile_config |= 2 << 4;
3445                         break;
3446                 }
3447         }
3448         rdev->config.evergreen.tile_config |= 0 << 8;
3449         rdev->config.evergreen.tile_config |=
3450                 ((gb_addr_config & 0x30000000) >> 28) << 12;
3451
3452         if ((rdev->family >= CHIP_CEDAR) && (rdev->family <= CHIP_HEMLOCK)) {
3453                 u32 efuse_straps_4;
3454                 u32 efuse_straps_3;
3455
3456                 efuse_straps_4 = RREG32_RCU(0x204);
3457                 efuse_straps_3 = RREG32_RCU(0x203);
3458                 tmp = (((efuse_straps_4 & 0xf) << 4) |
3459                       ((efuse_straps_3 & 0xf0000000) >> 28));
3460         } else {
3461                 tmp = 0;
3462                 for (i = (rdev->config.evergreen.num_ses - 1); i >= 0; i--) {
3463                         u32 rb_disable_bitmap;
3464
3465                         WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
3466                         WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
3467                         rb_disable_bitmap = (RREG32(CC_RB_BACKEND_DISABLE) & 0x00ff0000) >> 16;
3468                         tmp <<= 4;
3469                         tmp |= rb_disable_bitmap;
3470                 }
3471         }
3472         /* enabled rb are just the one not disabled :) */
3473         disabled_rb_mask = tmp;
3474         tmp = 0;
3475         for (i = 0; i < rdev->config.evergreen.max_backends; i++)
3476                 tmp |= (1 << i);
3477         /* if all the backends are disabled, fix it up here */
3478         if ((disabled_rb_mask & tmp) == tmp) {
3479                 for (i = 0; i < rdev->config.evergreen.max_backends; i++)
3480                         disabled_rb_mask &= ~(1 << i);
3481         }
3482
3483         for (i = 0; i < rdev->config.evergreen.num_ses; i++) {
3484                 u32 simd_disable_bitmap;
3485
3486                 WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
3487                 WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_INDEX(i));
3488                 simd_disable_bitmap = (RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffff0000) >> 16;
3489                 simd_disable_bitmap |= 0xffffffff << rdev->config.evergreen.max_simds;
3490                 tmp <<= 16;
3491                 tmp |= simd_disable_bitmap;
3492         }
3493         rdev->config.evergreen.active_simds = hweight32(~tmp);
3494
3495         WREG32(GRBM_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_BROADCAST_WRITES);
3496         WREG32(RLC_GFX_INDEX, INSTANCE_BROADCAST_WRITES | SE_BROADCAST_WRITES);
3497
3498         WREG32(GB_ADDR_CONFIG, gb_addr_config);
3499         WREG32(DMIF_ADDR_CONFIG, gb_addr_config);
3500         WREG32(HDP_ADDR_CONFIG, gb_addr_config);
3501         WREG32(DMA_TILING_CONFIG, gb_addr_config);
3502         WREG32(UVD_UDEC_ADDR_CONFIG, gb_addr_config);
3503         WREG32(UVD_UDEC_DB_ADDR_CONFIG, gb_addr_config);
3504         WREG32(UVD_UDEC_DBW_ADDR_CONFIG, gb_addr_config);
3505
3506         if ((rdev->config.evergreen.max_backends == 1) &&
3507             (rdev->flags & RADEON_IS_IGP)) {
3508                 if ((disabled_rb_mask & 3) == 1) {
3509                         /* RB0 disabled, RB1 enabled */
3510                         tmp = 0x11111111;
3511                 } else {
3512                         /* RB1 disabled, RB0 enabled */
3513                         tmp = 0x00000000;
3514                 }
3515         } else {
3516                 tmp = gb_addr_config & NUM_PIPES_MASK;
3517                 tmp = r6xx_remap_render_backend(rdev, tmp, rdev->config.evergreen.max_backends,
3518                                                 EVERGREEN_MAX_BACKENDS, disabled_rb_mask);
3519         }
3520         rdev->config.evergreen.backend_map = tmp;
3521         WREG32(GB_BACKEND_MAP, tmp);
3522
3523         WREG32(CGTS_SYS_TCC_DISABLE, 0);
3524         WREG32(CGTS_TCC_DISABLE, 0);
3525         WREG32(CGTS_USER_SYS_TCC_DISABLE, 0);
3526         WREG32(CGTS_USER_TCC_DISABLE, 0);
3527
3528         /* set HW defaults for 3D engine */
3529         WREG32(CP_QUEUE_THRESHOLDS, (ROQ_IB1_START(0x16) |
3530                                      ROQ_IB2_START(0x2b)));
3531
3532         WREG32(CP_MEQ_THRESHOLDS, STQ_SPLIT(0x30));
3533
3534         WREG32(TA_CNTL_AUX, (DISABLE_CUBE_ANISO |
3535                              SYNC_GRADIENT |
3536                              SYNC_WALKER |
3537                              SYNC_ALIGNER));
3538
3539         sx_debug_1 = RREG32(SX_DEBUG_1);
3540         sx_debug_1 |= ENABLE_NEW_SMX_ADDRESS;
3541         WREG32(SX_DEBUG_1, sx_debug_1);
3542
3543
3544         smx_dc_ctl0 = RREG32(SMX_DC_CTL0);
3545         smx_dc_ctl0 &= ~NUMBER_OF_SETS(0x1ff);
3546         smx_dc_ctl0 |= NUMBER_OF_SETS(rdev->config.evergreen.sx_num_of_sets);
3547         WREG32(SMX_DC_CTL0, smx_dc_ctl0);
3548
3549         if (rdev->family <= CHIP_SUMO2)
3550                 WREG32(SMX_SAR_CTL0, 0x00010000);
3551
3552         WREG32(SX_EXPORT_BUFFER_SIZES, (COLOR_BUFFER_SIZE((rdev->config.evergreen.sx_max_export_size / 4) - 1) |
3553                                         POSITION_BUFFER_SIZE((rdev->config.evergreen.sx_max_export_pos_size / 4) - 1) |
3554                                         SMX_BUFFER_SIZE((rdev->config.evergreen.sx_max_export_smx_size / 4) - 1)));
3555
3556         WREG32(PA_SC_FIFO_SIZE, (SC_PRIM_FIFO_SIZE(rdev->config.evergreen.sc_prim_fifo_size) |
3557                                  SC_HIZ_TILE_FIFO_SIZE(rdev->config.evergreen.sc_hiz_tile_fifo_size) |
3558                                  SC_EARLYZ_TILE_FIFO_SIZE(rdev->config.evergreen.sc_earlyz_tile_fifo_size)));
3559
3560         WREG32(VGT_NUM_INSTANCES, 1);
3561         WREG32(SPI_CONFIG_CNTL, 0);
3562         WREG32(SPI_CONFIG_CNTL_1, VTX_DONE_DELAY(4));
3563         WREG32(CP_PERFMON_CNTL, 0);
3564
3565         WREG32(SQ_MS_FIFO_SIZES, (CACHE_FIFO_SIZE(16 * rdev->config.evergreen.sq_num_cf_insts) |
3566                                   FETCH_FIFO_HIWATER(0x4) |
3567                                   DONE_FIFO_HIWATER(0xe0) |
3568                                   ALU_UPDATE_FIFO_HIWATER(0x8)));
3569
3570         sq_config = RREG32(SQ_CONFIG);
3571         sq_config &= ~(PS_PRIO(3) |
3572                        VS_PRIO(3) |
3573                        GS_PRIO(3) |
3574                        ES_PRIO(3));
3575         sq_config |= (VC_ENABLE |
3576                       EXPORT_SRC_C |
3577                       PS_PRIO(0) |
3578                       VS_PRIO(1) |
3579                       GS_PRIO(2) |
3580                       ES_PRIO(3));
3581
3582         switch (rdev->family) {
3583         case CHIP_CEDAR:
3584         case CHIP_PALM:
3585         case CHIP_SUMO:
3586         case CHIP_SUMO2:
3587         case CHIP_CAICOS:
3588                 /* no vertex cache */
3589                 sq_config &= ~VC_ENABLE;
3590                 break;
3591         default:
3592                 break;
3593         }
3594
3595         sq_lds_resource_mgmt = RREG32(SQ_LDS_RESOURCE_MGMT);
3596
3597         sq_gpr_resource_mgmt_1 = NUM_PS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2))* 12 / 32);
3598         sq_gpr_resource_mgmt_1 |= NUM_VS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 6 / 32);
3599         sq_gpr_resource_mgmt_1 |= NUM_CLAUSE_TEMP_GPRS(4);
3600         sq_gpr_resource_mgmt_2 = NUM_GS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 4 / 32);
3601         sq_gpr_resource_mgmt_2 |= NUM_ES_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 4 / 32);
3602         sq_gpr_resource_mgmt_3 = NUM_HS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 3 / 32);
3603         sq_gpr_resource_mgmt_3 |= NUM_LS_GPRS((rdev->config.evergreen.max_gprs - (4 * 2)) * 3 / 32);
3604
3605         switch (rdev->family) {
3606         case CHIP_CEDAR:
3607         case CHIP_PALM:
3608         case CHIP_SUMO:
3609         case CHIP_SUMO2:
3610                 ps_thread_count = 96;
3611                 break;
3612         default:
3613                 ps_thread_count = 128;
3614                 break;
3615         }
3616
3617         sq_thread_resource_mgmt = NUM_PS_THREADS(ps_thread_count);
3618         sq_thread_resource_mgmt |= NUM_VS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
3619         sq_thread_resource_mgmt |= NUM_GS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
3620         sq_thread_resource_mgmt |= NUM_ES_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
3621         sq_thread_resource_mgmt_2 = NUM_HS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
3622         sq_thread_resource_mgmt_2 |= NUM_LS_THREADS((((rdev->config.evergreen.max_threads - ps_thread_count) / 6) / 8) * 8);
3623
3624         sq_stack_resource_mgmt_1 = NUM_PS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
3625         sq_stack_resource_mgmt_1 |= NUM_VS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
3626         sq_stack_resource_mgmt_2 = NUM_GS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
3627         sq_stack_resource_mgmt_2 |= NUM_ES_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
3628         sq_stack_resource_mgmt_3 = NUM_HS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
3629         sq_stack_resource_mgmt_3 |= NUM_LS_STACK_ENTRIES((rdev->config.evergreen.max_stack_entries * 1) / 6);
3630
3631         WREG32(SQ_CONFIG, sq_config);
3632         WREG32(SQ_GPR_RESOURCE_MGMT_1, sq_gpr_resource_mgmt_1);
3633         WREG32(SQ_GPR_RESOURCE_MGMT_2, sq_gpr_resource_mgmt_2);
3634         WREG32(SQ_GPR_RESOURCE_MGMT_3, sq_gpr_resource_mgmt_3);
3635         WREG32(SQ_THREAD_RESOURCE_MGMT, sq_thread_resource_mgmt);
3636         WREG32(SQ_THREAD_RESOURCE_MGMT_2, sq_thread_resource_mgmt_2);
3637         WREG32(SQ_STACK_RESOURCE_MGMT_1, sq_stack_resource_mgmt_1);
3638         WREG32(SQ_STACK_RESOURCE_MGMT_2, sq_stack_resource_mgmt_2);
3639         WREG32(SQ_STACK_RESOURCE_MGMT_3, sq_stack_resource_mgmt_3);
3640         WREG32(SQ_DYN_GPR_CNTL_PS_FLUSH_REQ, 0);
3641         WREG32(SQ_LDS_RESOURCE_MGMT, sq_lds_resource_mgmt);
3642
3643         WREG32(PA_SC_FORCE_EOV_MAX_CNTS, (FORCE_EOV_MAX_CLK_CNT(4095) |
3644                                           FORCE_EOV_MAX_REZ_CNT(255)));
3645
3646         switch (rdev->family) {
3647         case CHIP_CEDAR:
3648         case CHIP_PALM:
3649         case CHIP_SUMO:
3650         case CHIP_SUMO2:
3651         case CHIP_CAICOS:
3652                 vgt_cache_invalidation = CACHE_INVALIDATION(TC_ONLY);
3653                 break;
3654         default:
3655                 vgt_cache_invalidation = CACHE_INVALIDATION(VC_AND_TC);
3656                 break;
3657         }
3658         vgt_cache_invalidation |= AUTO_INVLD_EN(ES_AND_GS_AUTO);
3659         WREG32(VGT_CACHE_INVALIDATION, vgt_cache_invalidation);
3660
3661         WREG32(VGT_GS_VERTEX_REUSE, 16);
3662         WREG32(PA_SU_LINE_STIPPLE_VALUE, 0);
3663         WREG32(PA_SC_LINE_STIPPLE_STATE, 0);
3664
3665         WREG32(VGT_VERTEX_REUSE_BLOCK_CNTL, 14);
3666         WREG32(VGT_OUT_DEALLOC_CNTL, 16);
3667
3668         WREG32(CB_PERF_CTR0_SEL_0, 0);
3669         WREG32(CB_PERF_CTR0_SEL_1, 0);
3670         WREG32(CB_PERF_CTR1_SEL_0, 0);
3671         WREG32(CB_PERF_CTR1_SEL_1, 0);
3672         WREG32(CB_PERF_CTR2_SEL_0, 0);
3673         WREG32(CB_PERF_CTR2_SEL_1, 0);
3674         WREG32(CB_PERF_CTR3_SEL_0, 0);
3675         WREG32(CB_PERF_CTR3_SEL_1, 0);
3676
3677         /* clear render buffer base addresses */
3678         WREG32(CB_COLOR0_BASE, 0);
3679         WREG32(CB_COLOR1_BASE, 0);
3680         WREG32(CB_COLOR2_BASE, 0);
3681         WREG32(CB_COLOR3_BASE, 0);
3682         WREG32(CB_COLOR4_BASE, 0);
3683         WREG32(CB_COLOR5_BASE, 0);
3684         WREG32(CB_COLOR6_BASE, 0);
3685         WREG32(CB_COLOR7_BASE, 0);
3686         WREG32(CB_COLOR8_BASE, 0);
3687         WREG32(CB_COLOR9_BASE, 0);
3688         WREG32(CB_COLOR10_BASE, 0);
3689         WREG32(CB_COLOR11_BASE, 0);
3690
3691         /* set the shader const cache sizes to 0 */
3692         for (i = SQ_ALU_CONST_BUFFER_SIZE_PS_0; i < 0x28200; i += 4)
3693                 WREG32(i, 0);
3694         for (i = SQ_ALU_CONST_BUFFER_SIZE_HS_0; i < 0x29000; i += 4)
3695                 WREG32(i, 0);
3696
3697         tmp = RREG32(HDP_MISC_CNTL);
3698         tmp |= HDP_FLUSH_INVALIDATE_CACHE;
3699         WREG32(HDP_MISC_CNTL, tmp);
3700
3701         hdp_host_path_cntl = RREG32(HDP_HOST_PATH_CNTL);
3702         WREG32(HDP_HOST_PATH_CNTL, hdp_host_path_cntl);
3703
3704         WREG32(PA_CL_ENHANCE, CLIP_VTX_REORDER_ENA | NUM_CLIP_SEQ(3));
3705
3706         udelay(50);
3707
3708 }
3709
3710 int evergreen_mc_init(struct radeon_device *rdev)
3711 {
3712         u32 tmp;
3713         int chansize, numchan;
3714
3715         /* Get VRAM informations */
3716         rdev->mc.vram_is_ddr = true;
3717         if ((rdev->family == CHIP_PALM) ||
3718             (rdev->family == CHIP_SUMO) ||
3719             (rdev->family == CHIP_SUMO2))
3720                 tmp = RREG32(FUS_MC_ARB_RAMCFG);
3721         else
3722                 tmp = RREG32(MC_ARB_RAMCFG);
3723         if (tmp & CHANSIZE_OVERRIDE) {
3724                 chansize = 16;
3725         } else if (tmp & CHANSIZE_MASK) {
3726                 chansize = 64;
3727         } else {
3728                 chansize = 32;
3729         }
3730         tmp = RREG32(MC_SHARED_CHMAP);
3731         switch ((tmp & NOOFCHAN_MASK) >> NOOFCHAN_SHIFT) {
3732         case 0:
3733         default:
3734                 numchan = 1;
3735                 break;
3736         case 1:
3737                 numchan = 2;
3738                 break;
3739         case 2:
3740                 numchan = 4;
3741                 break;
3742         case 3:
3743                 numchan = 8;
3744                 break;
3745         }
3746         rdev->mc.vram_width = numchan * chansize;
3747         /* Could aper size report 0 ? */
3748         rdev->mc.aper_base = pci_resource_start(rdev->pdev, 0);
3749         rdev->mc.aper_size = pci_resource_len(rdev->pdev, 0);
3750         /* Setup GPU memory space */
3751         if ((rdev->family == CHIP_PALM) ||
3752             (rdev->family == CHIP_SUMO) ||
3753             (rdev->family == CHIP_SUMO2)) {
3754                 /* size in bytes on fusion */
3755                 rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE);
3756                 rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE);
3757         } else {
3758                 /* size in MB on evergreen/cayman/tn */
3759                 rdev->mc.mc_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL;
3760                 rdev->mc.real_vram_size = RREG32(CONFIG_MEMSIZE) * 1024ULL * 1024ULL;
3761         }
3762         rdev->mc.visible_vram_size = rdev->mc.aper_size;
3763         r700_vram_gtt_location(rdev, &rdev->mc);
3764         radeon_update_bandwidth_info(rdev);
3765
3766         return 0;
3767 }
3768
3769 void evergreen_print_gpu_status_regs(struct radeon_device *rdev)
3770 {
3771         dev_info(rdev->dev, "  GRBM_STATUS               = 0x%08X\n",
3772                 RREG32(GRBM_STATUS));
3773         dev_info(rdev->dev, "  GRBM_STATUS_SE0           = 0x%08X\n",
3774                 RREG32(GRBM_STATUS_SE0));
3775         dev_info(rdev->dev, "  GRBM_STATUS_SE1           = 0x%08X\n",
3776                 RREG32(GRBM_STATUS_SE1));
3777         dev_info(rdev->dev, "  SRBM_STATUS               = 0x%08X\n",
3778                 RREG32(SRBM_STATUS));
3779         dev_info(rdev->dev, "  SRBM_STATUS2              = 0x%08X\n",
3780                 RREG32(SRBM_STATUS2));
3781         dev_info(rdev->dev, "  R_008674_CP_STALLED_STAT1 = 0x%08X\n",
3782                 RREG32(CP_STALLED_STAT1));
3783         dev_info(rdev->dev, "  R_008678_CP_STALLED_STAT2 = 0x%08X\n",
3784                 RREG32(CP_STALLED_STAT2));
3785         dev_info(rdev->dev, "  R_00867C_CP_BUSY_STAT     = 0x%08X\n",
3786                 RREG32(CP_BUSY_STAT));
3787         dev_info(rdev->dev, "  R_008680_CP_STAT          = 0x%08X\n",
3788                 RREG32(CP_STAT));
3789         dev_info(rdev->dev, "  R_00D034_DMA_STATUS_REG   = 0x%08X\n",
3790                 RREG32(DMA_STATUS_REG));
3791         if (rdev->family >= CHIP_CAYMAN) {
3792                 dev_info(rdev->dev, "  R_00D834_DMA_STATUS_REG   = 0x%08X\n",
3793                          RREG32(DMA_STATUS_REG + 0x800));
3794         }
3795 }
3796
3797 bool evergreen_is_display_hung(struct radeon_device *rdev)
3798 {
3799         u32 crtc_hung = 0;
3800         u32 crtc_status[6];
3801         u32 i, j, tmp;
3802
3803         for (i = 0; i < rdev->num_crtc; i++) {
3804                 if (RREG32(EVERGREEN_CRTC_CONTROL + crtc_offsets[i]) & EVERGREEN_CRTC_MASTER_EN) {
3805                         crtc_status[i] = RREG32(EVERGREEN_CRTC_STATUS_HV_COUNT + crtc_offsets[i]);
3806                         crtc_hung |= (1 << i);
3807                 }
3808         }
3809
3810         for (j = 0; j < 10; j++) {
3811                 for (i = 0; i < rdev->num_crtc; i++) {
3812                         if (crtc_hung & (1 << i)) {
3813                                 tmp = RREG32(EVERGREEN_CRTC_STATUS_HV_COUNT + crtc_offsets[i]);
3814                                 if (tmp != crtc_status[i])
3815                                         crtc_hung &= ~(1 << i);
3816                         }
3817                 }
3818                 if (crtc_hung == 0)
3819                         return false;
3820                 udelay(100);
3821         }
3822
3823         return true;
3824 }
3825
3826 u32 evergreen_gpu_check_soft_reset(struct radeon_device *rdev)
3827 {
3828         u32 reset_mask = 0;
3829         u32 tmp;
3830
3831         /* GRBM_STATUS */
3832         tmp = RREG32(GRBM_STATUS);
3833         if (tmp & (PA_BUSY | SC_BUSY |
3834                    SH_BUSY | SX_BUSY |
3835                    TA_BUSY | VGT_BUSY |
3836                    DB_BUSY | CB_BUSY |
3837                    SPI_BUSY | VGT_BUSY_NO_DMA))
3838                 reset_mask |= RADEON_RESET_GFX;
3839
3840         if (tmp & (CF_RQ_PENDING | PF_RQ_PENDING |
3841                    CP_BUSY | CP_COHERENCY_BUSY))
3842                 reset_mask |= RADEON_RESET_CP;
3843
3844         if (tmp & GRBM_EE_BUSY)
3845                 reset_mask |= RADEON_RESET_GRBM | RADEON_RESET_GFX | RADEON_RESET_CP;
3846
3847         /* DMA_STATUS_REG */
3848         tmp = RREG32(DMA_STATUS_REG);
3849         if (!(tmp & DMA_IDLE))
3850                 reset_mask |= RADEON_RESET_DMA;
3851
3852         /* SRBM_STATUS2 */
3853         tmp = RREG32(SRBM_STATUS2);
3854         if (tmp & DMA_BUSY)
3855                 reset_mask |= RADEON_RESET_DMA;
3856
3857         /* SRBM_STATUS */
3858         tmp = RREG32(SRBM_STATUS);
3859         if (tmp & (RLC_RQ_PENDING | RLC_BUSY))
3860                 reset_mask |= RADEON_RESET_RLC;
3861
3862         if (tmp & IH_BUSY)
3863                 reset_mask |= RADEON_RESET_IH;
3864
3865         if (tmp & SEM_BUSY)
3866                 reset_mask |= RADEON_RESET_SEM;
3867
3868         if (tmp & GRBM_RQ_PENDING)
3869                 reset_mask |= RADEON_RESET_GRBM;
3870
3871         if (tmp & VMC_BUSY)
3872                 reset_mask |= RADEON_RESET_VMC;
3873
3874         if (tmp & (MCB_BUSY | MCB_NON_DISPLAY_BUSY |
3875                    MCC_BUSY | MCD_BUSY))
3876                 reset_mask |= RADEON_RESET_MC;
3877
3878         if (evergreen_is_display_hung(rdev))
3879                 reset_mask |= RADEON_RESET_DISPLAY;
3880
3881         /* VM_L2_STATUS */
3882         tmp = RREG32(VM_L2_STATUS);
3883         if (tmp & L2_BUSY)
3884                 reset_mask |= RADEON_RESET_VMC;
3885
3886         /* Skip MC reset as it's mostly likely not hung, just busy */
3887         if (reset_mask & RADEON_RESET_MC) {
3888                 DRM_DEBUG("MC busy: 0x%08X, clearing.\n", reset_mask);
3889                 reset_mask &= ~RADEON_RESET_MC;
3890         }
3891
3892         return reset_mask;
3893 }
3894
3895 static void evergreen_gpu_soft_reset(struct radeon_device *rdev, u32 reset_mask)
3896 {
3897         struct evergreen_mc_save save;
3898         u32 grbm_soft_reset = 0, srbm_soft_reset = 0;
3899         u32 tmp;
3900
3901         if (reset_mask == 0)
3902                 return;
3903
3904         dev_info(rdev->dev, "GPU softreset: 0x%08X\n", reset_mask);
3905
3906         evergreen_print_gpu_status_regs(rdev);
3907
3908         /* Disable CP parsing/prefetching */
3909         WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT);
3910
3911         if (reset_mask & RADEON_RESET_DMA) {
3912                 /* Disable DMA */
3913                 tmp = RREG32(DMA_RB_CNTL);
3914                 tmp &= ~DMA_RB_ENABLE;
3915                 WREG32(DMA_RB_CNTL, tmp);
3916         }
3917
3918         udelay(50);
3919
3920         evergreen_mc_stop(rdev, &save);
3921         if (evergreen_mc_wait_for_idle(rdev)) {
3922                 dev_warn(rdev->dev, "Wait for MC idle timedout !\n");
3923         }
3924
3925         if (reset_mask & (RADEON_RESET_GFX | RADEON_RESET_COMPUTE)) {
3926                 grbm_soft_reset |= SOFT_RESET_DB |
3927                         SOFT_RESET_CB |
3928                         SOFT_RESET_PA |
3929                         SOFT_RESET_SC |
3930                         SOFT_RESET_SPI |
3931                         SOFT_RESET_SX |
3932                         SOFT_RESET_SH |
3933                         SOFT_RESET_TC |
3934                         SOFT_RESET_TA |
3935                         SOFT_RESET_VC |
3936                         SOFT_RESET_VGT;
3937         }
3938
3939         if (reset_mask & RADEON_RESET_CP) {
3940                 grbm_soft_reset |= SOFT_RESET_CP |
3941                         SOFT_RESET_VGT;
3942
3943                 srbm_soft_reset |= SOFT_RESET_GRBM;
3944         }
3945
3946         if (reset_mask & RADEON_RESET_DMA)
3947                 srbm_soft_reset |= SOFT_RESET_DMA;
3948
3949         if (reset_mask & RADEON_RESET_DISPLAY)
3950                 srbm_soft_reset |= SOFT_RESET_DC;
3951
3952         if (reset_mask & RADEON_RESET_RLC)
3953                 srbm_soft_reset |= SOFT_RESET_RLC;
3954
3955         if (reset_mask & RADEON_RESET_SEM)
3956                 srbm_soft_reset |= SOFT_RESET_SEM;
3957
3958         if (reset_mask & RADEON_RESET_IH)
3959                 srbm_soft_reset |= SOFT_RESET_IH;
3960
3961         if (reset_mask & RADEON_RESET_GRBM)
3962                 srbm_soft_reset |= SOFT_RESET_GRBM;
3963
3964         if (reset_mask & RADEON_RESET_VMC)
3965                 srbm_soft_reset |= SOFT_RESET_VMC;
3966
3967         if (!(rdev->flags & RADEON_IS_IGP)) {
3968                 if (reset_mask & RADEON_RESET_MC)
3969                         srbm_soft_reset |= SOFT_RESET_MC;
3970         }
3971
3972         if (grbm_soft_reset) {
3973                 tmp = RREG32(GRBM_SOFT_RESET);
3974                 tmp |= grbm_soft_reset;
3975                 dev_info(rdev->dev, "GRBM_SOFT_RESET=0x%08X\n", tmp);
3976                 WREG32(GRBM_SOFT_RESET, tmp);
3977                 tmp = RREG32(GRBM_SOFT_RESET);
3978
3979                 udelay(50);
3980
3981                 tmp &= ~grbm_soft_reset;
3982                 WREG32(GRBM_SOFT_RESET, tmp);
3983                 tmp = RREG32(GRBM_SOFT_RESET);
3984         }
3985
3986         if (srbm_soft_reset) {
3987                 tmp = RREG32(SRBM_SOFT_RESET);
3988                 tmp |= srbm_soft_reset;
3989                 dev_info(rdev->dev, "SRBM_SOFT_RESET=0x%08X\n", tmp);
3990                 WREG32(SRBM_SOFT_RESET, tmp);
3991                 tmp = RREG32(SRBM_SOFT_RESET);
3992
3993                 udelay(50);
3994
3995                 tmp &= ~srbm_soft_reset;
3996                 WREG32(SRBM_SOFT_RESET, tmp);
3997                 tmp = RREG32(SRBM_SOFT_RESET);
3998         }
3999
4000         /* Wait a little for things to settle down */
4001         udelay(50);
4002
4003         evergreen_mc_resume(rdev, &save);
4004         udelay(50);
4005
4006         evergreen_print_gpu_status_regs(rdev);
4007 }
4008
4009 void evergreen_gpu_pci_config_reset(struct radeon_device *rdev)
4010 {
4011         struct evergreen_mc_save save;
4012         u32 tmp, i;
4013
4014         dev_info(rdev->dev, "GPU pci config reset\n");
4015
4016         /* disable dpm? */
4017
4018         /* Disable CP parsing/prefetching */
4019         WREG32(CP_ME_CNTL, CP_ME_HALT | CP_PFP_HALT);
4020         udelay(50);
4021         /* Disable DMA */
4022         tmp = RREG32(DMA_RB_CNTL);
4023         tmp &= ~DMA_RB_ENABLE;
4024         WREG32(DMA_RB_CNTL, tmp);
4025         /* XXX other engines? */
4026
4027         /* halt the rlc */
4028         r600_rlc_stop(rdev);
4029
4030         udelay(50);
4031
4032         /* set mclk/sclk to bypass */
4033         rv770_set_clk_bypass_mode(rdev);
4034         /* disable BM */
4035         pci_clear_master(rdev->pdev);
4036         /* disable mem access */
4037         evergreen_mc_stop(rdev, &save);
4038         if (evergreen_mc_wait_for_idle(rdev)) {
4039                 dev_warn(rdev->dev, "Wait for MC idle timed out !\n");
4040         }
4041         /* reset */
4042         radeon_pci_config_reset(rdev);
4043         /* wait for asic to come out of reset */
4044         for (i = 0; i < rdev->usec_timeout; i++) {
4045                 if (RREG32(CONFIG_MEMSIZE) != 0xffffffff)
4046                         break;
4047                 udelay(1);
4048         }
4049 }
4050
4051 int evergreen_asic_reset(struct radeon_device *rdev, bool hard)
4052 {
4053         u32 reset_mask;
4054
4055         if (hard) {
4056                 evergreen_gpu_pci_config_reset(rdev);
4057                 return 0;
4058         }
4059
4060         reset_mask = evergreen_gpu_check_soft_reset(rdev);
4061
4062         if (reset_mask)
4063                 r600_set_bios_scratch_engine_hung(rdev, true);
4064
4065         /* try soft reset */
4066         evergreen_gpu_soft_reset(rdev, reset_mask);
4067
4068         reset_mask = evergreen_gpu_check_soft_reset(rdev);
4069
4070         /* try pci config reset */
4071         if (reset_mask && radeon_hard_reset)
4072                 evergreen_gpu_pci_config_reset(rdev);
4073
4074         reset_mask = evergreen_gpu_check_soft_reset(rdev);
4075
4076         if (!reset_mask)
4077                 r600_set_bios_scratch_engine_hung(rdev, false);
4078
4079         return 0;
4080 }
4081
4082 /**
4083  * evergreen_gfx_is_lockup - Check if the GFX engine is locked up
4084  *
4085  * @rdev: radeon_device pointer
4086  * @ring: radeon_ring structure holding ring information
4087  *
4088  * Check if the GFX engine is locked up.
4089  * Returns true if the engine appears to be locked up, false if not.
4090  */
4091 bool evergreen_gfx_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring)
4092 {
4093         u32 reset_mask = evergreen_gpu_check_soft_reset(rdev);
4094
4095         if (!(reset_mask & (RADEON_RESET_GFX |
4096                             RADEON_RESET_COMPUTE |
4097                             RADEON_RESET_CP))) {
4098                 radeon_ring_lockup_update(rdev, ring);
4099                 return false;
4100         }
4101         return radeon_ring_test_lockup(rdev, ring);
4102 }
4103
4104 /*
4105  * RLC
4106  */
4107 #define RLC_SAVE_RESTORE_LIST_END_MARKER    0x00000000
4108 #define RLC_CLEAR_STATE_END_MARKER          0x00000001
4109
4110 void sumo_rlc_fini(struct radeon_device *rdev)
4111 {
4112         int r;
4113
4114         /* save restore block */
4115         if (rdev->rlc.save_restore_obj) {
4116                 r = radeon_bo_reserve(rdev->rlc.save_restore_obj, false);
4117                 if (unlikely(r != 0))
4118                         dev_warn(rdev->dev, "(%d) reserve RLC sr bo failed\n", r);
4119                 radeon_bo_unpin(rdev->rlc.save_restore_obj);
4120                 radeon_bo_unreserve(rdev->rlc.save_restore_obj);
4121
4122                 radeon_bo_unref(&rdev->rlc.save_restore_obj);
4123                 rdev->rlc.save_restore_obj = NULL;
4124         }
4125
4126         /* clear state block */
4127         if (rdev->rlc.clear_state_obj) {
4128                 r = radeon_bo_reserve(rdev->rlc.clear_state_obj, false);
4129                 if (unlikely(r != 0))
4130                         dev_warn(rdev->dev, "(%d) reserve RLC c bo failed\n", r);
4131                 radeon_bo_unpin(rdev->rlc.clear_state_obj);
4132                 radeon_bo_unreserve(rdev->rlc.clear_state_obj);
4133
4134                 radeon_bo_unref(&rdev->rlc.clear_state_obj);
4135                 rdev->rlc.clear_state_obj = NULL;
4136         }
4137
4138         /* clear state block */
4139         if (rdev->rlc.cp_table_obj) {
4140                 r = radeon_bo_reserve(rdev->rlc.cp_table_obj, false);
4141                 if (unlikely(r != 0))
4142                         dev_warn(rdev->dev, "(%d) reserve RLC cp table bo failed\n", r);
4143                 radeon_bo_unpin(rdev->rlc.cp_table_obj);
4144                 radeon_bo_unreserve(rdev->rlc.cp_table_obj);
4145
4146                 radeon_bo_unref(&rdev->rlc.cp_table_obj);
4147                 rdev->rlc.cp_table_obj = NULL;
4148         }
4149 }
4150
4151 #define CP_ME_TABLE_SIZE    96
4152
4153 int sumo_rlc_init(struct radeon_device *rdev)
4154 {
4155         const u32 *src_ptr;
4156         volatile u32 *dst_ptr;
4157         u32 dws, data, i, j, k, reg_num;
4158         u32 reg_list_num, reg_list_hdr_blk_index, reg_list_blk_index = 0;
4159         u64 reg_list_mc_addr;
4160         const struct cs_section_def *cs_data;
4161         int r;
4162
4163         src_ptr = rdev->rlc.reg_list;
4164         dws = rdev->rlc.reg_list_size;
4165         if (rdev->family >= CHIP_BONAIRE) {
4166                 dws += (5 * 16) + 48 + 48 + 64;
4167         }
4168         cs_data = rdev->rlc.cs_data;
4169
4170         if (src_ptr) {
4171                 /* save restore block */
4172                 if (rdev->rlc.save_restore_obj == NULL) {
4173                         r = radeon_bo_create(rdev, dws * 4, PAGE_SIZE, true,
4174                                              RADEON_GEM_DOMAIN_VRAM, 0, NULL,
4175                                              NULL, &rdev->rlc.save_restore_obj);
4176                         if (r) {
4177                                 dev_warn(rdev->dev, "(%d) create RLC sr bo failed\n", r);
4178                                 return r;
4179                         }
4180                 }
4181
4182                 r = radeon_bo_reserve(rdev->rlc.save_restore_obj, false);
4183                 if (unlikely(r != 0)) {
4184                         sumo_rlc_fini(rdev);
4185                         return r;
4186                 }
4187                 r = radeon_bo_pin(rdev->rlc.save_restore_obj, RADEON_GEM_DOMAIN_VRAM,
4188                                   &rdev->rlc.save_restore_gpu_addr);
4189                 if (r) {
4190                         radeon_bo_unreserve(rdev->rlc.save_restore_obj);
4191                         dev_warn(rdev->dev, "(%d) pin RLC sr bo failed\n", r);
4192                         sumo_rlc_fini(rdev);
4193                         return r;
4194                 }
4195
4196                 r = radeon_bo_kmap(rdev->rlc.save_restore_obj, (void **)&rdev->rlc.sr_ptr);
4197                 if (r) {
4198                         dev_warn(rdev->dev, "(%d) map RLC sr bo failed\n", r);
4199                         sumo_rlc_fini(rdev);
4200                         return r;
4201                 }
4202                 /* write the sr buffer */
4203                 dst_ptr = rdev->rlc.sr_ptr;
4204                 if (rdev->family >= CHIP_TAHITI) {
4205                         /* SI */
4206                         for (i = 0; i < rdev->rlc.reg_list_size; i++)
4207                                 dst_ptr[i] = cpu_to_le32(src_ptr[i]);
4208                 } else {
4209                         /* ON/LN/TN */
4210                         /* format:
4211                          * dw0: (reg2 << 16) | reg1
4212                          * dw1: reg1 save space
4213                          * dw2: reg2 save space
4214                          */
4215                         for (i = 0; i < dws; i++) {
4216                                 data = src_ptr[i] >> 2;
4217                                 i++;
4218                                 if (i < dws)
4219                                         data |= (src_ptr[i] >> 2) << 16;
4220                                 j = (((i - 1) * 3) / 2);
4221                                 dst_ptr[j] = cpu_to_le32(data);
4222                         }
4223                         j = ((i * 3) / 2);
4224                         dst_ptr[j] = cpu_to_le32(RLC_SAVE_RESTORE_LIST_END_MARKER);
4225                 }
4226                 radeon_bo_kunmap(rdev->rlc.save_restore_obj);
4227                 radeon_bo_unreserve(rdev->rlc.save_restore_obj);
4228         }
4229
4230         if (cs_data) {
4231                 /* clear state block */
4232                 if (rdev->family >= CHIP_BONAIRE) {
4233                         rdev->rlc.clear_state_size = dws = cik_get_csb_size(rdev);
4234                 } else if (rdev->family >= CHIP_TAHITI) {
4235                         rdev->rlc.clear_state_size = si_get_csb_size(rdev);
4236                         dws = rdev->rlc.clear_state_size + (256 / 4);
4237                 } else {
4238                         reg_list_num = 0;
4239                         dws = 0;
4240                         for (i = 0; cs_data[i].section != NULL; i++) {
4241                                 for (j = 0; cs_data[i].section[j].extent != NULL; j++) {
4242                                         reg_list_num++;
4243                                         dws += cs_data[i].section[j].reg_count;
4244                                 }
4245                         }
4246                         reg_list_blk_index = (3 * reg_list_num + 2);
4247                         dws += reg_list_blk_index;
4248                         rdev->rlc.clear_state_size = dws;
4249                 }
4250
4251                 if (rdev->rlc.clear_state_obj == NULL) {
4252                         r = radeon_bo_create(rdev, dws * 4, PAGE_SIZE, true,
4253                                              RADEON_GEM_DOMAIN_VRAM, 0, NULL,
4254                                              NULL, &rdev->rlc.clear_state_obj);
4255                         if (r) {
4256                                 dev_warn(rdev->dev, "(%d) create RLC c bo failed\n", r);
4257                                 sumo_rlc_fini(rdev);
4258                                 return r;
4259                         }
4260                 }
4261                 r = radeon_bo_reserve(rdev->rlc.clear_state_obj, false);
4262                 if (unlikely(r != 0)) {
4263                         sumo_rlc_fini(rdev);
4264                         return r;
4265                 }
4266                 r = radeon_bo_pin(rdev->rlc.clear_state_obj, RADEON_GEM_DOMAIN_VRAM,
4267                                   &rdev->rlc.clear_state_gpu_addr);
4268                 if (r) {
4269                         radeon_bo_unreserve(rdev->rlc.clear_state_obj);
4270                         dev_warn(rdev->dev, "(%d) pin RLC c bo failed\n", r);
4271                         sumo_rlc_fini(rdev);
4272                         return r;
4273                 }
4274
4275                 r = radeon_bo_kmap(rdev->rlc.clear_state_obj, (void **)&rdev->rlc.cs_ptr);
4276                 if (r) {
4277                         dev_warn(rdev->dev, "(%d) map RLC c bo failed\n", r);
4278                         sumo_rlc_fini(rdev);
4279                         return r;
4280                 }
4281                 /* set up the cs buffer */
4282                 dst_ptr = rdev->rlc.cs_ptr;
4283                 if (rdev->family >= CHIP_BONAIRE) {
4284                         cik_get_csb_buffer(rdev, dst_ptr);
4285                 } else if (rdev->family >= CHIP_TAHITI) {
4286                         reg_list_mc_addr = rdev->rlc.clear_state_gpu_addr + 256;
4287                         dst_ptr[0] = cpu_to_le32(upper_32_bits(reg_list_mc_addr));
4288                         dst_ptr[1] = cpu_to_le32(lower_32_bits(reg_list_mc_addr));
4289                         dst_ptr[2] = cpu_to_le32(rdev->rlc.clear_state_size);
4290                         si_get_csb_buffer(rdev, &dst_ptr[(256/4)]);
4291                 } else {
4292                         reg_list_hdr_blk_index = 0;
4293                         reg_list_mc_addr = rdev->rlc.clear_state_gpu_addr + (reg_list_blk_index * 4);
4294                         data = upper_32_bits(reg_list_mc_addr);
4295                         dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
4296                         reg_list_hdr_blk_index++;
4297                         for (i = 0; cs_data[i].section != NULL; i++) {
4298                                 for (j = 0; cs_data[i].section[j].extent != NULL; j++) {
4299                                         reg_num = cs_data[i].section[j].reg_count;
4300                                         data = reg_list_mc_addr & 0xffffffff;
4301                                         dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
4302                                         reg_list_hdr_blk_index++;
4303
4304                                         data = (cs_data[i].section[j].reg_index * 4) & 0xffffffff;
4305                                         dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
4306                                         reg_list_hdr_blk_index++;
4307
4308                                         data = 0x08000000 | (reg_num * 4);
4309                                         dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(data);
4310                                         reg_list_hdr_blk_index++;
4311
4312                                         for (k = 0; k < reg_num; k++) {
4313                                                 data = cs_data[i].section[j].extent[k];
4314                                                 dst_ptr[reg_list_blk_index + k] = cpu_to_le32(data);
4315                                         }
4316                                         reg_list_mc_addr += reg_num * 4;
4317                                         reg_list_blk_index += reg_num;
4318                                 }
4319                         }
4320                         dst_ptr[reg_list_hdr_blk_index] = cpu_to_le32(RLC_CLEAR_STATE_END_MARKER);
4321                 }
4322                 radeon_bo_kunmap(rdev->rlc.clear_state_obj);
4323                 radeon_bo_unreserve(rdev->rlc.clear_state_obj);
4324         }
4325
4326         if (rdev->rlc.cp_table_size) {
4327                 if (rdev->rlc.cp_table_obj == NULL) {
4328                         r = radeon_bo_create(rdev, rdev->rlc.cp_table_size,
4329                                              PAGE_SIZE, true,
4330                                              RADEON_GEM_DOMAIN_VRAM, 0, NULL,
4331                                              NULL, &rdev->rlc.cp_table_obj);
4332                         if (r) {
4333                                 dev_warn(rdev->dev, "(%d) create RLC cp table bo failed\n", r);
4334                                 sumo_rlc_fini(rdev);
4335                                 return r;
4336                         }
4337                 }
4338
4339                 r = radeon_bo_reserve(rdev->rlc.cp_table_obj, false);
4340                 if (unlikely(r != 0)) {
4341                         dev_warn(rdev->dev, "(%d) reserve RLC cp table bo failed\n", r);
4342                         sumo_rlc_fini(rdev);
4343                         return r;
4344                 }
4345                 r = radeon_bo_pin(rdev->rlc.cp_table_obj, RADEON_GEM_DOMAIN_VRAM,
4346                                   &rdev->rlc.cp_table_gpu_addr);
4347                 if (r) {
4348                         radeon_bo_unreserve(rdev->rlc.cp_table_obj);
4349                         dev_warn(rdev->dev, "(%d) pin RLC cp_table bo failed\n", r);
4350                         sumo_rlc_fini(rdev);
4351                         return r;
4352                 }
4353                 r = radeon_bo_kmap(rdev->rlc.cp_table_obj, (void **)&rdev->rlc.cp_table_ptr);
4354                 if (r) {
4355                         dev_warn(rdev->dev, "(%d) map RLC cp table bo failed\n", r);
4356                         sumo_rlc_fini(rdev);
4357                         return r;
4358                 }
4359
4360                 cik_init_cp_pg_table(rdev);
4361
4362                 radeon_bo_kunmap(rdev->rlc.cp_table_obj);
4363                 radeon_bo_unreserve(rdev->rlc.cp_table_obj);
4364
4365         }
4366
4367         return 0;
4368 }
4369
4370 static void evergreen_rlc_start(struct radeon_device *rdev)
4371 {
4372         u32 mask = RLC_ENABLE;
4373
4374         if (rdev->flags & RADEON_IS_IGP) {
4375                 mask |= GFX_POWER_GATING_ENABLE | GFX_POWER_GATING_SRC;
4376         }
4377
4378         WREG32(RLC_CNTL, mask);
4379 }
4380
4381 int evergreen_rlc_resume(struct radeon_device *rdev)
4382 {
4383         u32 i;
4384         const __be32 *fw_data;
4385
4386         if (!rdev->rlc_fw)
4387                 return -EINVAL;
4388
4389         r600_rlc_stop(rdev);
4390
4391         WREG32(RLC_HB_CNTL, 0);
4392
4393         if (rdev->flags & RADEON_IS_IGP) {
4394                 if (rdev->family == CHIP_ARUBA) {
4395                         u32 always_on_bitmap =
4396                                 3 | (3 << (16 * rdev->config.cayman.max_shader_engines));
4397                         /* find out the number of active simds */
4398                         u32 tmp = (RREG32(CC_GC_SHADER_PIPE_CONFIG) & 0xffff0000) >> 16;
4399                         tmp |= 0xffffffff << rdev->config.cayman.max_simds_per_se;
4400                         tmp = hweight32(~tmp);
4401                         if (tmp == rdev->config.cayman.max_simds_per_se) {
4402                                 WREG32(TN_RLC_LB_ALWAYS_ACTIVE_SIMD_MASK, always_on_bitmap);
4403                                 WREG32(TN_RLC_LB_PARAMS, 0x00601004);
4404                                 WREG32(TN_RLC_LB_INIT_SIMD_MASK, 0xffffffff);
4405                                 WREG32(TN_RLC_LB_CNTR_INIT, 0x00000000);
4406                                 WREG32(TN_RLC_LB_CNTR_MAX, 0x00002000);
4407                         }
4408                 } else {
4409                         WREG32(RLC_HB_WPTR_LSB_ADDR, 0);
4410                         WREG32(RLC_HB_WPTR_MSB_ADDR, 0);
4411                 }
4412                 WREG32(TN_RLC_SAVE_AND_RESTORE_BASE, rdev->rlc.save_restore_gpu_addr >> 8);
4413                 WREG32(TN_RLC_CLEAR_STATE_RESTORE_BASE, rdev->rlc.clear_state_gpu_addr >> 8);
4414         } else {
4415                 WREG32(RLC_HB_BASE, 0);
4416                 WREG32(RLC_HB_RPTR, 0);
4417                 WREG32(RLC_HB_WPTR, 0);
4418                 WREG32(RLC_HB_WPTR_LSB_ADDR, 0);
4419                 WREG32(RLC_HB_WPTR_MSB_ADDR, 0);
4420         }
4421         WREG32(RLC_MC_CNTL, 0);
4422         WREG32(RLC_UCODE_CNTL, 0);
4423
4424         fw_data = (const __be32 *)rdev->rlc_fw->data;
4425         if (rdev->family >= CHIP_ARUBA) {
4426                 for (i = 0; i < ARUBA_RLC_UCODE_SIZE; i++) {
4427                         WREG32(RLC_UCODE_ADDR, i);
4428                         WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++));
4429                 }
4430         } else if (rdev->family >= CHIP_CAYMAN) {
4431                 for (i = 0; i < CAYMAN_RLC_UCODE_SIZE; i++) {
4432                         WREG32(RLC_UCODE_ADDR, i);
4433                         WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++));
4434                 }
4435         } else {
4436                 for (i = 0; i < EVERGREEN_RLC_UCODE_SIZE; i++) {
4437                         WREG32(RLC_UCODE_ADDR, i);
4438                         WREG32(RLC_UCODE_DATA, be32_to_cpup(fw_data++));
4439                 }
4440         }
4441         WREG32(RLC_UCODE_ADDR, 0);
4442
4443         evergreen_rlc_start(rdev);
4444
4445         return 0;
4446 }
4447
4448 /* Interrupts */
4449
4450 u32 evergreen_get_vblank_counter(struct radeon_device *rdev, int crtc)
4451 {
4452         if (crtc >= rdev->num_crtc)
4453                 return 0;
4454         else
4455                 return RREG32(CRTC_STATUS_FRAME_COUNT + crtc_offsets[crtc]);
4456 }
4457
4458 void evergreen_disable_interrupt_state(struct radeon_device *rdev)
4459 {
4460         int i;
4461         u32 tmp;
4462
4463         if (rdev->family >= CHIP_CAYMAN) {
4464                 cayman_cp_int_cntl_setup(rdev, 0,
4465                                          CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE);
4466                 cayman_cp_int_cntl_setup(rdev, 1, 0);
4467                 cayman_cp_int_cntl_setup(rdev, 2, 0);
4468                 tmp = RREG32(CAYMAN_DMA1_CNTL) & ~TRAP_ENABLE;
4469                 WREG32(CAYMAN_DMA1_CNTL, tmp);
4470         } else
4471                 WREG32(CP_INT_CNTL, CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE);
4472         tmp = RREG32(DMA_CNTL) & ~TRAP_ENABLE;
4473         WREG32(DMA_CNTL, tmp);
4474         WREG32(GRBM_INT_CNTL, 0);
4475         WREG32(SRBM_INT_CNTL, 0);
4476         for (i = 0; i < rdev->num_crtc; i++)
4477                 WREG32(INT_MASK + crtc_offsets[i], 0);
4478         for (i = 0; i < rdev->num_crtc; i++)
4479                 WREG32(GRPH_INT_CONTROL + crtc_offsets[i], 0);
4480
4481         /* only one DAC on DCE5 */
4482         if (!ASIC_IS_DCE5(rdev))
4483                 WREG32(DACA_AUTODETECT_INT_CONTROL, 0);
4484         WREG32(DACB_AUTODETECT_INT_CONTROL, 0);
4485
4486         for (i = 0; i < 6; i++)
4487                 WREG32_AND(DC_HPDx_INT_CONTROL(i), DC_HPDx_INT_POLARITY);
4488 }
4489
4490 /* Note that the order we write back regs here is important */
4491 int evergreen_irq_set(struct radeon_device *rdev)
4492 {
4493         int i;
4494         u32 cp_int_cntl = CNTX_BUSY_INT_ENABLE | CNTX_EMPTY_INT_ENABLE;
4495         u32 cp_int_cntl1 = 0, cp_int_cntl2 = 0;
4496         u32 grbm_int_cntl = 0;
4497         u32 dma_cntl, dma_cntl1 = 0;
4498         u32 thermal_int = 0;
4499
4500         if (!rdev->irq.installed) {
4501                 WARN(1, "Can't enable IRQ/MSI because no handler is installed\n");
4502                 return -EINVAL;
4503         }
4504         /* don't enable anything if the ih is disabled */
4505         if (!rdev->ih.enabled) {
4506                 r600_disable_interrupts(rdev);
4507                 /* force the active interrupt state to all disabled */
4508                 evergreen_disable_interrupt_state(rdev);
4509                 return 0;
4510         }
4511
4512         if (rdev->family == CHIP_ARUBA)
4513                 thermal_int = RREG32(TN_CG_THERMAL_INT_CTRL) &
4514                         ~(THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW);
4515         else
4516                 thermal_int = RREG32(CG_THERMAL_INT) &
4517                         ~(THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW);
4518
4519         dma_cntl = RREG32(DMA_CNTL) & ~TRAP_ENABLE;
4520
4521         if (rdev->family >= CHIP_CAYMAN) {
4522                 /* enable CP interrupts on all rings */
4523                 if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
4524                         DRM_DEBUG("evergreen_irq_set: sw int gfx\n");
4525                         cp_int_cntl |= TIME_STAMP_INT_ENABLE;
4526                 }
4527                 if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP1_INDEX])) {
4528                         DRM_DEBUG("evergreen_irq_set: sw int cp1\n");
4529                         cp_int_cntl1 |= TIME_STAMP_INT_ENABLE;
4530                 }
4531                 if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_CP2_INDEX])) {
4532                         DRM_DEBUG("evergreen_irq_set: sw int cp2\n");
4533                         cp_int_cntl2 |= TIME_STAMP_INT_ENABLE;
4534                 }
4535         } else {
4536                 if (atomic_read(&rdev->irq.ring_int[RADEON_RING_TYPE_GFX_INDEX])) {
4537                         DRM_DEBUG("evergreen_irq_set: sw int gfx\n");
4538                         cp_int_cntl |= RB_INT_ENABLE;
4539                         cp_int_cntl |= TIME_STAMP_INT_ENABLE;
4540                 }
4541         }
4542
4543         if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) {
4544                 DRM_DEBUG("r600_irq_set: sw int dma\n");
4545                 dma_cntl |= TRAP_ENABLE;
4546         }
4547
4548         if (rdev->family >= CHIP_CAYMAN) {
4549                 dma_cntl1 = RREG32(CAYMAN_DMA1_CNTL) & ~TRAP_ENABLE;
4550                 if (atomic_read(&rdev->irq.ring_int[CAYMAN_RING_TYPE_DMA1_INDEX])) {
4551                         DRM_DEBUG("r600_irq_set: sw int dma1\n");
4552                         dma_cntl1 |= TRAP_ENABLE;
4553                 }
4554         }
4555
4556         if (rdev->irq.dpm_thermal) {
4557                 DRM_DEBUG("dpm thermal\n");
4558                 thermal_int |= THERM_INT_MASK_HIGH | THERM_INT_MASK_LOW;
4559         }
4560
4561         if (rdev->family >= CHIP_CAYMAN) {
4562                 cayman_cp_int_cntl_setup(rdev, 0, cp_int_cntl);
4563                 cayman_cp_int_cntl_setup(rdev, 1, cp_int_cntl1);
4564                 cayman_cp_int_cntl_setup(rdev, 2, cp_int_cntl2);
4565         } else
4566                 WREG32(CP_INT_CNTL, cp_int_cntl);
4567
4568         WREG32(DMA_CNTL, dma_cntl);
4569
4570         if (rdev->family >= CHIP_CAYMAN)
4571                 WREG32(CAYMAN_DMA1_CNTL, dma_cntl1);
4572
4573         WREG32(GRBM_INT_CNTL, grbm_int_cntl);
4574
4575         for (i = 0; i < rdev->num_crtc; i++) {
4576                 radeon_irq_kms_set_irq_n_enabled(
4577                     rdev, INT_MASK + crtc_offsets[i],
4578                     VBLANK_INT_MASK,
4579                     rdev->irq.crtc_vblank_int[i] ||
4580                     atomic_read(&rdev->irq.pflip[i]), "vblank", i);
4581         }
4582
4583         for (i = 0; i < rdev->num_crtc; i++)
4584                 WREG32(GRPH_INT_CONTROL + crtc_offsets[i], GRPH_PFLIP_INT_MASK);
4585
4586         for (i = 0; i < 6; i++) {
4587                 radeon_irq_kms_set_irq_n_enabled(
4588                     rdev, DC_HPDx_INT_CONTROL(i),
4589                     DC_HPDx_INT_EN | DC_HPDx_RX_INT_EN,
4590                     rdev->irq.hpd[i], "HPD", i);
4591         }
4592
4593         if (rdev->family == CHIP_ARUBA)
4594                 WREG32(TN_CG_THERMAL_INT_CTRL, thermal_int);
4595         else
4596                 WREG32(CG_THERMAL_INT, thermal_int);
4597
4598         for (i = 0; i < 6; i++) {
4599                 radeon_irq_kms_set_irq_n_enabled(
4600                     rdev, AFMT_AUDIO_PACKET_CONTROL + crtc_offsets[i],
4601                     AFMT_AZ_FORMAT_WTRIG_MASK,
4602                     rdev->irq.afmt[i], "HDMI", i);
4603         }
4604
4605         /* posting read */
4606         RREG32(SRBM_STATUS);
4607
4608         return 0;
4609 }
4610
4611 /* Note that the order we write back regs here is important */
4612 static void evergreen_irq_ack(struct radeon_device *rdev)
4613 {
4614         int i, j;
4615         u32 *grph_int = rdev->irq.stat_regs.evergreen.grph_int;
4616         u32 *disp_int = rdev->irq.stat_regs.evergreen.disp_int;
4617         u32 *afmt_status = rdev->irq.stat_regs.evergreen.afmt_status;
4618
4619         for (i = 0; i < 6; i++) {
4620                 disp_int[i] = RREG32(evergreen_disp_int_status[i]);
4621                 afmt_status[i] = RREG32(AFMT_STATUS + crtc_offsets[i]);
4622                 if (i < rdev->num_crtc)
4623                         grph_int[i] = RREG32(GRPH_INT_STATUS + crtc_offsets[i]);
4624         }
4625
4626         /* We write back each interrupt register in pairs of two */
4627         for (i = 0; i < rdev->num_crtc; i += 2) {
4628                 for (j = i; j < (i + 2); j++) {
4629                         if (grph_int[j] & GRPH_PFLIP_INT_OCCURRED)
4630                                 WREG32(GRPH_INT_STATUS + crtc_offsets[j],
4631                                        GRPH_PFLIP_INT_CLEAR);
4632                 }
4633
4634                 for (j = i; j < (i + 2); j++) {
4635                         if (disp_int[j] & LB_D1_VBLANK_INTERRUPT)
4636                                 WREG32(VBLANK_STATUS + crtc_offsets[j],
4637                                        VBLANK_ACK);
4638                         if (disp_int[j] & LB_D1_VLINE_INTERRUPT)
4639                                 WREG32(VLINE_STATUS + crtc_offsets[j],
4640                                        VLINE_ACK);
4641                 }
4642         }
4643
4644         for (i = 0; i < 6; i++) {
4645                 if (disp_int[i] & DC_HPD1_INTERRUPT)
4646                         WREG32_OR(DC_HPDx_INT_CONTROL(i), DC_HPDx_INT_ACK);
4647         }
4648
4649         for (i = 0; i < 6; i++) {
4650                 if (disp_int[i] & DC_HPD1_RX_INTERRUPT)
4651                         WREG32_OR(DC_HPDx_INT_CONTROL(i), DC_HPDx_RX_INT_ACK);
4652         }
4653
4654         for (i = 0; i < 6; i++) {
4655                 if (afmt_status[i] & AFMT_AZ_FORMAT_WTRIG)
4656                         WREG32_OR(AFMT_AUDIO_PACKET_CONTROL + crtc_offsets[i],
4657                                   AFMT_AZ_FORMAT_WTRIG_ACK);
4658         }
4659 }
4660
4661 static void evergreen_irq_disable(struct radeon_device *rdev)
4662 {
4663         r600_disable_interrupts(rdev);
4664         /* Wait and acknowledge irq */
4665         mdelay(1);
4666         evergreen_irq_ack(rdev);
4667         evergreen_disable_interrupt_state(rdev);
4668 }
4669
4670 void evergreen_irq_suspend(struct radeon_device *rdev)
4671 {
4672         evergreen_irq_disable(rdev);
4673         r600_rlc_stop(rdev);
4674 }
4675
4676 static u32 evergreen_get_ih_wptr(struct radeon_device *rdev)
4677 {
4678         u32 wptr, tmp;
4679
4680         if (rdev->wb.enabled)
4681                 wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]);
4682         else
4683                 wptr = RREG32(IH_RB_WPTR);
4684
4685         if (wptr & RB_OVERFLOW) {
4686                 wptr &= ~RB_OVERFLOW;
4687                 /* When a ring buffer overflow happen start parsing interrupt
4688                  * from the last not overwritten vector (wptr + 16). Hopefully
4689                  * this should allow us to catchup.
4690                  */
4691                 dev_warn(rdev->dev, "IH ring buffer overflow (0x%08X, 0x%08X, 0x%08X)\n",
4692                          wptr, rdev->ih.rptr, (wptr + 16) & rdev->ih.ptr_mask);
4693                 rdev->ih.rptr = (wptr + 16) & rdev->ih.ptr_mask;
4694                 tmp = RREG32(IH_RB_CNTL);
4695                 tmp |= IH_WPTR_OVERFLOW_CLEAR;
4696                 WREG32(IH_RB_CNTL, tmp);
4697         }
4698         return (wptr & rdev->ih.ptr_mask);
4699 }
4700
4701 int evergreen_irq_process(struct radeon_device *rdev)
4702 {
4703         u32 *disp_int = rdev->irq.stat_regs.evergreen.disp_int;
4704         u32 *afmt_status = rdev->irq.stat_regs.evergreen.afmt_status;
4705         u32 crtc_idx, hpd_idx, afmt_idx;
4706         u32 mask;
4707         u32 wptr;
4708         u32 rptr;
4709         u32 src_id, src_data;
4710         u32 ring_index;
4711         bool queue_hotplug = false;
4712         bool queue_hdmi = false;
4713         bool queue_dp = false;
4714         bool queue_thermal = false;
4715         u32 status, addr;
4716         const char *event_name;
4717
4718         if (!rdev->ih.enabled || rdev->shutdown)
4719                 return IRQ_NONE;
4720
4721         wptr = evergreen_get_ih_wptr(rdev);
4722
4723 restart_ih:
4724         /* is somebody else already processing irqs? */
4725         if (atomic_xchg(&rdev->ih.lock, 1))
4726                 return IRQ_NONE;
4727
4728         rptr = rdev->ih.rptr;
4729         DRM_DEBUG("evergreen_irq_process start: rptr %d, wptr %d\n", rptr, wptr);
4730
4731         /* Order reading of wptr vs. reading of IH ring data */
4732         rmb();
4733
4734         /* display interrupts */
4735         evergreen_irq_ack(rdev);
4736
4737         while (rptr != wptr) {
4738                 /* wptr/rptr are in bytes! */
4739                 ring_index = rptr / 4;
4740                 src_id =  le32_to_cpu(rdev->ih.ring[ring_index]) & 0xff;
4741                 src_data = le32_to_cpu(rdev->ih.ring[ring_index + 1]) & 0xfffffff;
4742
4743                 switch (src_id) {
4744                 case 1: /* D1 vblank/vline */
4745                 case 2: /* D2 vblank/vline */
4746                 case 3: /* D3 vblank/vline */
4747                 case 4: /* D4 vblank/vline */
4748                 case 5: /* D5 vblank/vline */
4749                 case 6: /* D6 vblank/vline */
4750                         crtc_idx = src_id - 1;
4751
4752                         if (src_data == 0) { /* vblank */
4753                                 mask = LB_D1_VBLANK_INTERRUPT;
4754                                 event_name = "vblank";
4755
4756                                 if (rdev->irq.crtc_vblank_int[crtc_idx]) {
4757                                         drm_handle_vblank(rdev->ddev, crtc_idx);
4758                                         rdev->pm.vblank_sync = true;
4759                                         wake_up(&rdev->irq.vblank_queue);
4760                                 }
4761                                 if (atomic_read(&rdev->irq.pflip[crtc_idx])) {
4762                                         radeon_crtc_handle_vblank(rdev,
4763                                                                   crtc_idx);
4764                                 }
4765
4766                         } else if (src_data == 1) { /* vline */
4767                                 mask = LB_D1_VLINE_INTERRUPT;
4768                                 event_name = "vline";
4769                         } else {
4770                                 DRM_DEBUG("Unhandled interrupt: %d %d\n",
4771                                           src_id, src_data);
4772                                 break;
4773                         }
4774
4775                         if (!(disp_int[crtc_idx] & mask)) {
4776                                 DRM_DEBUG("IH: D%d %s - IH event w/o asserted irq bit?\n",
4777                                           crtc_idx + 1, event_name);
4778                         }
4779
4780                         disp_int[crtc_idx] &= ~mask;
4781                         DRM_DEBUG("IH: D%d %s\n", crtc_idx + 1, event_name);
4782
4783                         break;
4784                 case 8: /* D1 page flip */
4785                 case 10: /* D2 page flip */
4786                 case 12: /* D3 page flip */
4787                 case 14: /* D4 page flip */
4788                 case 16: /* D5 page flip */
4789                 case 18: /* D6 page flip */
4790                         DRM_DEBUG("IH: D%d flip\n", ((src_id - 8) >> 1) + 1);
4791                         if (radeon_use_pflipirq > 0)
4792                                 radeon_crtc_handle_flip(rdev, (src_id - 8) >> 1);
4793                         break;
4794                 case 42: /* HPD hotplug */
4795                         if (src_data <= 5) {
4796                                 hpd_idx = src_data;
4797                                 mask = DC_HPD1_INTERRUPT;
4798                                 queue_hotplug = true;
4799                                 event_name = "HPD";
4800
4801                         } else if (src_data <= 11) {
4802                                 hpd_idx = src_data - 6;
4803                                 mask = DC_HPD1_RX_INTERRUPT;
4804                                 queue_dp = true;
4805                                 event_name = "HPD_RX";
4806
4807                         } else {
4808                                 DRM_DEBUG("Unhandled interrupt: %d %d\n",
4809                                           src_id, src_data);
4810                                 break;
4811                         }
4812
4813                         if (!(disp_int[hpd_idx] & mask))
4814                                 DRM_DEBUG("IH: IH event w/o asserted irq bit?\n");
4815
4816                         disp_int[hpd_idx] &= ~mask;
4817                         DRM_DEBUG("IH: %s%d\n", event_name, hpd_idx + 1);
4818
4819                         break;
4820                 case 44: /* hdmi */
4821                         afmt_idx = src_data;
4822                         if (!(afmt_status[afmt_idx] & AFMT_AZ_FORMAT_WTRIG))
4823                                 DRM_DEBUG("IH: IH event w/o asserted irq bit?\n");
4824
4825                         if (afmt_idx > 5) {
4826                                 DRM_ERROR("Unhandled interrupt: %d %d\n",
4827                                           src_id, src_data);
4828                                 break;
4829                         }
4830                         afmt_status[afmt_idx] &= ~AFMT_AZ_FORMAT_WTRIG;
4831                         queue_hdmi = true;
4832                         DRM_DEBUG("IH: HDMI%d\n", afmt_idx + 1);
4833                         break;
4834                 case 96:
4835                         DRM_ERROR("SRBM_READ_ERROR: 0x%x\n", RREG32(SRBM_READ_ERROR));
4836                         WREG32(SRBM_INT_ACK, 0x1);
4837                         break;
4838                 case 124: /* UVD */
4839                         DRM_DEBUG("IH: UVD int: 0x%08x\n", src_data);
4840                         radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX);
4841                         break;
4842                 case 146:
4843                 case 147:
4844                         addr = RREG32(VM_CONTEXT1_PROTECTION_FAULT_ADDR);
4845                         status = RREG32(VM_CONTEXT1_PROTECTION_FAULT_STATUS);
4846                         /* reset addr and status */
4847                         WREG32_P(VM_CONTEXT1_CNTL2, 1, ~1);
4848                         if (addr == 0x0 && status == 0x0)
4849                                 break;
4850                         dev_err(rdev->dev, "GPU fault detected: %d 0x%08x\n", src_id, src_data);
4851                         dev_err(rdev->dev, "  VM_CONTEXT1_PROTECTION_FAULT_ADDR   0x%08X\n",
4852                                 addr);
4853                         dev_err(rdev->dev, "  VM_CONTEXT1_PROTECTION_FAULT_STATUS 0x%08X\n",
4854                                 status);
4855                         cayman_vm_decode_fault(rdev, status, addr);
4856                         break;
4857                 case 176: /* CP_INT in ring buffer */
4858                 case 177: /* CP_INT in IB1 */
4859                 case 178: /* CP_INT in IB2 */
4860                         DRM_DEBUG("IH: CP int: 0x%08x\n", src_data);
4861                         radeon_fence_process(rdev, RADEON_RING_TYPE_GFX_INDEX);
4862                         break;
4863                 case 181: /* CP EOP event */
4864                         DRM_DEBUG("IH: CP EOP\n");
4865                         if (rdev->family >= CHIP_CAYMAN) {
4866                                 switch (src_data) {
4867                                 case 0:
4868                                         radeon_fence_process(rdev, RADEON_RING_TYPE_GFX_INDEX);
4869                                         break;
4870                                 case 1:
4871                                         radeon_fence_process(rdev, CAYMAN_RING_TYPE_CP1_INDEX);
4872                                         break;
4873                                 case 2:
4874                                         radeon_fence_process(rdev, CAYMAN_RING_TYPE_CP2_INDEX);
4875                                         break;
4876                                 }
4877                         } else
4878                                 radeon_fence_process(rdev, RADEON_RING_TYPE_GFX_INDEX);
4879                         break;
4880                 case 224: /* DMA trap event */
4881                         DRM_DEBUG("IH: DMA trap\n");
4882                         radeon_fence_process(rdev, R600_RING_TYPE_DMA_INDEX);
4883                         break;
4884                 case 230: /* thermal low to high */
4885                         DRM_DEBUG("IH: thermal low to high\n");
4886                         rdev->pm.dpm.thermal.high_to_low = false;
4887                         queue_thermal = true;
4888                         break;
4889                 case 231: /* thermal high to low */
4890                         DRM_DEBUG("IH: thermal high to low\n");
4891                         rdev->pm.dpm.thermal.high_to_low = true;
4892                         queue_thermal = true;
4893                         break;
4894                 case 233: /* GUI IDLE */
4895                         DRM_DEBUG("IH: GUI idle\n");
4896                         break;
4897                 case 244: /* DMA trap event */
4898                         if (rdev->family >= CHIP_CAYMAN) {
4899                                 DRM_DEBUG("IH: DMA1 trap\n");
4900                                 radeon_fence_process(rdev, CAYMAN_RING_TYPE_DMA1_INDEX);
4901                         }
4902                         break;
4903                 default:
4904                         DRM_DEBUG("Unhandled interrupt: %d %d\n", src_id, src_data);
4905                         break;
4906                 }
4907
4908                 /* wptr/rptr are in bytes! */
4909                 rptr += 16;
4910                 rptr &= rdev->ih.ptr_mask;
4911                 WREG32(IH_RB_RPTR, rptr);
4912         }
4913         if (queue_dp)
4914                 schedule_work(&rdev->dp_work);
4915         if (queue_hotplug)
4916                 schedule_delayed_work(&rdev->hotplug_work, 0);
4917         if (queue_hdmi)
4918                 schedule_work(&rdev->audio_work);
4919         if (queue_thermal && rdev->pm.dpm_enabled)
4920                 schedule_work(&rdev->pm.dpm.thermal.work);
4921         rdev->ih.rptr = rptr;
4922         atomic_set(&rdev->ih.lock, 0);
4923
4924         /* make sure wptr hasn't changed while processing */
4925         wptr = evergreen_get_ih_wptr(rdev);
4926         if (wptr != rptr)
4927                 goto restart_ih;
4928
4929         return IRQ_HANDLED;
4930 }
4931
4932 static void evergreen_uvd_init(struct radeon_device *rdev)
4933 {
4934         int r;
4935
4936         if (!rdev->has_uvd)
4937                 return;
4938
4939         r = radeon_uvd_init(rdev);
4940         if (r) {
4941                 dev_err(rdev->dev, "failed UVD (%d) init.\n", r);
4942                 /*
4943                  * At this point rdev->uvd.vcpu_bo is NULL which trickles down
4944                  * to early fails uvd_v2_2_resume() and thus nothing happens
4945                  * there. So it is pointless to try to go through that code
4946                  * hence why we disable uvd here.
4947                  */
4948                 rdev->has_uvd = false;
4949                 return;
4950         }
4951         rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL;
4952         r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096);
4953 }
4954
4955 static void evergreen_uvd_start(struct radeon_device *rdev)
4956 {
4957         int r;
4958
4959         if (!rdev->has_uvd)
4960                 return;
4961
4962         r = uvd_v2_2_resume(rdev);
4963         if (r) {
4964                 dev_err(rdev->dev, "failed UVD resume (%d).\n", r);
4965                 goto error;
4966         }
4967         r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX);
4968         if (r) {
4969                 dev_err(rdev->dev, "failed initializing UVD fences (%d).\n", r);
4970                 goto error;
4971         }
4972         return;
4973
4974 error:
4975         rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0;
4976 }
4977
4978 static void evergreen_uvd_resume(struct radeon_device *rdev)
4979 {
4980         struct radeon_ring *ring;
4981         int r;
4982
4983         if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size)
4984                 return;
4985
4986         ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
4987         r = radeon_ring_init(rdev, ring, ring->ring_size, 0, PACKET0(UVD_NO_OP, 0));
4988         if (r) {
4989                 dev_err(rdev->dev, "failed initializing UVD ring (%d).\n", r);
4990                 return;
4991         }
4992         r = uvd_v1_0_init(rdev);
4993         if (r) {
4994                 dev_err(rdev->dev, "failed initializing UVD (%d).\n", r);
4995                 return;
4996         }
4997 }
4998
4999 static int evergreen_startup(struct radeon_device *rdev)
5000 {
5001         struct radeon_ring *ring;
5002         int r;
5003
5004         /* enable pcie gen2 link */
5005         evergreen_pcie_gen2_enable(rdev);
5006         /* enable aspm */
5007         evergreen_program_aspm(rdev);
5008
5009         /* scratch needs to be initialized before MC */
5010         r = r600_vram_scratch_init(rdev);
5011         if (r)
5012                 return r;
5013
5014         evergreen_mc_program(rdev);
5015
5016         if (ASIC_IS_DCE5(rdev) && !rdev->pm.dpm_enabled) {
5017                 r = ni_mc_load_microcode(rdev);
5018                 if (r) {
5019                         DRM_ERROR("Failed to load MC firmware!\n");
5020                         return r;
5021                 }
5022         }
5023
5024         if (rdev->flags & RADEON_IS_AGP) {
5025                 evergreen_agp_enable(rdev);
5026         } else {
5027                 r = evergreen_pcie_gart_enable(rdev);
5028                 if (r)
5029                         return r;
5030         }
5031         evergreen_gpu_init(rdev);
5032
5033         /* allocate rlc buffers */
5034         if (rdev->flags & RADEON_IS_IGP) {
5035                 rdev->rlc.reg_list = sumo_rlc_save_restore_register_list;
5036                 rdev->rlc.reg_list_size =
5037                         (u32)ARRAY_SIZE(sumo_rlc_save_restore_register_list);
5038                 rdev->rlc.cs_data = evergreen_cs_data;
5039                 r = sumo_rlc_init(rdev);
5040                 if (r) {
5041                         DRM_ERROR("Failed to init rlc BOs!\n");
5042                         return r;
5043                 }
5044         }
5045
5046         /* allocate wb buffer */
5047         r = radeon_wb_init(rdev);
5048         if (r)
5049                 return r;
5050
5051         r = radeon_fence_driver_start_ring(rdev, RADEON_RING_TYPE_GFX_INDEX);
5052         if (r) {
5053                 dev_err(rdev->dev, "failed initializing CP fences (%d).\n", r);
5054                 return r;
5055         }
5056
5057         r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX);
5058         if (r) {
5059                 dev_err(rdev->dev, "failed initializing DMA fences (%d).\n", r);
5060                 return r;
5061         }
5062
5063         evergreen_uvd_start(rdev);
5064
5065         /* Enable IRQ */
5066         if (!rdev->irq.installed) {
5067                 r = radeon_irq_kms_init(rdev);
5068                 if (r)
5069                         return r;
5070         }
5071
5072         r = r600_irq_init(rdev);
5073         if (r) {
5074                 DRM_ERROR("radeon: IH init failed (%d).\n", r);
5075                 radeon_irq_kms_fini(rdev);
5076                 return r;
5077         }
5078         evergreen_irq_set(rdev);
5079
5080         ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX];
5081         r = radeon_ring_init(rdev, ring, ring->ring_size, RADEON_WB_CP_RPTR_OFFSET,
5082                              RADEON_CP_PACKET2);
5083         if (r)
5084                 return r;
5085
5086         ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX];
5087         r = radeon_ring_init(rdev, ring, ring->ring_size, R600_WB_DMA_RPTR_OFFSET,
5088                              DMA_PACKET(DMA_PACKET_NOP, 0, 0));
5089         if (r)
5090                 return r;
5091
5092         r = evergreen_cp_load_microcode(rdev);
5093         if (r)
5094                 return r;
5095         r = evergreen_cp_resume(rdev);
5096         if (r)
5097                 return r;
5098         r = r600_dma_resume(rdev);
5099         if (r)
5100                 return r;
5101
5102         evergreen_uvd_resume(rdev);
5103
5104         r = radeon_ib_pool_init(rdev);
5105         if (r) {
5106                 dev_err(rdev->dev, "IB initialization failed (%d).\n", r);
5107                 return r;
5108         }
5109
5110         r = radeon_audio_init(rdev);
5111         if (r) {
5112                 DRM_ERROR("radeon: audio init failed\n");
5113                 return r;
5114         }
5115
5116         return 0;
5117 }
5118
5119 int evergreen_resume(struct radeon_device *rdev)
5120 {
5121         int r;
5122
5123         /* reset the asic, the gfx blocks are often in a bad state
5124          * after the driver is unloaded or after a resume
5125          */
5126         if (radeon_asic_reset(rdev))
5127                 dev_warn(rdev->dev, "GPU reset failed !\n");
5128         /* Do not reset GPU before posting, on rv770 hw unlike on r500 hw,
5129          * posting will perform necessary task to bring back GPU into good
5130          * shape.
5131          */
5132         /* post card */
5133         atom_asic_init(rdev->mode_info.atom_context);
5134
5135         /* init golden registers */
5136         evergreen_init_golden_registers(rdev);
5137
5138         if (rdev->pm.pm_method == PM_METHOD_DPM)
5139                 radeon_pm_resume(rdev);
5140
5141         rdev->accel_working = true;
5142         r = evergreen_startup(rdev);
5143         if (r) {
5144                 DRM_ERROR("evergreen startup failed on resume\n");
5145                 rdev->accel_working = false;
5146                 return r;
5147         }
5148
5149         return r;
5150
5151 }
5152
5153 int evergreen_suspend(struct radeon_device *rdev)
5154 {
5155         radeon_pm_suspend(rdev);
5156         radeon_audio_fini(rdev);
5157         if (rdev->has_uvd) {
5158                 uvd_v1_0_fini(rdev);
5159                 radeon_uvd_suspend(rdev);
5160         }
5161         r700_cp_stop(rdev);
5162         r600_dma_stop(rdev);
5163         evergreen_irq_suspend(rdev);
5164         radeon_wb_disable(rdev);
5165         evergreen_pcie_gart_disable(rdev);
5166
5167         return 0;
5168 }
5169
5170 /* Plan is to move initialization in that function and use
5171  * helper function so that radeon_device_init pretty much
5172  * do nothing more than calling asic specific function. This
5173  * should also allow to remove a bunch of callback function
5174  * like vram_info.
5175  */
5176 int evergreen_init(struct radeon_device *rdev)
5177 {
5178         int r;
5179
5180         /* Read BIOS */
5181         if (!radeon_get_bios(rdev)) {
5182                 if (ASIC_IS_AVIVO(rdev))
5183                         return -EINVAL;
5184         }
5185         /* Must be an ATOMBIOS */
5186         if (!rdev->is_atom_bios) {
5187                 dev_err(rdev->dev, "Expecting atombios for evergreen GPU\n");
5188                 return -EINVAL;
5189         }
5190         r = radeon_atombios_init(rdev);
5191         if (r)
5192                 return r;
5193         /* reset the asic, the gfx blocks are often in a bad state
5194          * after the driver is unloaded or after a resume
5195          */
5196         if (radeon_asic_reset(rdev))
5197                 dev_warn(rdev->dev, "GPU reset failed !\n");
5198         /* Post card if necessary */
5199         if (!radeon_card_posted(rdev)) {
5200                 if (!rdev->bios) {
5201                         dev_err(rdev->dev, "Card not posted and no BIOS - ignoring\n");
5202                         return -EINVAL;
5203                 }
5204                 DRM_INFO("GPU not posted. posting now...\n");
5205                 atom_asic_init(rdev->mode_info.atom_context);
5206         }
5207         /* init golden registers */
5208         evergreen_init_golden_registers(rdev);
5209         /* Initialize scratch registers */
5210         r600_scratch_init(rdev);
5211         /* Initialize surface registers */
5212         radeon_surface_init(rdev);
5213         /* Initialize clocks */
5214         radeon_get_clock_info(rdev->ddev);
5215         /* Fence driver */
5216         r = radeon_fence_driver_init(rdev);
5217         if (r)
5218                 return r;
5219         /* initialize AGP */
5220         if (rdev->flags & RADEON_IS_AGP) {
5221                 r = radeon_agp_init(rdev);
5222                 if (r)
5223                         radeon_agp_disable(rdev);
5224         }
5225         /* initialize memory controller */
5226         r = evergreen_mc_init(rdev);
5227         if (r)
5228                 return r;
5229         /* Memory manager */
5230         r = radeon_bo_init(rdev);
5231         if (r)
5232                 return r;
5233
5234         if (ASIC_IS_DCE5(rdev)) {
5235                 if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw || !rdev->mc_fw) {
5236                         r = ni_init_microcode(rdev);
5237                         if (r) {
5238                                 DRM_ERROR("Failed to load firmware!\n");
5239                                 return r;
5240                         }
5241                 }
5242         } else {
5243                 if (!rdev->me_fw || !rdev->pfp_fw || !rdev->rlc_fw) {
5244                         r = r600_init_microcode(rdev);
5245                         if (r) {
5246                                 DRM_ERROR("Failed to load firmware!\n");
5247                                 /*(DEBLOBBED)*/
5248                         }
5249                 }
5250         }
5251
5252         /* Initialize power management */
5253         radeon_pm_init(rdev);
5254
5255         rdev->ring[RADEON_RING_TYPE_GFX_INDEX].ring_obj = NULL;
5256         r600_ring_init(rdev, &rdev->ring[RADEON_RING_TYPE_GFX_INDEX], 1024 * 1024);
5257
5258         rdev->ring[R600_RING_TYPE_DMA_INDEX].ring_obj = NULL;
5259         r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX], 64 * 1024);
5260
5261         evergreen_uvd_init(rdev);
5262
5263         rdev->ih.ring_obj = NULL;
5264         r600_ih_ring_init(rdev, 64 * 1024);
5265
5266         r = r600_pcie_gart_init(rdev);
5267         if (r)
5268                 return r;
5269
5270         rdev->accel_working = true;
5271         r = evergreen_startup(rdev);
5272         if (r) {
5273                 dev_err(rdev->dev, "disabling GPU acceleration\n");
5274                 r700_cp_fini(rdev);
5275                 r600_dma_fini(rdev);
5276                 r600_irq_fini(rdev);
5277                 if (rdev->flags & RADEON_IS_IGP)
5278                         sumo_rlc_fini(rdev);
5279                 radeon_wb_fini(rdev);
5280                 radeon_ib_pool_fini(rdev);
5281                 radeon_irq_kms_fini(rdev);
5282                 evergreen_pcie_gart_fini(rdev);
5283                 rdev->accel_working = false;
5284         }
5285
5286         /* Don't start up if the MC ucode is missing on BTC parts.
5287          * The default clocks and voltages before the MC ucode
5288          * is loaded are not suffient for advanced operations.
5289          */
5290         if (ASIC_IS_DCE5(rdev)) {
5291                 if (!rdev->mc_fw && !(rdev->flags & RADEON_IS_IGP)) {
5292                         DRM_ERROR("radeon: MC ucode required for NI+.\n");
5293                         return -EINVAL;
5294                 }
5295         }
5296
5297         return 0;
5298 }
5299
5300 void evergreen_fini(struct radeon_device *rdev)
5301 {
5302         radeon_pm_fini(rdev);
5303         radeon_audio_fini(rdev);
5304         r700_cp_fini(rdev);
5305         r600_dma_fini(rdev);
5306         r600_irq_fini(rdev);
5307         if (rdev->flags & RADEON_IS_IGP)
5308                 sumo_rlc_fini(rdev);
5309         radeon_wb_fini(rdev);
5310         radeon_ib_pool_fini(rdev);
5311         radeon_irq_kms_fini(rdev);
5312         uvd_v1_0_fini(rdev);
5313         radeon_uvd_fini(rdev);
5314         evergreen_pcie_gart_fini(rdev);
5315         r600_vram_scratch_fini(rdev);
5316         radeon_gem_fini(rdev);
5317         radeon_fence_driver_fini(rdev);
5318         radeon_agp_fini(rdev);
5319         radeon_bo_fini(rdev);
5320         radeon_atombios_fini(rdev);
5321         kfree(rdev->bios);
5322         rdev->bios = NULL;
5323 }
5324
5325 void evergreen_pcie_gen2_enable(struct radeon_device *rdev)
5326 {
5327         u32 link_width_cntl, speed_cntl;
5328
5329         if (radeon_pcie_gen2 == 0)
5330                 return;
5331
5332         if (rdev->flags & RADEON_IS_IGP)
5333                 return;
5334
5335         if (!(rdev->flags & RADEON_IS_PCIE))
5336                 return;
5337
5338         /* x2 cards have a special sequence */
5339         if (ASIC_IS_X2(rdev))
5340                 return;
5341
5342         if ((rdev->pdev->bus->max_bus_speed != PCIE_SPEED_5_0GT) &&
5343                 (rdev->pdev->bus->max_bus_speed != PCIE_SPEED_8_0GT))
5344                 return;
5345
5346         speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
5347         if (speed_cntl & LC_CURRENT_DATA_RATE) {
5348                 DRM_INFO("PCIE gen 2 link speeds already enabled\n");
5349                 return;
5350         }
5351
5352         DRM_INFO("enabling PCIE gen 2 link speeds, disable with radeon.pcie_gen2=0\n");
5353
5354         if ((speed_cntl & LC_OTHER_SIDE_EVER_SENT_GEN2) ||
5355             (speed_cntl & LC_OTHER_SIDE_SUPPORTS_GEN2)) {
5356
5357                 link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
5358                 link_width_cntl &= ~LC_UPCONFIGURE_DIS;
5359                 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
5360
5361                 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
5362                 speed_cntl &= ~LC_TARGET_LINK_SPEED_OVERRIDE_EN;
5363                 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
5364
5365                 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
5366                 speed_cntl |= LC_CLR_FAILED_SPD_CHANGE_CNT;
5367                 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
5368
5369                 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
5370                 speed_cntl &= ~LC_CLR_FAILED_SPD_CHANGE_CNT;
5371                 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
5372
5373                 speed_cntl = RREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL);
5374                 speed_cntl |= LC_GEN2_EN_STRAP;
5375                 WREG32_PCIE_PORT(PCIE_LC_SPEED_CNTL, speed_cntl);
5376
5377         } else {
5378                 link_width_cntl = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
5379                 /* XXX: only disable it if gen1 bridge vendor == 0x111d or 0x1106 */
5380                 if (1)
5381                         link_width_cntl |= LC_UPCONFIGURE_DIS;
5382                 else
5383                         link_width_cntl &= ~LC_UPCONFIGURE_DIS;
5384                 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, link_width_cntl);
5385         }
5386 }
5387
5388 void evergreen_program_aspm(struct radeon_device *rdev)
5389 {
5390         u32 data, orig;
5391         u32 pcie_lc_cntl, pcie_lc_cntl_old;
5392         bool disable_l0s, disable_l1 = false, disable_plloff_in_l1 = false;
5393         /* fusion_platform = true
5394          * if the system is a fusion system
5395          * (APU or DGPU in a fusion system).
5396          * todo: check if the system is a fusion platform.
5397          */
5398         bool fusion_platform = false;
5399
5400         if (radeon_aspm == 0)
5401                 return;
5402
5403         if (!(rdev->flags & RADEON_IS_PCIE))
5404                 return;
5405
5406         switch (rdev->family) {
5407         case CHIP_CYPRESS:
5408         case CHIP_HEMLOCK:
5409         case CHIP_JUNIPER:
5410         case CHIP_REDWOOD:
5411         case CHIP_CEDAR:
5412         case CHIP_SUMO:
5413         case CHIP_SUMO2:
5414         case CHIP_PALM:
5415         case CHIP_ARUBA:
5416                 disable_l0s = true;
5417                 break;
5418         default:
5419                 disable_l0s = false;
5420                 break;
5421         }
5422
5423         if (rdev->flags & RADEON_IS_IGP)
5424                 fusion_platform = true; /* XXX also dGPUs in a fusion system */
5425
5426         data = orig = RREG32_PIF_PHY0(PB0_PIF_PAIRING);
5427         if (fusion_platform)
5428                 data &= ~MULTI_PIF;
5429         else
5430                 data |= MULTI_PIF;
5431         if (data != orig)
5432                 WREG32_PIF_PHY0(PB0_PIF_PAIRING, data);
5433
5434         data = orig = RREG32_PIF_PHY1(PB1_PIF_PAIRING);
5435         if (fusion_platform)
5436                 data &= ~MULTI_PIF;
5437         else
5438                 data |= MULTI_PIF;
5439         if (data != orig)
5440                 WREG32_PIF_PHY1(PB1_PIF_PAIRING, data);
5441
5442         pcie_lc_cntl = pcie_lc_cntl_old = RREG32_PCIE_PORT(PCIE_LC_CNTL);
5443         pcie_lc_cntl &= ~(LC_L0S_INACTIVITY_MASK | LC_L1_INACTIVITY_MASK);
5444         if (!disable_l0s) {
5445                 if (rdev->family >= CHIP_BARTS)
5446                         pcie_lc_cntl |= LC_L0S_INACTIVITY(7);
5447                 else
5448                         pcie_lc_cntl |= LC_L0S_INACTIVITY(3);
5449         }
5450
5451         if (!disable_l1) {
5452                 if (rdev->family >= CHIP_BARTS)
5453                         pcie_lc_cntl |= LC_L1_INACTIVITY(7);
5454                 else
5455                         pcie_lc_cntl |= LC_L1_INACTIVITY(8);
5456
5457                 if (!disable_plloff_in_l1) {
5458                         data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0);
5459                         data &= ~(PLL_POWER_STATE_IN_OFF_0_MASK | PLL_POWER_STATE_IN_TXS2_0_MASK);
5460                         data |= PLL_POWER_STATE_IN_OFF_0(7) | PLL_POWER_STATE_IN_TXS2_0(7);
5461                         if (data != orig)
5462                                 WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0, data);
5463
5464                         data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1);
5465                         data &= ~(PLL_POWER_STATE_IN_OFF_1_MASK | PLL_POWER_STATE_IN_TXS2_1_MASK);
5466                         data |= PLL_POWER_STATE_IN_OFF_1(7) | PLL_POWER_STATE_IN_TXS2_1(7);
5467                         if (data != orig)
5468                                 WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1, data);
5469
5470                         data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0);
5471                         data &= ~(PLL_POWER_STATE_IN_OFF_0_MASK | PLL_POWER_STATE_IN_TXS2_0_MASK);
5472                         data |= PLL_POWER_STATE_IN_OFF_0(7) | PLL_POWER_STATE_IN_TXS2_0(7);
5473                         if (data != orig)
5474                                 WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0, data);
5475
5476                         data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1);
5477                         data &= ~(PLL_POWER_STATE_IN_OFF_1_MASK | PLL_POWER_STATE_IN_TXS2_1_MASK);
5478                         data |= PLL_POWER_STATE_IN_OFF_1(7) | PLL_POWER_STATE_IN_TXS2_1(7);
5479                         if (data != orig)
5480                                 WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1, data);
5481
5482                         if (rdev->family >= CHIP_BARTS) {
5483                                 data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0);
5484                                 data &= ~PLL_RAMP_UP_TIME_0_MASK;
5485                                 data |= PLL_RAMP_UP_TIME_0(4);
5486                                 if (data != orig)
5487                                         WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_0, data);
5488
5489                                 data = orig = RREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1);
5490                                 data &= ~PLL_RAMP_UP_TIME_1_MASK;
5491                                 data |= PLL_RAMP_UP_TIME_1(4);
5492                                 if (data != orig)
5493                                         WREG32_PIF_PHY0(PB0_PIF_PWRDOWN_1, data);
5494
5495                                 data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0);
5496                                 data &= ~PLL_RAMP_UP_TIME_0_MASK;
5497                                 data |= PLL_RAMP_UP_TIME_0(4);
5498                                 if (data != orig)
5499                                         WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_0, data);
5500
5501                                 data = orig = RREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1);
5502                                 data &= ~PLL_RAMP_UP_TIME_1_MASK;
5503                                 data |= PLL_RAMP_UP_TIME_1(4);
5504                                 if (data != orig)
5505                                         WREG32_PIF_PHY1(PB1_PIF_PWRDOWN_1, data);
5506                         }
5507
5508                         data = orig = RREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL);
5509                         data &= ~LC_DYN_LANES_PWR_STATE_MASK;
5510                         data |= LC_DYN_LANES_PWR_STATE(3);
5511                         if (data != orig)
5512                                 WREG32_PCIE_PORT(PCIE_LC_LINK_WIDTH_CNTL, data);
5513
5514                         if (rdev->family >= CHIP_BARTS) {
5515                                 data = orig = RREG32_PIF_PHY0(PB0_PIF_CNTL);
5516                                 data &= ~LS2_EXIT_TIME_MASK;
5517                                 data |= LS2_EXIT_TIME(1);
5518                                 if (data != orig)
5519                                         WREG32_PIF_PHY0(PB0_PIF_CNTL, data);
5520
5521                                 data = orig = RREG32_PIF_PHY1(PB1_PIF_CNTL);
5522                                 data &= ~LS2_EXIT_TIME_MASK;
5523                                 data |= LS2_EXIT_TIME(1);
5524                                 if (data != orig)
5525                                         WREG32_PIF_PHY1(PB1_PIF_CNTL, data);
5526                         }
5527                 }
5528         }
5529
5530         /* evergreen parts only */
5531         if (rdev->family < CHIP_BARTS)
5532                 pcie_lc_cntl |= LC_PMI_TO_L1_DIS;
5533
5534         if (pcie_lc_cntl != pcie_lc_cntl_old)
5535                 WREG32_PCIE_PORT(PCIE_LC_CNTL, pcie_lc_cntl);
5536 }