Merge git://git.kernel.org/pub/scm/linux/kernel/git/pablo/nf
[linux-2.6-microblaze.git] / drivers / gpu / drm / i915 / intel_runtime_pm.c
1 /*
2  * Copyright © 2012-2014 Intel Corporation
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice (including the next
12  * paragraph) shall be included in all copies or substantial portions of the
13  * Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21  * IN THE SOFTWARE.
22  *
23  * Authors:
24  *    Eugeni Dodonov <eugeni.dodonov@intel.com>
25  *    Daniel Vetter <daniel.vetter@ffwll.ch>
26  *
27  */
28
29 #include <linux/pm_runtime.h>
30 #include <linux/vgaarb.h>
31
32 #include <drm/drm_print.h>
33
34 #include "i915_drv.h"
35 #include "intel_drv.h"
36
37 /**
38  * DOC: runtime pm
39  *
40  * The i915 driver supports dynamic enabling and disabling of entire hardware
41  * blocks at runtime. This is especially important on the display side where
42  * software is supposed to control many power gates manually on recent hardware,
43  * since on the GT side a lot of the power management is done by the hardware.
44  * But even there some manual control at the device level is required.
45  *
46  * Since i915 supports a diverse set of platforms with a unified codebase and
47  * hardware engineers just love to shuffle functionality around between power
48  * domains there's a sizeable amount of indirection required. This file provides
49  * generic functions to the driver for grabbing and releasing references for
50  * abstract power domains. It then maps those to the actual power wells
51  * present for a given platform.
52  */
53
54 #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
55
56 #include <linux/sort.h>
57
58 #define STACKDEPTH 8
59
60 static noinline depot_stack_handle_t __save_depot_stack(void)
61 {
62         unsigned long entries[STACKDEPTH];
63         struct stack_trace trace = {
64                 .entries = entries,
65                 .max_entries = ARRAY_SIZE(entries),
66                 .skip = 1,
67         };
68
69         save_stack_trace(&trace);
70         if (trace.nr_entries &&
71             trace.entries[trace.nr_entries - 1] == ULONG_MAX)
72                 trace.nr_entries--;
73
74         return depot_save_stack(&trace, GFP_NOWAIT | __GFP_NOWARN);
75 }
76
77 static void __print_depot_stack(depot_stack_handle_t stack,
78                                 char *buf, int sz, int indent)
79 {
80         unsigned long entries[STACKDEPTH];
81         struct stack_trace trace = {
82                 .entries = entries,
83                 .max_entries = ARRAY_SIZE(entries),
84         };
85
86         depot_fetch_stack(stack, &trace);
87         snprint_stack_trace(buf, sz, &trace, indent);
88 }
89
90 static void init_intel_runtime_pm_wakeref(struct drm_i915_private *i915)
91 {
92         struct i915_runtime_pm *rpm = &i915->runtime_pm;
93
94         spin_lock_init(&rpm->debug.lock);
95 }
96
97 static noinline depot_stack_handle_t
98 track_intel_runtime_pm_wakeref(struct drm_i915_private *i915)
99 {
100         struct i915_runtime_pm *rpm = &i915->runtime_pm;
101         depot_stack_handle_t stack, *stacks;
102         unsigned long flags;
103
104         atomic_inc(&rpm->wakeref_count);
105         assert_rpm_wakelock_held(i915);
106
107         if (!HAS_RUNTIME_PM(i915))
108                 return -1;
109
110         stack = __save_depot_stack();
111         if (!stack)
112                 return -1;
113
114         spin_lock_irqsave(&rpm->debug.lock, flags);
115
116         if (!rpm->debug.count)
117                 rpm->debug.last_acquire = stack;
118
119         stacks = krealloc(rpm->debug.owners,
120                           (rpm->debug.count + 1) * sizeof(*stacks),
121                           GFP_NOWAIT | __GFP_NOWARN);
122         if (stacks) {
123                 stacks[rpm->debug.count++] = stack;
124                 rpm->debug.owners = stacks;
125         } else {
126                 stack = -1;
127         }
128
129         spin_unlock_irqrestore(&rpm->debug.lock, flags);
130
131         return stack;
132 }
133
134 static void cancel_intel_runtime_pm_wakeref(struct drm_i915_private *i915,
135                                             depot_stack_handle_t stack)
136 {
137         struct i915_runtime_pm *rpm = &i915->runtime_pm;
138         unsigned long flags, n;
139         bool found = false;
140
141         if (unlikely(stack == -1))
142                 return;
143
144         spin_lock_irqsave(&rpm->debug.lock, flags);
145         for (n = rpm->debug.count; n--; ) {
146                 if (rpm->debug.owners[n] == stack) {
147                         memmove(rpm->debug.owners + n,
148                                 rpm->debug.owners + n + 1,
149                                 (--rpm->debug.count - n) * sizeof(stack));
150                         found = true;
151                         break;
152                 }
153         }
154         spin_unlock_irqrestore(&rpm->debug.lock, flags);
155
156         if (WARN(!found,
157                  "Unmatched wakeref (tracking %lu), count %u\n",
158                  rpm->debug.count, atomic_read(&rpm->wakeref_count))) {
159                 char *buf;
160
161                 buf = kmalloc(PAGE_SIZE, GFP_KERNEL);
162                 if (!buf)
163                         return;
164
165                 __print_depot_stack(stack, buf, PAGE_SIZE, 2);
166                 DRM_DEBUG_DRIVER("wakeref %x from\n%s", stack, buf);
167
168                 stack = READ_ONCE(rpm->debug.last_release);
169                 if (stack) {
170                         __print_depot_stack(stack, buf, PAGE_SIZE, 2);
171                         DRM_DEBUG_DRIVER("wakeref last released at\n%s", buf);
172                 }
173
174                 kfree(buf);
175         }
176 }
177
178 static int cmphandle(const void *_a, const void *_b)
179 {
180         const depot_stack_handle_t * const a = _a, * const b = _b;
181
182         if (*a < *b)
183                 return -1;
184         else if (*a > *b)
185                 return 1;
186         else
187                 return 0;
188 }
189
190 static void
191 __print_intel_runtime_pm_wakeref(struct drm_printer *p,
192                                  const struct intel_runtime_pm_debug *dbg)
193 {
194         unsigned long i;
195         char *buf;
196
197         buf = kmalloc(PAGE_SIZE, GFP_KERNEL);
198         if (!buf)
199                 return;
200
201         if (dbg->last_acquire) {
202                 __print_depot_stack(dbg->last_acquire, buf, PAGE_SIZE, 2);
203                 drm_printf(p, "Wakeref last acquired:\n%s", buf);
204         }
205
206         if (dbg->last_release) {
207                 __print_depot_stack(dbg->last_release, buf, PAGE_SIZE, 2);
208                 drm_printf(p, "Wakeref last released:\n%s", buf);
209         }
210
211         drm_printf(p, "Wakeref count: %lu\n", dbg->count);
212
213         sort(dbg->owners, dbg->count, sizeof(*dbg->owners), cmphandle, NULL);
214
215         for (i = 0; i < dbg->count; i++) {
216                 depot_stack_handle_t stack = dbg->owners[i];
217                 unsigned long rep;
218
219                 rep = 1;
220                 while (i + 1 < dbg->count && dbg->owners[i + 1] == stack)
221                         rep++, i++;
222                 __print_depot_stack(stack, buf, PAGE_SIZE, 2);
223                 drm_printf(p, "Wakeref x%lu taken at:\n%s", rep, buf);
224         }
225
226         kfree(buf);
227 }
228
229 static noinline void
230 untrack_intel_runtime_pm_wakeref(struct drm_i915_private *i915)
231 {
232         struct i915_runtime_pm *rpm = &i915->runtime_pm;
233         struct intel_runtime_pm_debug dbg = {};
234         struct drm_printer p;
235         unsigned long flags;
236
237         assert_rpm_wakelock_held(i915);
238         if (atomic_dec_and_lock_irqsave(&rpm->wakeref_count,
239                                         &rpm->debug.lock,
240                                         flags)) {
241                 dbg = rpm->debug;
242
243                 rpm->debug.owners = NULL;
244                 rpm->debug.count = 0;
245                 rpm->debug.last_release = __save_depot_stack();
246
247                 spin_unlock_irqrestore(&rpm->debug.lock, flags);
248         }
249         if (!dbg.count)
250                 return;
251
252         p = drm_debug_printer("i915");
253         __print_intel_runtime_pm_wakeref(&p, &dbg);
254
255         kfree(dbg.owners);
256 }
257
258 void print_intel_runtime_pm_wakeref(struct drm_i915_private *i915,
259                                     struct drm_printer *p)
260 {
261         struct intel_runtime_pm_debug dbg = {};
262
263         do {
264                 struct i915_runtime_pm *rpm = &i915->runtime_pm;
265                 unsigned long alloc = dbg.count;
266                 depot_stack_handle_t *s;
267
268                 spin_lock_irq(&rpm->debug.lock);
269                 dbg.count = rpm->debug.count;
270                 if (dbg.count <= alloc) {
271                         memcpy(dbg.owners,
272                                rpm->debug.owners,
273                                dbg.count * sizeof(*s));
274                 }
275                 dbg.last_acquire = rpm->debug.last_acquire;
276                 dbg.last_release = rpm->debug.last_release;
277                 spin_unlock_irq(&rpm->debug.lock);
278                 if (dbg.count <= alloc)
279                         break;
280
281                 s = krealloc(dbg.owners, dbg.count * sizeof(*s), GFP_KERNEL);
282                 if (!s)
283                         goto out;
284
285                 dbg.owners = s;
286         } while (1);
287
288         __print_intel_runtime_pm_wakeref(p, &dbg);
289
290 out:
291         kfree(dbg.owners);
292 }
293
294 #else
295
296 static void init_intel_runtime_pm_wakeref(struct drm_i915_private *i915)
297 {
298 }
299
300 static depot_stack_handle_t
301 track_intel_runtime_pm_wakeref(struct drm_i915_private *i915)
302 {
303         atomic_inc(&i915->runtime_pm.wakeref_count);
304         assert_rpm_wakelock_held(i915);
305         return -1;
306 }
307
308 static void untrack_intel_runtime_pm_wakeref(struct drm_i915_private *i915)
309 {
310         assert_rpm_wakelock_held(i915);
311         atomic_dec(&i915->runtime_pm.wakeref_count);
312 }
313
314 #endif
315
316 bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv,
317                                          enum i915_power_well_id power_well_id);
318
319 const char *
320 intel_display_power_domain_str(enum intel_display_power_domain domain)
321 {
322         switch (domain) {
323         case POWER_DOMAIN_PIPE_A:
324                 return "PIPE_A";
325         case POWER_DOMAIN_PIPE_B:
326                 return "PIPE_B";
327         case POWER_DOMAIN_PIPE_C:
328                 return "PIPE_C";
329         case POWER_DOMAIN_PIPE_A_PANEL_FITTER:
330                 return "PIPE_A_PANEL_FITTER";
331         case POWER_DOMAIN_PIPE_B_PANEL_FITTER:
332                 return "PIPE_B_PANEL_FITTER";
333         case POWER_DOMAIN_PIPE_C_PANEL_FITTER:
334                 return "PIPE_C_PANEL_FITTER";
335         case POWER_DOMAIN_TRANSCODER_A:
336                 return "TRANSCODER_A";
337         case POWER_DOMAIN_TRANSCODER_B:
338                 return "TRANSCODER_B";
339         case POWER_DOMAIN_TRANSCODER_C:
340                 return "TRANSCODER_C";
341         case POWER_DOMAIN_TRANSCODER_EDP:
342                 return "TRANSCODER_EDP";
343         case POWER_DOMAIN_TRANSCODER_EDP_VDSC:
344                 return "TRANSCODER_EDP_VDSC";
345         case POWER_DOMAIN_TRANSCODER_DSI_A:
346                 return "TRANSCODER_DSI_A";
347         case POWER_DOMAIN_TRANSCODER_DSI_C:
348                 return "TRANSCODER_DSI_C";
349         case POWER_DOMAIN_PORT_DDI_A_LANES:
350                 return "PORT_DDI_A_LANES";
351         case POWER_DOMAIN_PORT_DDI_B_LANES:
352                 return "PORT_DDI_B_LANES";
353         case POWER_DOMAIN_PORT_DDI_C_LANES:
354                 return "PORT_DDI_C_LANES";
355         case POWER_DOMAIN_PORT_DDI_D_LANES:
356                 return "PORT_DDI_D_LANES";
357         case POWER_DOMAIN_PORT_DDI_E_LANES:
358                 return "PORT_DDI_E_LANES";
359         case POWER_DOMAIN_PORT_DDI_F_LANES:
360                 return "PORT_DDI_F_LANES";
361         case POWER_DOMAIN_PORT_DDI_A_IO:
362                 return "PORT_DDI_A_IO";
363         case POWER_DOMAIN_PORT_DDI_B_IO:
364                 return "PORT_DDI_B_IO";
365         case POWER_DOMAIN_PORT_DDI_C_IO:
366                 return "PORT_DDI_C_IO";
367         case POWER_DOMAIN_PORT_DDI_D_IO:
368                 return "PORT_DDI_D_IO";
369         case POWER_DOMAIN_PORT_DDI_E_IO:
370                 return "PORT_DDI_E_IO";
371         case POWER_DOMAIN_PORT_DDI_F_IO:
372                 return "PORT_DDI_F_IO";
373         case POWER_DOMAIN_PORT_DSI:
374                 return "PORT_DSI";
375         case POWER_DOMAIN_PORT_CRT:
376                 return "PORT_CRT";
377         case POWER_DOMAIN_PORT_OTHER:
378                 return "PORT_OTHER";
379         case POWER_DOMAIN_VGA:
380                 return "VGA";
381         case POWER_DOMAIN_AUDIO:
382                 return "AUDIO";
383         case POWER_DOMAIN_PLLS:
384                 return "PLLS";
385         case POWER_DOMAIN_AUX_A:
386                 return "AUX_A";
387         case POWER_DOMAIN_AUX_B:
388                 return "AUX_B";
389         case POWER_DOMAIN_AUX_C:
390                 return "AUX_C";
391         case POWER_DOMAIN_AUX_D:
392                 return "AUX_D";
393         case POWER_DOMAIN_AUX_E:
394                 return "AUX_E";
395         case POWER_DOMAIN_AUX_F:
396                 return "AUX_F";
397         case POWER_DOMAIN_AUX_IO_A:
398                 return "AUX_IO_A";
399         case POWER_DOMAIN_AUX_TBT1:
400                 return "AUX_TBT1";
401         case POWER_DOMAIN_AUX_TBT2:
402                 return "AUX_TBT2";
403         case POWER_DOMAIN_AUX_TBT3:
404                 return "AUX_TBT3";
405         case POWER_DOMAIN_AUX_TBT4:
406                 return "AUX_TBT4";
407         case POWER_DOMAIN_GMBUS:
408                 return "GMBUS";
409         case POWER_DOMAIN_INIT:
410                 return "INIT";
411         case POWER_DOMAIN_MODESET:
412                 return "MODESET";
413         case POWER_DOMAIN_GT_IRQ:
414                 return "GT_IRQ";
415         default:
416                 MISSING_CASE(domain);
417                 return "?";
418         }
419 }
420
421 static void intel_power_well_enable(struct drm_i915_private *dev_priv,
422                                     struct i915_power_well *power_well)
423 {
424         DRM_DEBUG_KMS("enabling %s\n", power_well->desc->name);
425         power_well->desc->ops->enable(dev_priv, power_well);
426         power_well->hw_enabled = true;
427 }
428
429 static void intel_power_well_disable(struct drm_i915_private *dev_priv,
430                                      struct i915_power_well *power_well)
431 {
432         DRM_DEBUG_KMS("disabling %s\n", power_well->desc->name);
433         power_well->hw_enabled = false;
434         power_well->desc->ops->disable(dev_priv, power_well);
435 }
436
437 static void intel_power_well_get(struct drm_i915_private *dev_priv,
438                                  struct i915_power_well *power_well)
439 {
440         if (!power_well->count++)
441                 intel_power_well_enable(dev_priv, power_well);
442 }
443
444 static void intel_power_well_put(struct drm_i915_private *dev_priv,
445                                  struct i915_power_well *power_well)
446 {
447         WARN(!power_well->count, "Use count on power well %s is already zero",
448              power_well->desc->name);
449
450         if (!--power_well->count)
451                 intel_power_well_disable(dev_priv, power_well);
452 }
453
454 /**
455  * __intel_display_power_is_enabled - unlocked check for a power domain
456  * @dev_priv: i915 device instance
457  * @domain: power domain to check
458  *
459  * This is the unlocked version of intel_display_power_is_enabled() and should
460  * only be used from error capture and recovery code where deadlocks are
461  * possible.
462  *
463  * Returns:
464  * True when the power domain is enabled, false otherwise.
465  */
466 bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
467                                       enum intel_display_power_domain domain)
468 {
469         struct i915_power_well *power_well;
470         bool is_enabled;
471
472         if (dev_priv->runtime_pm.suspended)
473                 return false;
474
475         is_enabled = true;
476
477         for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain)) {
478                 if (power_well->desc->always_on)
479                         continue;
480
481                 if (!power_well->hw_enabled) {
482                         is_enabled = false;
483                         break;
484                 }
485         }
486
487         return is_enabled;
488 }
489
490 /**
491  * intel_display_power_is_enabled - check for a power domain
492  * @dev_priv: i915 device instance
493  * @domain: power domain to check
494  *
495  * This function can be used to check the hw power domain state. It is mostly
496  * used in hardware state readout functions. Everywhere else code should rely
497  * upon explicit power domain reference counting to ensure that the hardware
498  * block is powered up before accessing it.
499  *
500  * Callers must hold the relevant modesetting locks to ensure that concurrent
501  * threads can't disable the power well while the caller tries to read a few
502  * registers.
503  *
504  * Returns:
505  * True when the power domain is enabled, false otherwise.
506  */
507 bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv,
508                                     enum intel_display_power_domain domain)
509 {
510         struct i915_power_domains *power_domains;
511         bool ret;
512
513         power_domains = &dev_priv->power_domains;
514
515         mutex_lock(&power_domains->lock);
516         ret = __intel_display_power_is_enabled(dev_priv, domain);
517         mutex_unlock(&power_domains->lock);
518
519         return ret;
520 }
521
522 /*
523  * Starting with Haswell, we have a "Power Down Well" that can be turned off
524  * when not needed anymore. We have 4 registers that can request the power well
525  * to be enabled, and it will only be disabled if none of the registers is
526  * requesting it to be enabled.
527  */
528 static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv,
529                                        u8 irq_pipe_mask, bool has_vga)
530 {
531         struct pci_dev *pdev = dev_priv->drm.pdev;
532
533         /*
534          * After we re-enable the power well, if we touch VGA register 0x3d5
535          * we'll get unclaimed register interrupts. This stops after we write
536          * anything to the VGA MSR register. The vgacon module uses this
537          * register all the time, so if we unbind our driver and, as a
538          * consequence, bind vgacon, we'll get stuck in an infinite loop at
539          * console_unlock(). So make here we touch the VGA MSR register, making
540          * sure vgacon can keep working normally without triggering interrupts
541          * and error messages.
542          */
543         if (has_vga) {
544                 vga_get_uninterruptible(pdev, VGA_RSRC_LEGACY_IO);
545                 outb(inb(VGA_MSR_READ), VGA_MSR_WRITE);
546                 vga_put(pdev, VGA_RSRC_LEGACY_IO);
547         }
548
549         if (irq_pipe_mask)
550                 gen8_irq_power_well_post_enable(dev_priv, irq_pipe_mask);
551 }
552
553 static void hsw_power_well_pre_disable(struct drm_i915_private *dev_priv,
554                                        u8 irq_pipe_mask)
555 {
556         if (irq_pipe_mask)
557                 gen8_irq_power_well_pre_disable(dev_priv, irq_pipe_mask);
558 }
559
560
561 static void hsw_wait_for_power_well_enable(struct drm_i915_private *dev_priv,
562                                            struct i915_power_well *power_well)
563 {
564         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
565         int pw_idx = power_well->desc->hsw.idx;
566
567         /* Timeout for PW1:10 us, AUX:not specified, other PWs:20 us. */
568         WARN_ON(intel_wait_for_register(dev_priv,
569                                         regs->driver,
570                                         HSW_PWR_WELL_CTL_STATE(pw_idx),
571                                         HSW_PWR_WELL_CTL_STATE(pw_idx),
572                                         1));
573 }
574
575 static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv,
576                                      const struct i915_power_well_regs *regs,
577                                      int pw_idx)
578 {
579         u32 req_mask = HSW_PWR_WELL_CTL_REQ(pw_idx);
580         u32 ret;
581
582         ret = I915_READ(regs->bios) & req_mask ? 1 : 0;
583         ret |= I915_READ(regs->driver) & req_mask ? 2 : 0;
584         if (regs->kvmr.reg)
585                 ret |= I915_READ(regs->kvmr) & req_mask ? 4 : 0;
586         ret |= I915_READ(regs->debug) & req_mask ? 8 : 0;
587
588         return ret;
589 }
590
591 static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv,
592                                             struct i915_power_well *power_well)
593 {
594         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
595         int pw_idx = power_well->desc->hsw.idx;
596         bool disabled;
597         u32 reqs;
598
599         /*
600          * Bspec doesn't require waiting for PWs to get disabled, but still do
601          * this for paranoia. The known cases where a PW will be forced on:
602          * - a KVMR request on any power well via the KVMR request register
603          * - a DMC request on PW1 and MISC_IO power wells via the BIOS and
604          *   DEBUG request registers
605          * Skip the wait in case any of the request bits are set and print a
606          * diagnostic message.
607          */
608         wait_for((disabled = !(I915_READ(regs->driver) &
609                                HSW_PWR_WELL_CTL_STATE(pw_idx))) ||
610                  (reqs = hsw_power_well_requesters(dev_priv, regs, pw_idx)), 1);
611         if (disabled)
612                 return;
613
614         DRM_DEBUG_KMS("%s forced on (bios:%d driver:%d kvmr:%d debug:%d)\n",
615                       power_well->desc->name,
616                       !!(reqs & 1), !!(reqs & 2), !!(reqs & 4), !!(reqs & 8));
617 }
618
619 static void gen9_wait_for_power_well_fuses(struct drm_i915_private *dev_priv,
620                                            enum skl_power_gate pg)
621 {
622         /* Timeout 5us for PG#0, for other PGs 1us */
623         WARN_ON(intel_wait_for_register(dev_priv, SKL_FUSE_STATUS,
624                                         SKL_FUSE_PG_DIST_STATUS(pg),
625                                         SKL_FUSE_PG_DIST_STATUS(pg), 1));
626 }
627
628 static void hsw_power_well_enable(struct drm_i915_private *dev_priv,
629                                   struct i915_power_well *power_well)
630 {
631         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
632         int pw_idx = power_well->desc->hsw.idx;
633         bool wait_fuses = power_well->desc->hsw.has_fuses;
634         enum skl_power_gate uninitialized_var(pg);
635         u32 val;
636
637         if (wait_fuses) {
638                 pg = INTEL_GEN(dev_priv) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) :
639                                                  SKL_PW_CTL_IDX_TO_PG(pw_idx);
640                 /*
641                  * For PW1 we have to wait both for the PW0/PG0 fuse state
642                  * before enabling the power well and PW1/PG1's own fuse
643                  * state after the enabling. For all other power wells with
644                  * fuses we only have to wait for that PW/PG's fuse state
645                  * after the enabling.
646                  */
647                 if (pg == SKL_PG1)
648                         gen9_wait_for_power_well_fuses(dev_priv, SKL_PG0);
649         }
650
651         val = I915_READ(regs->driver);
652         I915_WRITE(regs->driver, val | HSW_PWR_WELL_CTL_REQ(pw_idx));
653         hsw_wait_for_power_well_enable(dev_priv, power_well);
654
655         /* Display WA #1178: cnl */
656         if (IS_CANNONLAKE(dev_priv) &&
657             pw_idx >= GLK_PW_CTL_IDX_AUX_B &&
658             pw_idx <= CNL_PW_CTL_IDX_AUX_F) {
659                 val = I915_READ(CNL_AUX_ANAOVRD1(pw_idx));
660                 val |= CNL_AUX_ANAOVRD1_ENABLE | CNL_AUX_ANAOVRD1_LDO_BYPASS;
661                 I915_WRITE(CNL_AUX_ANAOVRD1(pw_idx), val);
662         }
663
664         if (wait_fuses)
665                 gen9_wait_for_power_well_fuses(dev_priv, pg);
666
667         hsw_power_well_post_enable(dev_priv,
668                                    power_well->desc->hsw.irq_pipe_mask,
669                                    power_well->desc->hsw.has_vga);
670 }
671
672 static void hsw_power_well_disable(struct drm_i915_private *dev_priv,
673                                    struct i915_power_well *power_well)
674 {
675         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
676         int pw_idx = power_well->desc->hsw.idx;
677         u32 val;
678
679         hsw_power_well_pre_disable(dev_priv,
680                                    power_well->desc->hsw.irq_pipe_mask);
681
682         val = I915_READ(regs->driver);
683         I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx));
684         hsw_wait_for_power_well_disable(dev_priv, power_well);
685 }
686
687 #define ICL_AUX_PW_TO_PORT(pw_idx)      ((pw_idx) - ICL_PW_CTL_IDX_AUX_A)
688
689 static void
690 icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
691                                     struct i915_power_well *power_well)
692 {
693         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
694         int pw_idx = power_well->desc->hsw.idx;
695         enum port port = ICL_AUX_PW_TO_PORT(pw_idx);
696         u32 val;
697
698         val = I915_READ(regs->driver);
699         I915_WRITE(regs->driver, val | HSW_PWR_WELL_CTL_REQ(pw_idx));
700
701         val = I915_READ(ICL_PORT_CL_DW12(port));
702         I915_WRITE(ICL_PORT_CL_DW12(port), val | ICL_LANE_ENABLE_AUX);
703
704         hsw_wait_for_power_well_enable(dev_priv, power_well);
705
706         /* Display WA #1178: icl */
707         if (IS_ICELAKE(dev_priv) &&
708             pw_idx >= ICL_PW_CTL_IDX_AUX_A && pw_idx <= ICL_PW_CTL_IDX_AUX_B &&
709             !intel_bios_is_port_edp(dev_priv, port)) {
710                 val = I915_READ(ICL_AUX_ANAOVRD1(pw_idx));
711                 val |= ICL_AUX_ANAOVRD1_ENABLE | ICL_AUX_ANAOVRD1_LDO_BYPASS;
712                 I915_WRITE(ICL_AUX_ANAOVRD1(pw_idx), val);
713         }
714 }
715
716 static void
717 icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv,
718                                      struct i915_power_well *power_well)
719 {
720         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
721         int pw_idx = power_well->desc->hsw.idx;
722         enum port port = ICL_AUX_PW_TO_PORT(pw_idx);
723         u32 val;
724
725         val = I915_READ(ICL_PORT_CL_DW12(port));
726         I915_WRITE(ICL_PORT_CL_DW12(port), val & ~ICL_LANE_ENABLE_AUX);
727
728         val = I915_READ(regs->driver);
729         I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx));
730
731         hsw_wait_for_power_well_disable(dev_priv, power_well);
732 }
733
734 #define ICL_AUX_PW_TO_CH(pw_idx)        \
735         ((pw_idx) - ICL_PW_CTL_IDX_AUX_A + AUX_CH_A)
736
737 static void
738 icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv,
739                                  struct i915_power_well *power_well)
740 {
741         enum aux_ch aux_ch = ICL_AUX_PW_TO_CH(power_well->desc->hsw.idx);
742         u32 val;
743
744         val = I915_READ(DP_AUX_CH_CTL(aux_ch));
745         val &= ~DP_AUX_CH_CTL_TBT_IO;
746         if (power_well->desc->hsw.is_tc_tbt)
747                 val |= DP_AUX_CH_CTL_TBT_IO;
748         I915_WRITE(DP_AUX_CH_CTL(aux_ch), val);
749
750         hsw_power_well_enable(dev_priv, power_well);
751 }
752
753 /*
754  * We should only use the power well if we explicitly asked the hardware to
755  * enable it, so check if it's enabled and also check if we've requested it to
756  * be enabled.
757  */
758 static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv,
759                                    struct i915_power_well *power_well)
760 {
761         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
762         enum i915_power_well_id id = power_well->desc->id;
763         int pw_idx = power_well->desc->hsw.idx;
764         u32 mask = HSW_PWR_WELL_CTL_REQ(pw_idx) |
765                    HSW_PWR_WELL_CTL_STATE(pw_idx);
766         u32 val;
767
768         val = I915_READ(regs->driver);
769
770         /*
771          * On GEN9 big core due to a DMC bug the driver's request bits for PW1
772          * and the MISC_IO PW will be not restored, so check instead for the
773          * BIOS's own request bits, which are forced-on for these power wells
774          * when exiting DC5/6.
775          */
776         if (IS_GEN(dev_priv, 9) && !IS_GEN9_LP(dev_priv) &&
777             (id == SKL_DISP_PW_1 || id == SKL_DISP_PW_MISC_IO))
778                 val |= I915_READ(regs->bios);
779
780         return (val & mask) == mask;
781 }
782
783 static void assert_can_enable_dc9(struct drm_i915_private *dev_priv)
784 {
785         WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_DC9),
786                   "DC9 already programmed to be enabled.\n");
787         WARN_ONCE(I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5,
788                   "DC5 still not disabled to enable DC9.\n");
789         WARN_ONCE(I915_READ(HSW_PWR_WELL_CTL2) &
790                   HSW_PWR_WELL_CTL_REQ(SKL_PW_CTL_IDX_PW_2),
791                   "Power well 2 on.\n");
792         WARN_ONCE(intel_irqs_enabled(dev_priv),
793                   "Interrupts not disabled yet.\n");
794
795          /*
796           * TODO: check for the following to verify the conditions to enter DC9
797           * state are satisfied:
798           * 1] Check relevant display engine registers to verify if mode set
799           * disable sequence was followed.
800           * 2] Check if display uninitialize sequence is initialized.
801           */
802 }
803
804 static void assert_can_disable_dc9(struct drm_i915_private *dev_priv)
805 {
806         WARN_ONCE(intel_irqs_enabled(dev_priv),
807                   "Interrupts not disabled yet.\n");
808         WARN_ONCE(I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5,
809                   "DC5 still not disabled.\n");
810
811          /*
812           * TODO: check for the following to verify DC9 state was indeed
813           * entered before programming to disable it:
814           * 1] Check relevant display engine registers to verify if mode
815           *  set disable sequence was followed.
816           * 2] Check if display uninitialize sequence is initialized.
817           */
818 }
819
820 static void gen9_write_dc_state(struct drm_i915_private *dev_priv,
821                                 u32 state)
822 {
823         int rewrites = 0;
824         int rereads = 0;
825         u32 v;
826
827         I915_WRITE(DC_STATE_EN, state);
828
829         /* It has been observed that disabling the dc6 state sometimes
830          * doesn't stick and dmc keeps returning old value. Make sure
831          * the write really sticks enough times and also force rewrite until
832          * we are confident that state is exactly what we want.
833          */
834         do  {
835                 v = I915_READ(DC_STATE_EN);
836
837                 if (v != state) {
838                         I915_WRITE(DC_STATE_EN, state);
839                         rewrites++;
840                         rereads = 0;
841                 } else if (rereads++ > 5) {
842                         break;
843                 }
844
845         } while (rewrites < 100);
846
847         if (v != state)
848                 DRM_ERROR("Writing dc state to 0x%x failed, now 0x%x\n",
849                           state, v);
850
851         /* Most of the times we need one retry, avoid spam */
852         if (rewrites > 1)
853                 DRM_DEBUG_KMS("Rewrote dc state to 0x%x %d times\n",
854                               state, rewrites);
855 }
856
857 static u32 gen9_dc_mask(struct drm_i915_private *dev_priv)
858 {
859         u32 mask;
860
861         mask = DC_STATE_EN_UPTO_DC5;
862         if (INTEL_GEN(dev_priv) >= 11)
863                 mask |= DC_STATE_EN_UPTO_DC6 | DC_STATE_EN_DC9;
864         else if (IS_GEN9_LP(dev_priv))
865                 mask |= DC_STATE_EN_DC9;
866         else
867                 mask |= DC_STATE_EN_UPTO_DC6;
868
869         return mask;
870 }
871
872 void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv)
873 {
874         u32 val;
875
876         val = I915_READ(DC_STATE_EN) & gen9_dc_mask(dev_priv);
877
878         DRM_DEBUG_KMS("Resetting DC state tracking from %02x to %02x\n",
879                       dev_priv->csr.dc_state, val);
880         dev_priv->csr.dc_state = val;
881 }
882
883 /**
884  * gen9_set_dc_state - set target display C power state
885  * @dev_priv: i915 device instance
886  * @state: target DC power state
887  * - DC_STATE_DISABLE
888  * - DC_STATE_EN_UPTO_DC5
889  * - DC_STATE_EN_UPTO_DC6
890  * - DC_STATE_EN_DC9
891  *
892  * Signal to DMC firmware/HW the target DC power state passed in @state.
893  * DMC/HW can turn off individual display clocks and power rails when entering
894  * a deeper DC power state (higher in number) and turns these back when exiting
895  * that state to a shallower power state (lower in number). The HW will decide
896  * when to actually enter a given state on an on-demand basis, for instance
897  * depending on the active state of display pipes. The state of display
898  * registers backed by affected power rails are saved/restored as needed.
899  *
900  * Based on the above enabling a deeper DC power state is asynchronous wrt.
901  * enabling it. Disabling a deeper power state is synchronous: for instance
902  * setting %DC_STATE_DISABLE won't complete until all HW resources are turned
903  * back on and register state is restored. This is guaranteed by the MMIO write
904  * to DC_STATE_EN blocking until the state is restored.
905  */
906 static void gen9_set_dc_state(struct drm_i915_private *dev_priv, u32 state)
907 {
908         u32 val;
909         u32 mask;
910
911         if (WARN_ON_ONCE(state & ~dev_priv->csr.allowed_dc_mask))
912                 state &= dev_priv->csr.allowed_dc_mask;
913
914         val = I915_READ(DC_STATE_EN);
915         mask = gen9_dc_mask(dev_priv);
916         DRM_DEBUG_KMS("Setting DC state from %02x to %02x\n",
917                       val & mask, state);
918
919         /* Check if DMC is ignoring our DC state requests */
920         if ((val & mask) != dev_priv->csr.dc_state)
921                 DRM_ERROR("DC state mismatch (0x%x -> 0x%x)\n",
922                           dev_priv->csr.dc_state, val & mask);
923
924         val &= ~mask;
925         val |= state;
926
927         gen9_write_dc_state(dev_priv, val);
928
929         dev_priv->csr.dc_state = val & mask;
930 }
931
932 void bxt_enable_dc9(struct drm_i915_private *dev_priv)
933 {
934         assert_can_enable_dc9(dev_priv);
935
936         DRM_DEBUG_KMS("Enabling DC9\n");
937         /*
938          * Power sequencer reset is not needed on
939          * platforms with South Display Engine on PCH,
940          * because PPS registers are always on.
941          */
942         if (!HAS_PCH_SPLIT(dev_priv))
943                 intel_power_sequencer_reset(dev_priv);
944         gen9_set_dc_state(dev_priv, DC_STATE_EN_DC9);
945 }
946
947 void bxt_disable_dc9(struct drm_i915_private *dev_priv)
948 {
949         assert_can_disable_dc9(dev_priv);
950
951         DRM_DEBUG_KMS("Disabling DC9\n");
952
953         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
954
955         intel_pps_unlock_regs_wa(dev_priv);
956 }
957
958 static void assert_csr_loaded(struct drm_i915_private *dev_priv)
959 {
960         WARN_ONCE(!I915_READ(CSR_PROGRAM(0)),
961                   "CSR program storage start is NULL\n");
962         WARN_ONCE(!I915_READ(CSR_SSP_BASE), "CSR SSP Base Not fine\n");
963         WARN_ONCE(!I915_READ(CSR_HTP_SKL), "CSR HTP Not fine\n");
964 }
965
966 static struct i915_power_well *
967 lookup_power_well(struct drm_i915_private *dev_priv,
968                   enum i915_power_well_id power_well_id)
969 {
970         struct i915_power_well *power_well;
971
972         for_each_power_well(dev_priv, power_well)
973                 if (power_well->desc->id == power_well_id)
974                         return power_well;
975
976         /*
977          * It's not feasible to add error checking code to the callers since
978          * this condition really shouldn't happen and it doesn't even make sense
979          * to abort things like display initialization sequences. Just return
980          * the first power well and hope the WARN gets reported so we can fix
981          * our driver.
982          */
983         WARN(1, "Power well %d not defined for this platform\n", power_well_id);
984         return &dev_priv->power_domains.power_wells[0];
985 }
986
987 static void assert_can_enable_dc5(struct drm_i915_private *dev_priv)
988 {
989         bool pg2_enabled = intel_display_power_well_is_enabled(dev_priv,
990                                         SKL_DISP_PW_2);
991
992         WARN_ONCE(pg2_enabled, "PG2 not disabled to enable DC5.\n");
993
994         WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5),
995                   "DC5 already programmed to be enabled.\n");
996         assert_rpm_wakelock_held(dev_priv);
997
998         assert_csr_loaded(dev_priv);
999 }
1000
1001 void gen9_enable_dc5(struct drm_i915_private *dev_priv)
1002 {
1003         assert_can_enable_dc5(dev_priv);
1004
1005         DRM_DEBUG_KMS("Enabling DC5\n");
1006
1007         /* Wa Display #1183: skl,kbl,cfl */
1008         if (IS_GEN9_BC(dev_priv))
1009                 I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) |
1010                            SKL_SELECT_ALTERNATE_DC_EXIT);
1011
1012         gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC5);
1013 }
1014
1015 static void assert_can_enable_dc6(struct drm_i915_private *dev_priv)
1016 {
1017         WARN_ONCE(I915_READ(UTIL_PIN_CTL) & UTIL_PIN_ENABLE,
1018                   "Backlight is not disabled.\n");
1019         WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC6),
1020                   "DC6 already programmed to be enabled.\n");
1021
1022         assert_csr_loaded(dev_priv);
1023 }
1024
1025 void skl_enable_dc6(struct drm_i915_private *dev_priv)
1026 {
1027         assert_can_enable_dc6(dev_priv);
1028
1029         DRM_DEBUG_KMS("Enabling DC6\n");
1030
1031         /* Wa Display #1183: skl,kbl,cfl */
1032         if (IS_GEN9_BC(dev_priv))
1033                 I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) |
1034                            SKL_SELECT_ALTERNATE_DC_EXIT);
1035
1036         gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC6);
1037 }
1038
1039 static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv,
1040                                    struct i915_power_well *power_well)
1041 {
1042         const struct i915_power_well_regs *regs = power_well->desc->hsw.regs;
1043         int pw_idx = power_well->desc->hsw.idx;
1044         u32 mask = HSW_PWR_WELL_CTL_REQ(pw_idx);
1045         u32 bios_req = I915_READ(regs->bios);
1046
1047         /* Take over the request bit if set by BIOS. */
1048         if (bios_req & mask) {
1049                 u32 drv_req = I915_READ(regs->driver);
1050
1051                 if (!(drv_req & mask))
1052                         I915_WRITE(regs->driver, drv_req | mask);
1053                 I915_WRITE(regs->bios, bios_req & ~mask);
1054         }
1055 }
1056
1057 static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
1058                                            struct i915_power_well *power_well)
1059 {
1060         bxt_ddi_phy_init(dev_priv, power_well->desc->bxt.phy);
1061 }
1062
1063 static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
1064                                             struct i915_power_well *power_well)
1065 {
1066         bxt_ddi_phy_uninit(dev_priv, power_well->desc->bxt.phy);
1067 }
1068
1069 static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *dev_priv,
1070                                             struct i915_power_well *power_well)
1071 {
1072         return bxt_ddi_phy_is_enabled(dev_priv, power_well->desc->bxt.phy);
1073 }
1074
1075 static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *dev_priv)
1076 {
1077         struct i915_power_well *power_well;
1078
1079         power_well = lookup_power_well(dev_priv, BXT_DISP_PW_DPIO_CMN_A);
1080         if (power_well->count > 0)
1081                 bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy);
1082
1083         power_well = lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
1084         if (power_well->count > 0)
1085                 bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy);
1086
1087         if (IS_GEMINILAKE(dev_priv)) {
1088                 power_well = lookup_power_well(dev_priv,
1089                                                GLK_DISP_PW_DPIO_CMN_C);
1090                 if (power_well->count > 0)
1091                         bxt_ddi_phy_verify_state(dev_priv,
1092                                                  power_well->desc->bxt.phy);
1093         }
1094 }
1095
1096 static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *dev_priv,
1097                                            struct i915_power_well *power_well)
1098 {
1099         return (I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5_DC6_MASK) == 0;
1100 }
1101
1102 static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv)
1103 {
1104         u32 tmp = I915_READ(DBUF_CTL);
1105
1106         WARN((tmp & (DBUF_POWER_STATE | DBUF_POWER_REQUEST)) !=
1107              (DBUF_POWER_STATE | DBUF_POWER_REQUEST),
1108              "Unexpected DBuf power power state (0x%08x)\n", tmp);
1109 }
1110
1111 static void gen9_dc_off_power_well_enable(struct drm_i915_private *dev_priv,
1112                                           struct i915_power_well *power_well)
1113 {
1114         struct intel_cdclk_state cdclk_state = {};
1115
1116         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
1117
1118         dev_priv->display.get_cdclk(dev_priv, &cdclk_state);
1119         /* Can't read out voltage_level so can't use intel_cdclk_changed() */
1120         WARN_ON(intel_cdclk_needs_modeset(&dev_priv->cdclk.hw, &cdclk_state));
1121
1122         gen9_assert_dbuf_enabled(dev_priv);
1123
1124         if (IS_GEN9_LP(dev_priv))
1125                 bxt_verify_ddi_phy_power_wells(dev_priv);
1126
1127         if (INTEL_GEN(dev_priv) >= 11)
1128                 /*
1129                  * DMC retains HW context only for port A, the other combo
1130                  * PHY's HW context for port B is lost after DC transitions,
1131                  * so we need to restore it manually.
1132                  */
1133                 icl_combo_phys_init(dev_priv);
1134 }
1135
1136 static void gen9_dc_off_power_well_disable(struct drm_i915_private *dev_priv,
1137                                            struct i915_power_well *power_well)
1138 {
1139         if (!dev_priv->csr.dmc_payload)
1140                 return;
1141
1142         if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6)
1143                 skl_enable_dc6(dev_priv);
1144         else if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)
1145                 gen9_enable_dc5(dev_priv);
1146 }
1147
1148 static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *dev_priv,
1149                                          struct i915_power_well *power_well)
1150 {
1151 }
1152
1153 static void i9xx_always_on_power_well_noop(struct drm_i915_private *dev_priv,
1154                                            struct i915_power_well *power_well)
1155 {
1156 }
1157
1158 static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *dev_priv,
1159                                              struct i915_power_well *power_well)
1160 {
1161         return true;
1162 }
1163
1164 static void i830_pipes_power_well_enable(struct drm_i915_private *dev_priv,
1165                                          struct i915_power_well *power_well)
1166 {
1167         if ((I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE) == 0)
1168                 i830_enable_pipe(dev_priv, PIPE_A);
1169         if ((I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE) == 0)
1170                 i830_enable_pipe(dev_priv, PIPE_B);
1171 }
1172
1173 static void i830_pipes_power_well_disable(struct drm_i915_private *dev_priv,
1174                                           struct i915_power_well *power_well)
1175 {
1176         i830_disable_pipe(dev_priv, PIPE_B);
1177         i830_disable_pipe(dev_priv, PIPE_A);
1178 }
1179
1180 static bool i830_pipes_power_well_enabled(struct drm_i915_private *dev_priv,
1181                                           struct i915_power_well *power_well)
1182 {
1183         return I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE &&
1184                 I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE;
1185 }
1186
1187 static void i830_pipes_power_well_sync_hw(struct drm_i915_private *dev_priv,
1188                                           struct i915_power_well *power_well)
1189 {
1190         if (power_well->count > 0)
1191                 i830_pipes_power_well_enable(dev_priv, power_well);
1192         else
1193                 i830_pipes_power_well_disable(dev_priv, power_well);
1194 }
1195
1196 static void vlv_set_power_well(struct drm_i915_private *dev_priv,
1197                                struct i915_power_well *power_well, bool enable)
1198 {
1199         int pw_idx = power_well->desc->vlv.idx;
1200         u32 mask;
1201         u32 state;
1202         u32 ctrl;
1203
1204         mask = PUNIT_PWRGT_MASK(pw_idx);
1205         state = enable ? PUNIT_PWRGT_PWR_ON(pw_idx) :
1206                          PUNIT_PWRGT_PWR_GATE(pw_idx);
1207
1208         mutex_lock(&dev_priv->pcu_lock);
1209
1210 #define COND \
1211         ((vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask) == state)
1212
1213         if (COND)
1214                 goto out;
1215
1216         ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL);
1217         ctrl &= ~mask;
1218         ctrl |= state;
1219         vlv_punit_write(dev_priv, PUNIT_REG_PWRGT_CTRL, ctrl);
1220
1221         if (wait_for(COND, 100))
1222                 DRM_ERROR("timeout setting power well state %08x (%08x)\n",
1223                           state,
1224                           vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL));
1225
1226 #undef COND
1227
1228 out:
1229         mutex_unlock(&dev_priv->pcu_lock);
1230 }
1231
1232 static void vlv_power_well_enable(struct drm_i915_private *dev_priv,
1233                                   struct i915_power_well *power_well)
1234 {
1235         vlv_set_power_well(dev_priv, power_well, true);
1236 }
1237
1238 static void vlv_power_well_disable(struct drm_i915_private *dev_priv,
1239                                    struct i915_power_well *power_well)
1240 {
1241         vlv_set_power_well(dev_priv, power_well, false);
1242 }
1243
1244 static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv,
1245                                    struct i915_power_well *power_well)
1246 {
1247         int pw_idx = power_well->desc->vlv.idx;
1248         bool enabled = false;
1249         u32 mask;
1250         u32 state;
1251         u32 ctrl;
1252
1253         mask = PUNIT_PWRGT_MASK(pw_idx);
1254         ctrl = PUNIT_PWRGT_PWR_ON(pw_idx);
1255
1256         mutex_lock(&dev_priv->pcu_lock);
1257
1258         state = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask;
1259         /*
1260          * We only ever set the power-on and power-gate states, anything
1261          * else is unexpected.
1262          */
1263         WARN_ON(state != PUNIT_PWRGT_PWR_ON(pw_idx) &&
1264                 state != PUNIT_PWRGT_PWR_GATE(pw_idx));
1265         if (state == ctrl)
1266                 enabled = true;
1267
1268         /*
1269          * A transient state at this point would mean some unexpected party
1270          * is poking at the power controls too.
1271          */
1272         ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL) & mask;
1273         WARN_ON(ctrl != state);
1274
1275         mutex_unlock(&dev_priv->pcu_lock);
1276
1277         return enabled;
1278 }
1279
1280 static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv)
1281 {
1282         u32 val;
1283
1284         /*
1285          * On driver load, a pipe may be active and driving a DSI display.
1286          * Preserve DPOUNIT_CLOCK_GATE_DISABLE to avoid the pipe getting stuck
1287          * (and never recovering) in this case. intel_dsi_post_disable() will
1288          * clear it when we turn off the display.
1289          */
1290         val = I915_READ(DSPCLK_GATE_D);
1291         val &= DPOUNIT_CLOCK_GATE_DISABLE;
1292         val |= VRHUNIT_CLOCK_GATE_DISABLE;
1293         I915_WRITE(DSPCLK_GATE_D, val);
1294
1295         /*
1296          * Disable trickle feed and enable pnd deadline calculation
1297          */
1298         I915_WRITE(MI_ARB_VLV, MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE);
1299         I915_WRITE(CBR1_VLV, 0);
1300
1301         WARN_ON(dev_priv->rawclk_freq == 0);
1302
1303         I915_WRITE(RAWCLK_FREQ_VLV,
1304                    DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 1000));
1305 }
1306
1307 static void vlv_display_power_well_init(struct drm_i915_private *dev_priv)
1308 {
1309         struct intel_encoder *encoder;
1310         enum pipe pipe;
1311
1312         /*
1313          * Enable the CRI clock source so we can get at the
1314          * display and the reference clock for VGA
1315          * hotplug / manual detection. Supposedly DSI also
1316          * needs the ref clock up and running.
1317          *
1318          * CHV DPLL B/C have some issues if VGA mode is enabled.
1319          */
1320         for_each_pipe(dev_priv, pipe) {
1321                 u32 val = I915_READ(DPLL(pipe));
1322
1323                 val |= DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS;
1324                 if (pipe != PIPE_A)
1325                         val |= DPLL_INTEGRATED_CRI_CLK_VLV;
1326
1327                 I915_WRITE(DPLL(pipe), val);
1328         }
1329
1330         vlv_init_display_clock_gating(dev_priv);
1331
1332         spin_lock_irq(&dev_priv->irq_lock);
1333         valleyview_enable_display_irqs(dev_priv);
1334         spin_unlock_irq(&dev_priv->irq_lock);
1335
1336         /*
1337          * During driver initialization/resume we can avoid restoring the
1338          * part of the HW/SW state that will be inited anyway explicitly.
1339          */
1340         if (dev_priv->power_domains.initializing)
1341                 return;
1342
1343         intel_hpd_init(dev_priv);
1344
1345         /* Re-enable the ADPA, if we have one */
1346         for_each_intel_encoder(&dev_priv->drm, encoder) {
1347                 if (encoder->type == INTEL_OUTPUT_ANALOG)
1348                         intel_crt_reset(&encoder->base);
1349         }
1350
1351         i915_redisable_vga_power_on(dev_priv);
1352
1353         intel_pps_unlock_regs_wa(dev_priv);
1354 }
1355
1356 static void vlv_display_power_well_deinit(struct drm_i915_private *dev_priv)
1357 {
1358         spin_lock_irq(&dev_priv->irq_lock);
1359         valleyview_disable_display_irqs(dev_priv);
1360         spin_unlock_irq(&dev_priv->irq_lock);
1361
1362         /* make sure we're done processing display irqs */
1363         synchronize_irq(dev_priv->drm.irq);
1364
1365         intel_power_sequencer_reset(dev_priv);
1366
1367         /* Prevent us from re-enabling polling on accident in late suspend */
1368         if (!dev_priv->drm.dev->power.is_suspended)
1369                 intel_hpd_poll_init(dev_priv);
1370 }
1371
1372 static void vlv_display_power_well_enable(struct drm_i915_private *dev_priv,
1373                                           struct i915_power_well *power_well)
1374 {
1375         vlv_set_power_well(dev_priv, power_well, true);
1376
1377         vlv_display_power_well_init(dev_priv);
1378 }
1379
1380 static void vlv_display_power_well_disable(struct drm_i915_private *dev_priv,
1381                                            struct i915_power_well *power_well)
1382 {
1383         vlv_display_power_well_deinit(dev_priv);
1384
1385         vlv_set_power_well(dev_priv, power_well, false);
1386 }
1387
1388 static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
1389                                            struct i915_power_well *power_well)
1390 {
1391         /* since ref/cri clock was enabled */
1392         udelay(1); /* >10ns for cmnreset, >0ns for sidereset */
1393
1394         vlv_set_power_well(dev_priv, power_well, true);
1395
1396         /*
1397          * From VLV2A0_DP_eDP_DPIO_driver_vbios_notes_10.docx -
1398          *  6.  De-assert cmn_reset/side_reset. Same as VLV X0.
1399          *   a. GUnit 0x2110 bit[0] set to 1 (def 0)
1400          *   b. The other bits such as sfr settings / modesel may all
1401          *      be set to 0.
1402          *
1403          * This should only be done on init and resume from S3 with
1404          * both PLLs disabled, or we risk losing DPIO and PLL
1405          * synchronization.
1406          */
1407         I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) | DPIO_CMNRST);
1408 }
1409
1410 static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
1411                                             struct i915_power_well *power_well)
1412 {
1413         enum pipe pipe;
1414
1415         for_each_pipe(dev_priv, pipe)
1416                 assert_pll_disabled(dev_priv, pipe);
1417
1418         /* Assert common reset */
1419         I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) & ~DPIO_CMNRST);
1420
1421         vlv_set_power_well(dev_priv, power_well, false);
1422 }
1423
1424 #define POWER_DOMAIN_MASK (GENMASK_ULL(POWER_DOMAIN_NUM - 1, 0))
1425
1426 #define BITS_SET(val, bits) (((val) & (bits)) == (bits))
1427
1428 static void assert_chv_phy_status(struct drm_i915_private *dev_priv)
1429 {
1430         struct i915_power_well *cmn_bc =
1431                 lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
1432         struct i915_power_well *cmn_d =
1433                 lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D);
1434         u32 phy_control = dev_priv->chv_phy_control;
1435         u32 phy_status = 0;
1436         u32 phy_status_mask = 0xffffffff;
1437
1438         /*
1439          * The BIOS can leave the PHY is some weird state
1440          * where it doesn't fully power down some parts.
1441          * Disable the asserts until the PHY has been fully
1442          * reset (ie. the power well has been disabled at
1443          * least once).
1444          */
1445         if (!dev_priv->chv_phy_assert[DPIO_PHY0])
1446                 phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH0) |
1447                                      PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 0) |
1448                                      PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 1) |
1449                                      PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH1) |
1450                                      PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 0) |
1451                                      PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1));
1452
1453         if (!dev_priv->chv_phy_assert[DPIO_PHY1])
1454                 phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY1, DPIO_CH0) |
1455                                      PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 0) |
1456                                      PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 1));
1457
1458         if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) {
1459                 phy_status |= PHY_POWERGOOD(DPIO_PHY0);
1460
1461                 /* this assumes override is only used to enable lanes */
1462                 if ((phy_control & PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH0)) == 0)
1463                         phy_control |= PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY0, DPIO_CH0);
1464
1465                 if ((phy_control & PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH1)) == 0)
1466                         phy_control |= PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY0, DPIO_CH1);
1467
1468                 /* CL1 is on whenever anything is on in either channel */
1469                 if (BITS_SET(phy_control,
1470                              PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY0, DPIO_CH0) |
1471                              PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY0, DPIO_CH1)))
1472                         phy_status |= PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH0);
1473
1474                 /*
1475                  * The DPLLB check accounts for the pipe B + port A usage
1476                  * with CL2 powered up but all the lanes in the second channel
1477                  * powered down.
1478                  */
1479                 if (BITS_SET(phy_control,
1480                              PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY0, DPIO_CH1)) &&
1481                     (I915_READ(DPLL(PIPE_B)) & DPLL_VCO_ENABLE) == 0)
1482                         phy_status |= PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH1);
1483
1484                 if (BITS_SET(phy_control,
1485                              PHY_CH_POWER_DOWN_OVRD(0x3, DPIO_PHY0, DPIO_CH0)))
1486                         phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 0);
1487                 if (BITS_SET(phy_control,
1488                              PHY_CH_POWER_DOWN_OVRD(0xc, DPIO_PHY0, DPIO_CH0)))
1489                         phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 1);
1490
1491                 if (BITS_SET(phy_control,
1492                              PHY_CH_POWER_DOWN_OVRD(0x3, DPIO_PHY0, DPIO_CH1)))
1493                         phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 0);
1494                 if (BITS_SET(phy_control,
1495                              PHY_CH_POWER_DOWN_OVRD(0xc, DPIO_PHY0, DPIO_CH1)))
1496                         phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1);
1497         }
1498
1499         if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) {
1500                 phy_status |= PHY_POWERGOOD(DPIO_PHY1);
1501
1502                 /* this assumes override is only used to enable lanes */
1503                 if ((phy_control & PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY1, DPIO_CH0)) == 0)
1504                         phy_control |= PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY1, DPIO_CH0);
1505
1506                 if (BITS_SET(phy_control,
1507                              PHY_CH_POWER_DOWN_OVRD(0xf, DPIO_PHY1, DPIO_CH0)))
1508                         phy_status |= PHY_STATUS_CMN_LDO(DPIO_PHY1, DPIO_CH0);
1509
1510                 if (BITS_SET(phy_control,
1511                              PHY_CH_POWER_DOWN_OVRD(0x3, DPIO_PHY1, DPIO_CH0)))
1512                         phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 0);
1513                 if (BITS_SET(phy_control,
1514                              PHY_CH_POWER_DOWN_OVRD(0xc, DPIO_PHY1, DPIO_CH0)))
1515                         phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 1);
1516         }
1517
1518         phy_status &= phy_status_mask;
1519
1520         /*
1521          * The PHY may be busy with some initial calibration and whatnot,
1522          * so the power state can take a while to actually change.
1523          */
1524         if (intel_wait_for_register(dev_priv,
1525                                     DISPLAY_PHY_STATUS,
1526                                     phy_status_mask,
1527                                     phy_status,
1528                                     10))
1529                 DRM_ERROR("Unexpected PHY_STATUS 0x%08x, expected 0x%08x (PHY_CONTROL=0x%08x)\n",
1530                           I915_READ(DISPLAY_PHY_STATUS) & phy_status_mask,
1531                            phy_status, dev_priv->chv_phy_control);
1532 }
1533
1534 #undef BITS_SET
1535
1536 static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv,
1537                                            struct i915_power_well *power_well)
1538 {
1539         enum dpio_phy phy;
1540         enum pipe pipe;
1541         u32 tmp;
1542
1543         WARN_ON_ONCE(power_well->desc->id != VLV_DISP_PW_DPIO_CMN_BC &&
1544                      power_well->desc->id != CHV_DISP_PW_DPIO_CMN_D);
1545
1546         if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
1547                 pipe = PIPE_A;
1548                 phy = DPIO_PHY0;
1549         } else {
1550                 pipe = PIPE_C;
1551                 phy = DPIO_PHY1;
1552         }
1553
1554         /* since ref/cri clock was enabled */
1555         udelay(1); /* >10ns for cmnreset, >0ns for sidereset */
1556         vlv_set_power_well(dev_priv, power_well, true);
1557
1558         /* Poll for phypwrgood signal */
1559         if (intel_wait_for_register(dev_priv,
1560                                     DISPLAY_PHY_STATUS,
1561                                     PHY_POWERGOOD(phy),
1562                                     PHY_POWERGOOD(phy),
1563                                     1))
1564                 DRM_ERROR("Display PHY %d is not power up\n", phy);
1565
1566         mutex_lock(&dev_priv->sb_lock);
1567
1568         /* Enable dynamic power down */
1569         tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW28);
1570         tmp |= DPIO_DYNPWRDOWNEN_CH0 | DPIO_CL1POWERDOWNEN |
1571                 DPIO_SUS_CLK_CONFIG_GATE_CLKREQ;
1572         vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW28, tmp);
1573
1574         if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
1575                 tmp = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW6_CH1);
1576                 tmp |= DPIO_DYNPWRDOWNEN_CH1;
1577                 vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW6_CH1, tmp);
1578         } else {
1579                 /*
1580                  * Force the non-existing CL2 off. BXT does this
1581                  * too, so maybe it saves some power even though
1582                  * CL2 doesn't exist?
1583                  */
1584                 tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW30);
1585                 tmp |= DPIO_CL2_LDOFUSE_PWRENB;
1586                 vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW30, tmp);
1587         }
1588
1589         mutex_unlock(&dev_priv->sb_lock);
1590
1591         dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy);
1592         I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
1593
1594         DRM_DEBUG_KMS("Enabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n",
1595                       phy, dev_priv->chv_phy_control);
1596
1597         assert_chv_phy_status(dev_priv);
1598 }
1599
1600 static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv,
1601                                             struct i915_power_well *power_well)
1602 {
1603         enum dpio_phy phy;
1604
1605         WARN_ON_ONCE(power_well->desc->id != VLV_DISP_PW_DPIO_CMN_BC &&
1606                      power_well->desc->id != CHV_DISP_PW_DPIO_CMN_D);
1607
1608         if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) {
1609                 phy = DPIO_PHY0;
1610                 assert_pll_disabled(dev_priv, PIPE_A);
1611                 assert_pll_disabled(dev_priv, PIPE_B);
1612         } else {
1613                 phy = DPIO_PHY1;
1614                 assert_pll_disabled(dev_priv, PIPE_C);
1615         }
1616
1617         dev_priv->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy);
1618         I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
1619
1620         vlv_set_power_well(dev_priv, power_well, false);
1621
1622         DRM_DEBUG_KMS("Disabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n",
1623                       phy, dev_priv->chv_phy_control);
1624
1625         /* PHY is fully reset now, so we can enable the PHY state asserts */
1626         dev_priv->chv_phy_assert[phy] = true;
1627
1628         assert_chv_phy_status(dev_priv);
1629 }
1630
1631 static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpio_phy phy,
1632                                      enum dpio_channel ch, bool override, unsigned int mask)
1633 {
1634         enum pipe pipe = phy == DPIO_PHY0 ? PIPE_A : PIPE_C;
1635         u32 reg, val, expected, actual;
1636
1637         /*
1638          * The BIOS can leave the PHY is some weird state
1639          * where it doesn't fully power down some parts.
1640          * Disable the asserts until the PHY has been fully
1641          * reset (ie. the power well has been disabled at
1642          * least once).
1643          */
1644         if (!dev_priv->chv_phy_assert[phy])
1645                 return;
1646
1647         if (ch == DPIO_CH0)
1648                 reg = _CHV_CMN_DW0_CH0;
1649         else
1650                 reg = _CHV_CMN_DW6_CH1;
1651
1652         mutex_lock(&dev_priv->sb_lock);
1653         val = vlv_dpio_read(dev_priv, pipe, reg);
1654         mutex_unlock(&dev_priv->sb_lock);
1655
1656         /*
1657          * This assumes !override is only used when the port is disabled.
1658          * All lanes should power down even without the override when
1659          * the port is disabled.
1660          */
1661         if (!override || mask == 0xf) {
1662                 expected = DPIO_ALLDL_POWERDOWN | DPIO_ANYDL_POWERDOWN;
1663                 /*
1664                  * If CH1 common lane is not active anymore
1665                  * (eg. for pipe B DPLL) the entire channel will
1666                  * shut down, which causes the common lane registers
1667                  * to read as 0. That means we can't actually check
1668                  * the lane power down status bits, but as the entire
1669                  * register reads as 0 it's a good indication that the
1670                  * channel is indeed entirely powered down.
1671                  */
1672                 if (ch == DPIO_CH1 && val == 0)
1673                         expected = 0;
1674         } else if (mask != 0x0) {
1675                 expected = DPIO_ANYDL_POWERDOWN;
1676         } else {
1677                 expected = 0;
1678         }
1679
1680         if (ch == DPIO_CH0)
1681                 actual = val >> DPIO_ANYDL_POWERDOWN_SHIFT_CH0;
1682         else
1683                 actual = val >> DPIO_ANYDL_POWERDOWN_SHIFT_CH1;
1684         actual &= DPIO_ALLDL_POWERDOWN | DPIO_ANYDL_POWERDOWN;
1685
1686         WARN(actual != expected,
1687              "Unexpected DPIO lane power down: all %d, any %d. Expected: all %d, any %d. (0x%x = 0x%08x)\n",
1688              !!(actual & DPIO_ALLDL_POWERDOWN), !!(actual & DPIO_ANYDL_POWERDOWN),
1689              !!(expected & DPIO_ALLDL_POWERDOWN), !!(expected & DPIO_ANYDL_POWERDOWN),
1690              reg, val);
1691 }
1692
1693 bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy,
1694                           enum dpio_channel ch, bool override)
1695 {
1696         struct i915_power_domains *power_domains = &dev_priv->power_domains;
1697         bool was_override;
1698
1699         mutex_lock(&power_domains->lock);
1700
1701         was_override = dev_priv->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
1702
1703         if (override == was_override)
1704                 goto out;
1705
1706         if (override)
1707                 dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
1708         else
1709                 dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
1710
1711         I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
1712
1713         DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d (DPIO_PHY_CONTROL=0x%08x)\n",
1714                       phy, ch, dev_priv->chv_phy_control);
1715
1716         assert_chv_phy_status(dev_priv);
1717
1718 out:
1719         mutex_unlock(&power_domains->lock);
1720
1721         return was_override;
1722 }
1723
1724 void chv_phy_powergate_lanes(struct intel_encoder *encoder,
1725                              bool override, unsigned int mask)
1726 {
1727         struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
1728         struct i915_power_domains *power_domains = &dev_priv->power_domains;
1729         enum dpio_phy phy = vlv_dport_to_phy(enc_to_dig_port(&encoder->base));
1730         enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base));
1731
1732         mutex_lock(&power_domains->lock);
1733
1734         dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch);
1735         dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch);
1736
1737         if (override)
1738                 dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
1739         else
1740                 dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch);
1741
1742         I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
1743
1744         DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d lanes 0x%x (PHY_CONTROL=0x%08x)\n",
1745                       phy, ch, mask, dev_priv->chv_phy_control);
1746
1747         assert_chv_phy_status(dev_priv);
1748
1749         assert_chv_phy_powergate(dev_priv, phy, ch, override, mask);
1750
1751         mutex_unlock(&power_domains->lock);
1752 }
1753
1754 static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv,
1755                                         struct i915_power_well *power_well)
1756 {
1757         enum pipe pipe = PIPE_A;
1758         bool enabled;
1759         u32 state, ctrl;
1760
1761         mutex_lock(&dev_priv->pcu_lock);
1762
1763         state = vlv_punit_read(dev_priv, PUNIT_REG_DSPFREQ) & DP_SSS_MASK(pipe);
1764         /*
1765          * We only ever set the power-on and power-gate states, anything
1766          * else is unexpected.
1767          */
1768         WARN_ON(state != DP_SSS_PWR_ON(pipe) && state != DP_SSS_PWR_GATE(pipe));
1769         enabled = state == DP_SSS_PWR_ON(pipe);
1770
1771         /*
1772          * A transient state at this point would mean some unexpected party
1773          * is poking at the power controls too.
1774          */
1775         ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPFREQ) & DP_SSC_MASK(pipe);
1776         WARN_ON(ctrl << 16 != state);
1777
1778         mutex_unlock(&dev_priv->pcu_lock);
1779
1780         return enabled;
1781 }
1782
1783 static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv,
1784                                     struct i915_power_well *power_well,
1785                                     bool enable)
1786 {
1787         enum pipe pipe = PIPE_A;
1788         u32 state;
1789         u32 ctrl;
1790
1791         state = enable ? DP_SSS_PWR_ON(pipe) : DP_SSS_PWR_GATE(pipe);
1792
1793         mutex_lock(&dev_priv->pcu_lock);
1794
1795 #define COND \
1796         ((vlv_punit_read(dev_priv, PUNIT_REG_DSPFREQ) & DP_SSS_MASK(pipe)) == state)
1797
1798         if (COND)
1799                 goto out;
1800
1801         ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPFREQ);
1802         ctrl &= ~DP_SSC_MASK(pipe);
1803         ctrl |= enable ? DP_SSC_PWR_ON(pipe) : DP_SSC_PWR_GATE(pipe);
1804         vlv_punit_write(dev_priv, PUNIT_REG_DSPFREQ, ctrl);
1805
1806         if (wait_for(COND, 100))
1807                 DRM_ERROR("timeout setting power well state %08x (%08x)\n",
1808                           state,
1809                           vlv_punit_read(dev_priv, PUNIT_REG_DSPFREQ));
1810
1811 #undef COND
1812
1813 out:
1814         mutex_unlock(&dev_priv->pcu_lock);
1815 }
1816
1817 static void chv_pipe_power_well_enable(struct drm_i915_private *dev_priv,
1818                                        struct i915_power_well *power_well)
1819 {
1820         chv_set_pipe_power_well(dev_priv, power_well, true);
1821
1822         vlv_display_power_well_init(dev_priv);
1823 }
1824
1825 static void chv_pipe_power_well_disable(struct drm_i915_private *dev_priv,
1826                                         struct i915_power_well *power_well)
1827 {
1828         vlv_display_power_well_deinit(dev_priv);
1829
1830         chv_set_pipe_power_well(dev_priv, power_well, false);
1831 }
1832
1833 static void
1834 __intel_display_power_get_domain(struct drm_i915_private *dev_priv,
1835                                  enum intel_display_power_domain domain)
1836 {
1837         struct i915_power_domains *power_domains = &dev_priv->power_domains;
1838         struct i915_power_well *power_well;
1839
1840         for_each_power_domain_well(dev_priv, power_well, BIT_ULL(domain))
1841                 intel_power_well_get(dev_priv, power_well);
1842
1843         power_domains->domain_use_count[domain]++;
1844 }
1845
1846 /**
1847  * intel_display_power_get - grab a power domain reference
1848  * @dev_priv: i915 device instance
1849  * @domain: power domain to reference
1850  *
1851  * This function grabs a power domain reference for @domain and ensures that the
1852  * power domain and all its parents are powered up. Therefore users should only
1853  * grab a reference to the innermost power domain they need.
1854  *
1855  * Any power domain reference obtained by this function must have a symmetric
1856  * call to intel_display_power_put() to release the reference again.
1857  */
1858 intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv,
1859                                         enum intel_display_power_domain domain)
1860 {
1861         struct i915_power_domains *power_domains = &dev_priv->power_domains;
1862         intel_wakeref_t wakeref = intel_runtime_pm_get(dev_priv);
1863
1864         mutex_lock(&power_domains->lock);
1865
1866         __intel_display_power_get_domain(dev_priv, domain);
1867
1868         mutex_unlock(&power_domains->lock);
1869
1870         return wakeref;
1871 }
1872
1873 /**
1874  * intel_display_power_get_if_enabled - grab a reference for an enabled display power domain
1875  * @dev_priv: i915 device instance
1876  * @domain: power domain to reference
1877  *
1878  * This function grabs a power domain reference for @domain and ensures that the
1879  * power domain and all its parents are powered up. Therefore users should only
1880  * grab a reference to the innermost power domain they need.
1881  *
1882  * Any power domain reference obtained by this function must have a symmetric
1883  * call to intel_display_power_put() to release the reference again.
1884  */
1885 intel_wakeref_t
1886 intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv,
1887                                    enum intel_display_power_domain domain)
1888 {
1889         struct i915_power_domains *power_domains = &dev_priv->power_domains;
1890         intel_wakeref_t wakeref;
1891         bool is_enabled;
1892
1893         wakeref = intel_runtime_pm_get_if_in_use(dev_priv);
1894         if (!wakeref)
1895                 return false;
1896
1897         mutex_lock(&power_domains->lock);
1898
1899         if (__intel_display_power_is_enabled(dev_priv, domain)) {
1900                 __intel_display_power_get_domain(dev_priv, domain);
1901                 is_enabled = true;
1902         } else {
1903                 is_enabled = false;
1904         }
1905
1906         mutex_unlock(&power_domains->lock);
1907
1908         if (!is_enabled) {
1909                 intel_runtime_pm_put(dev_priv, wakeref);
1910                 wakeref = 0;
1911         }
1912
1913         return wakeref;
1914 }
1915
1916 static void __intel_display_power_put(struct drm_i915_private *dev_priv,
1917                                       enum intel_display_power_domain domain)
1918 {
1919         struct i915_power_domains *power_domains;
1920         struct i915_power_well *power_well;
1921
1922         power_domains = &dev_priv->power_domains;
1923
1924         mutex_lock(&power_domains->lock);
1925
1926         WARN(!power_domains->domain_use_count[domain],
1927              "Use count on domain %s is already zero\n",
1928              intel_display_power_domain_str(domain));
1929         power_domains->domain_use_count[domain]--;
1930
1931         for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain))
1932                 intel_power_well_put(dev_priv, power_well);
1933
1934         mutex_unlock(&power_domains->lock);
1935 }
1936
1937 /**
1938  * intel_display_power_put - release a power domain reference
1939  * @dev_priv: i915 device instance
1940  * @domain: power domain to reference
1941  *
1942  * This function drops the power domain reference obtained by
1943  * intel_display_power_get() and might power down the corresponding hardware
1944  * block right away if this is the last reference.
1945  */
1946 void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv,
1947                                        enum intel_display_power_domain domain)
1948 {
1949         __intel_display_power_put(dev_priv, domain);
1950         intel_runtime_pm_put_unchecked(dev_priv);
1951 }
1952
1953 #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
1954 void intel_display_power_put(struct drm_i915_private *dev_priv,
1955                              enum intel_display_power_domain domain,
1956                              intel_wakeref_t wakeref)
1957 {
1958         __intel_display_power_put(dev_priv, domain);
1959         intel_runtime_pm_put(dev_priv, wakeref);
1960 }
1961 #endif
1962
1963 #define I830_PIPES_POWER_DOMAINS (              \
1964         BIT_ULL(POWER_DOMAIN_PIPE_A) |          \
1965         BIT_ULL(POWER_DOMAIN_PIPE_B) |          \
1966         BIT_ULL(POWER_DOMAIN_PIPE_A_PANEL_FITTER) |     \
1967         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |     \
1968         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |    \
1969         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |    \
1970         BIT_ULL(POWER_DOMAIN_INIT))
1971
1972 #define VLV_DISPLAY_POWER_DOMAINS (             \
1973         BIT_ULL(POWER_DOMAIN_PIPE_A) |          \
1974         BIT_ULL(POWER_DOMAIN_PIPE_B) |          \
1975         BIT_ULL(POWER_DOMAIN_PIPE_A_PANEL_FITTER) |     \
1976         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |     \
1977         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |    \
1978         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |    \
1979         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
1980         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
1981         BIT_ULL(POWER_DOMAIN_PORT_DSI) |                \
1982         BIT_ULL(POWER_DOMAIN_PORT_CRT) |                \
1983         BIT_ULL(POWER_DOMAIN_VGA) |                     \
1984         BIT_ULL(POWER_DOMAIN_AUDIO) |           \
1985         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
1986         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
1987         BIT_ULL(POWER_DOMAIN_GMBUS) |           \
1988         BIT_ULL(POWER_DOMAIN_INIT))
1989
1990 #define VLV_DPIO_CMN_BC_POWER_DOMAINS (         \
1991         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
1992         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
1993         BIT_ULL(POWER_DOMAIN_PORT_CRT) |                \
1994         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
1995         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
1996         BIT_ULL(POWER_DOMAIN_INIT))
1997
1998 #define VLV_DPIO_TX_B_LANES_01_POWER_DOMAINS (  \
1999         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
2000         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
2001         BIT_ULL(POWER_DOMAIN_INIT))
2002
2003 #define VLV_DPIO_TX_B_LANES_23_POWER_DOMAINS (  \
2004         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
2005         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
2006         BIT_ULL(POWER_DOMAIN_INIT))
2007
2008 #define VLV_DPIO_TX_C_LANES_01_POWER_DOMAINS (  \
2009         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
2010         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
2011         BIT_ULL(POWER_DOMAIN_INIT))
2012
2013 #define VLV_DPIO_TX_C_LANES_23_POWER_DOMAINS (  \
2014         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
2015         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
2016         BIT_ULL(POWER_DOMAIN_INIT))
2017
2018 #define CHV_DISPLAY_POWER_DOMAINS (             \
2019         BIT_ULL(POWER_DOMAIN_PIPE_A) |          \
2020         BIT_ULL(POWER_DOMAIN_PIPE_B) |          \
2021         BIT_ULL(POWER_DOMAIN_PIPE_C) |          \
2022         BIT_ULL(POWER_DOMAIN_PIPE_A_PANEL_FITTER) |     \
2023         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |     \
2024         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |     \
2025         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |    \
2026         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |    \
2027         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |    \
2028         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
2029         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
2030         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |        \
2031         BIT_ULL(POWER_DOMAIN_PORT_DSI) |                \
2032         BIT_ULL(POWER_DOMAIN_VGA) |                     \
2033         BIT_ULL(POWER_DOMAIN_AUDIO) |           \
2034         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
2035         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
2036         BIT_ULL(POWER_DOMAIN_AUX_D) |           \
2037         BIT_ULL(POWER_DOMAIN_GMBUS) |           \
2038         BIT_ULL(POWER_DOMAIN_INIT))
2039
2040 #define CHV_DPIO_CMN_BC_POWER_DOMAINS (         \
2041         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
2042         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
2043         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
2044         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
2045         BIT_ULL(POWER_DOMAIN_INIT))
2046
2047 #define CHV_DPIO_CMN_D_POWER_DOMAINS (          \
2048         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |        \
2049         BIT_ULL(POWER_DOMAIN_AUX_D) |           \
2050         BIT_ULL(POWER_DOMAIN_INIT))
2051
2052 #define HSW_DISPLAY_POWER_DOMAINS (                     \
2053         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2054         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2055         BIT_ULL(POWER_DOMAIN_PIPE_A_PANEL_FITTER) |             \
2056         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |             \
2057         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |             \
2058         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2059         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2060         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2061         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2062         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2063         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |                \
2064         BIT_ULL(POWER_DOMAIN_PORT_CRT) | /* DDI E */    \
2065         BIT_ULL(POWER_DOMAIN_VGA) |                             \
2066         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2067         BIT_ULL(POWER_DOMAIN_INIT))
2068
2069 #define BDW_DISPLAY_POWER_DOMAINS (                     \
2070         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2071         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2072         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |             \
2073         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |             \
2074         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2075         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2076         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2077         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2078         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2079         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |                \
2080         BIT_ULL(POWER_DOMAIN_PORT_CRT) | /* DDI E */    \
2081         BIT_ULL(POWER_DOMAIN_VGA) |                             \
2082         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2083         BIT_ULL(POWER_DOMAIN_INIT))
2084
2085 #define SKL_DISPLAY_POWERWELL_2_POWER_DOMAINS (         \
2086         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2087         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2088         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2089         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2090         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2091         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |             \
2092         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |             \
2093         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2094         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2095         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |                \
2096         BIT_ULL(POWER_DOMAIN_PORT_DDI_E_LANES) |                \
2097         BIT_ULL(POWER_DOMAIN_AUX_B) |                       \
2098         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2099         BIT_ULL(POWER_DOMAIN_AUX_D) |                   \
2100         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2101         BIT_ULL(POWER_DOMAIN_VGA) |                             \
2102         BIT_ULL(POWER_DOMAIN_INIT))
2103 #define SKL_DISPLAY_DDI_IO_A_E_POWER_DOMAINS (          \
2104         BIT_ULL(POWER_DOMAIN_PORT_DDI_A_IO) |           \
2105         BIT_ULL(POWER_DOMAIN_PORT_DDI_E_IO) |           \
2106         BIT_ULL(POWER_DOMAIN_INIT))
2107 #define SKL_DISPLAY_DDI_IO_B_POWER_DOMAINS (            \
2108         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_IO) |           \
2109         BIT_ULL(POWER_DOMAIN_INIT))
2110 #define SKL_DISPLAY_DDI_IO_C_POWER_DOMAINS (            \
2111         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_IO) |           \
2112         BIT_ULL(POWER_DOMAIN_INIT))
2113 #define SKL_DISPLAY_DDI_IO_D_POWER_DOMAINS (            \
2114         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_IO) |           \
2115         BIT_ULL(POWER_DOMAIN_INIT))
2116 #define SKL_DISPLAY_DC_OFF_POWER_DOMAINS (              \
2117         SKL_DISPLAY_POWERWELL_2_POWER_DOMAINS |         \
2118         BIT_ULL(POWER_DOMAIN_GT_IRQ) |                  \
2119         BIT_ULL(POWER_DOMAIN_MODESET) |                 \
2120         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2121         BIT_ULL(POWER_DOMAIN_INIT))
2122
2123 #define BXT_DISPLAY_POWERWELL_2_POWER_DOMAINS (         \
2124         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2125         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2126         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2127         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2128         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2129         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |             \
2130         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |             \
2131         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2132         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2133         BIT_ULL(POWER_DOMAIN_AUX_B) |                   \
2134         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2135         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2136         BIT_ULL(POWER_DOMAIN_VGA) |                             \
2137         BIT_ULL(POWER_DOMAIN_INIT))
2138 #define BXT_DISPLAY_DC_OFF_POWER_DOMAINS (              \
2139         BXT_DISPLAY_POWERWELL_2_POWER_DOMAINS |         \
2140         BIT_ULL(POWER_DOMAIN_GT_IRQ) |                  \
2141         BIT_ULL(POWER_DOMAIN_MODESET) |                 \
2142         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2143         BIT_ULL(POWER_DOMAIN_GMBUS) |                   \
2144         BIT_ULL(POWER_DOMAIN_INIT))
2145 #define BXT_DPIO_CMN_A_POWER_DOMAINS (                  \
2146         BIT_ULL(POWER_DOMAIN_PORT_DDI_A_LANES) |                \
2147         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2148         BIT_ULL(POWER_DOMAIN_INIT))
2149 #define BXT_DPIO_CMN_BC_POWER_DOMAINS (                 \
2150         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2151         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2152         BIT_ULL(POWER_DOMAIN_AUX_B) |                   \
2153         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2154         BIT_ULL(POWER_DOMAIN_INIT))
2155
2156 #define GLK_DISPLAY_POWERWELL_2_POWER_DOMAINS (         \
2157         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2158         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2159         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2160         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2161         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2162         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |             \
2163         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |             \
2164         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2165         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2166         BIT_ULL(POWER_DOMAIN_AUX_B) |                       \
2167         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2168         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2169         BIT_ULL(POWER_DOMAIN_VGA) |                             \
2170         BIT_ULL(POWER_DOMAIN_INIT))
2171 #define GLK_DISPLAY_DDI_IO_A_POWER_DOMAINS (            \
2172         BIT_ULL(POWER_DOMAIN_PORT_DDI_A_IO))
2173 #define GLK_DISPLAY_DDI_IO_B_POWER_DOMAINS (            \
2174         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_IO))
2175 #define GLK_DISPLAY_DDI_IO_C_POWER_DOMAINS (            \
2176         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_IO))
2177 #define GLK_DPIO_CMN_A_POWER_DOMAINS (                  \
2178         BIT_ULL(POWER_DOMAIN_PORT_DDI_A_LANES) |                \
2179         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2180         BIT_ULL(POWER_DOMAIN_INIT))
2181 #define GLK_DPIO_CMN_B_POWER_DOMAINS (                  \
2182         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2183         BIT_ULL(POWER_DOMAIN_AUX_B) |                   \
2184         BIT_ULL(POWER_DOMAIN_INIT))
2185 #define GLK_DPIO_CMN_C_POWER_DOMAINS (                  \
2186         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2187         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2188         BIT_ULL(POWER_DOMAIN_INIT))
2189 #define GLK_DISPLAY_AUX_A_POWER_DOMAINS (               \
2190         BIT_ULL(POWER_DOMAIN_AUX_A) |           \
2191         BIT_ULL(POWER_DOMAIN_AUX_IO_A) |                \
2192         BIT_ULL(POWER_DOMAIN_INIT))
2193 #define GLK_DISPLAY_AUX_B_POWER_DOMAINS (               \
2194         BIT_ULL(POWER_DOMAIN_AUX_B) |           \
2195         BIT_ULL(POWER_DOMAIN_INIT))
2196 #define GLK_DISPLAY_AUX_C_POWER_DOMAINS (               \
2197         BIT_ULL(POWER_DOMAIN_AUX_C) |           \
2198         BIT_ULL(POWER_DOMAIN_INIT))
2199 #define GLK_DISPLAY_DC_OFF_POWER_DOMAINS (              \
2200         GLK_DISPLAY_POWERWELL_2_POWER_DOMAINS |         \
2201         BIT_ULL(POWER_DOMAIN_GT_IRQ) |                  \
2202         BIT_ULL(POWER_DOMAIN_MODESET) |                 \
2203         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2204         BIT_ULL(POWER_DOMAIN_GMBUS) |                   \
2205         BIT_ULL(POWER_DOMAIN_INIT))
2206
2207 #define CNL_DISPLAY_POWERWELL_2_POWER_DOMAINS (         \
2208         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2209         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2210         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2211         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2212         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2213         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |             \
2214         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |             \
2215         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |                \
2216         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |                \
2217         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |                \
2218         BIT_ULL(POWER_DOMAIN_PORT_DDI_F_LANES) |                \
2219         BIT_ULL(POWER_DOMAIN_AUX_B) |                       \
2220         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2221         BIT_ULL(POWER_DOMAIN_AUX_D) |                   \
2222         BIT_ULL(POWER_DOMAIN_AUX_F) |                   \
2223         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2224         BIT_ULL(POWER_DOMAIN_VGA) |                             \
2225         BIT_ULL(POWER_DOMAIN_INIT))
2226 #define CNL_DISPLAY_DDI_A_IO_POWER_DOMAINS (            \
2227         BIT_ULL(POWER_DOMAIN_PORT_DDI_A_IO) |           \
2228         BIT_ULL(POWER_DOMAIN_INIT))
2229 #define CNL_DISPLAY_DDI_B_IO_POWER_DOMAINS (            \
2230         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_IO) |           \
2231         BIT_ULL(POWER_DOMAIN_INIT))
2232 #define CNL_DISPLAY_DDI_C_IO_POWER_DOMAINS (            \
2233         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_IO) |           \
2234         BIT_ULL(POWER_DOMAIN_INIT))
2235 #define CNL_DISPLAY_DDI_D_IO_POWER_DOMAINS (            \
2236         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_IO) |           \
2237         BIT_ULL(POWER_DOMAIN_INIT))
2238 #define CNL_DISPLAY_AUX_A_POWER_DOMAINS (               \
2239         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2240         BIT_ULL(POWER_DOMAIN_AUX_IO_A) |                \
2241         BIT_ULL(POWER_DOMAIN_INIT))
2242 #define CNL_DISPLAY_AUX_B_POWER_DOMAINS (               \
2243         BIT_ULL(POWER_DOMAIN_AUX_B) |                   \
2244         BIT_ULL(POWER_DOMAIN_INIT))
2245 #define CNL_DISPLAY_AUX_C_POWER_DOMAINS (               \
2246         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2247         BIT_ULL(POWER_DOMAIN_INIT))
2248 #define CNL_DISPLAY_AUX_D_POWER_DOMAINS (               \
2249         BIT_ULL(POWER_DOMAIN_AUX_D) |                   \
2250         BIT_ULL(POWER_DOMAIN_INIT))
2251 #define CNL_DISPLAY_AUX_F_POWER_DOMAINS (               \
2252         BIT_ULL(POWER_DOMAIN_AUX_F) |                   \
2253         BIT_ULL(POWER_DOMAIN_INIT))
2254 #define CNL_DISPLAY_DDI_F_IO_POWER_DOMAINS (            \
2255         BIT_ULL(POWER_DOMAIN_PORT_DDI_F_IO) |           \
2256         BIT_ULL(POWER_DOMAIN_INIT))
2257 #define CNL_DISPLAY_DC_OFF_POWER_DOMAINS (              \
2258         CNL_DISPLAY_POWERWELL_2_POWER_DOMAINS |         \
2259         BIT_ULL(POWER_DOMAIN_GT_IRQ) |                  \
2260         BIT_ULL(POWER_DOMAIN_MODESET) |                 \
2261         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2262         BIT_ULL(POWER_DOMAIN_INIT))
2263
2264 /*
2265  * ICL PW_0/PG_0 domains (HW/DMC control):
2266  * - PCI
2267  * - clocks except port PLL
2268  * - central power except FBC
2269  * - shared functions except pipe interrupts, pipe MBUS, DBUF registers
2270  * ICL PW_1/PG_1 domains (HW/DMC control):
2271  * - DBUF function
2272  * - PIPE_A and its planes, except VGA
2273  * - transcoder EDP + PSR
2274  * - transcoder DSI
2275  * - DDI_A
2276  * - FBC
2277  */
2278 #define ICL_PW_4_POWER_DOMAINS (                        \
2279         BIT_ULL(POWER_DOMAIN_PIPE_C) |                  \
2280         BIT_ULL(POWER_DOMAIN_PIPE_C_PANEL_FITTER) |     \
2281         BIT_ULL(POWER_DOMAIN_INIT))
2282         /* VDSC/joining */
2283 #define ICL_PW_3_POWER_DOMAINS (                        \
2284         ICL_PW_4_POWER_DOMAINS |                        \
2285         BIT_ULL(POWER_DOMAIN_PIPE_B) |                  \
2286         BIT_ULL(POWER_DOMAIN_TRANSCODER_A) |            \
2287         BIT_ULL(POWER_DOMAIN_TRANSCODER_B) |            \
2288         BIT_ULL(POWER_DOMAIN_TRANSCODER_C) |            \
2289         BIT_ULL(POWER_DOMAIN_PIPE_B_PANEL_FITTER) |     \
2290         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_LANES) |        \
2291         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_IO) |           \
2292         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_LANES) |        \
2293         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_IO) |           \
2294         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_LANES) |        \
2295         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_IO) |           \
2296         BIT_ULL(POWER_DOMAIN_PORT_DDI_E_LANES) |        \
2297         BIT_ULL(POWER_DOMAIN_PORT_DDI_E_IO) |           \
2298         BIT_ULL(POWER_DOMAIN_PORT_DDI_F_LANES) |        \
2299         BIT_ULL(POWER_DOMAIN_PORT_DDI_F_IO) |           \
2300         BIT_ULL(POWER_DOMAIN_AUX_B) |                   \
2301         BIT_ULL(POWER_DOMAIN_AUX_C) |                   \
2302         BIT_ULL(POWER_DOMAIN_AUX_D) |                   \
2303         BIT_ULL(POWER_DOMAIN_AUX_E) |                   \
2304         BIT_ULL(POWER_DOMAIN_AUX_F) |                   \
2305         BIT_ULL(POWER_DOMAIN_AUX_TBT1) |                \
2306         BIT_ULL(POWER_DOMAIN_AUX_TBT2) |                \
2307         BIT_ULL(POWER_DOMAIN_AUX_TBT3) |                \
2308         BIT_ULL(POWER_DOMAIN_AUX_TBT4) |                \
2309         BIT_ULL(POWER_DOMAIN_VGA) |                     \
2310         BIT_ULL(POWER_DOMAIN_AUDIO) |                   \
2311         BIT_ULL(POWER_DOMAIN_INIT))
2312         /*
2313          * - transcoder WD
2314          * - KVMR (HW control)
2315          */
2316 #define ICL_PW_2_POWER_DOMAINS (                        \
2317         ICL_PW_3_POWER_DOMAINS |                        \
2318         BIT_ULL(POWER_DOMAIN_TRANSCODER_EDP_VDSC) |             \
2319         BIT_ULL(POWER_DOMAIN_INIT))
2320         /*
2321          * - KVMR (HW control)
2322          */
2323 #define ICL_DISPLAY_DC_OFF_POWER_DOMAINS (              \
2324         ICL_PW_2_POWER_DOMAINS |                        \
2325         BIT_ULL(POWER_DOMAIN_MODESET) |                 \
2326         BIT_ULL(POWER_DOMAIN_AUX_A) |                   \
2327         BIT_ULL(POWER_DOMAIN_INIT))
2328
2329 #define ICL_DDI_IO_A_POWER_DOMAINS (                    \
2330         BIT_ULL(POWER_DOMAIN_PORT_DDI_A_IO))
2331 #define ICL_DDI_IO_B_POWER_DOMAINS (                    \
2332         BIT_ULL(POWER_DOMAIN_PORT_DDI_B_IO))
2333 #define ICL_DDI_IO_C_POWER_DOMAINS (                    \
2334         BIT_ULL(POWER_DOMAIN_PORT_DDI_C_IO))
2335 #define ICL_DDI_IO_D_POWER_DOMAINS (                    \
2336         BIT_ULL(POWER_DOMAIN_PORT_DDI_D_IO))
2337 #define ICL_DDI_IO_E_POWER_DOMAINS (                    \
2338         BIT_ULL(POWER_DOMAIN_PORT_DDI_E_IO))
2339 #define ICL_DDI_IO_F_POWER_DOMAINS (                    \
2340         BIT_ULL(POWER_DOMAIN_PORT_DDI_F_IO))
2341
2342 #define ICL_AUX_A_IO_POWER_DOMAINS (                    \
2343         BIT_ULL(POWER_DOMAIN_AUX_IO_A) |                \
2344         BIT_ULL(POWER_DOMAIN_AUX_A))
2345 #define ICL_AUX_B_IO_POWER_DOMAINS (                    \
2346         BIT_ULL(POWER_DOMAIN_AUX_B))
2347 #define ICL_AUX_C_IO_POWER_DOMAINS (                    \
2348         BIT_ULL(POWER_DOMAIN_AUX_C))
2349 #define ICL_AUX_D_IO_POWER_DOMAINS (                    \
2350         BIT_ULL(POWER_DOMAIN_AUX_D))
2351 #define ICL_AUX_E_IO_POWER_DOMAINS (                    \
2352         BIT_ULL(POWER_DOMAIN_AUX_E))
2353 #define ICL_AUX_F_IO_POWER_DOMAINS (                    \
2354         BIT_ULL(POWER_DOMAIN_AUX_F))
2355 #define ICL_AUX_TBT1_IO_POWER_DOMAINS (                 \
2356         BIT_ULL(POWER_DOMAIN_AUX_TBT1))
2357 #define ICL_AUX_TBT2_IO_POWER_DOMAINS (                 \
2358         BIT_ULL(POWER_DOMAIN_AUX_TBT2))
2359 #define ICL_AUX_TBT3_IO_POWER_DOMAINS (                 \
2360         BIT_ULL(POWER_DOMAIN_AUX_TBT3))
2361 #define ICL_AUX_TBT4_IO_POWER_DOMAINS (                 \
2362         BIT_ULL(POWER_DOMAIN_AUX_TBT4))
2363
2364 static const struct i915_power_well_ops i9xx_always_on_power_well_ops = {
2365         .sync_hw = i9xx_power_well_sync_hw_noop,
2366         .enable = i9xx_always_on_power_well_noop,
2367         .disable = i9xx_always_on_power_well_noop,
2368         .is_enabled = i9xx_always_on_power_well_enabled,
2369 };
2370
2371 static const struct i915_power_well_ops chv_pipe_power_well_ops = {
2372         .sync_hw = i9xx_power_well_sync_hw_noop,
2373         .enable = chv_pipe_power_well_enable,
2374         .disable = chv_pipe_power_well_disable,
2375         .is_enabled = chv_pipe_power_well_enabled,
2376 };
2377
2378 static const struct i915_power_well_ops chv_dpio_cmn_power_well_ops = {
2379         .sync_hw = i9xx_power_well_sync_hw_noop,
2380         .enable = chv_dpio_cmn_power_well_enable,
2381         .disable = chv_dpio_cmn_power_well_disable,
2382         .is_enabled = vlv_power_well_enabled,
2383 };
2384
2385 static const struct i915_power_well_desc i9xx_always_on_power_well[] = {
2386         {
2387                 .name = "always-on",
2388                 .always_on = true,
2389                 .domains = POWER_DOMAIN_MASK,
2390                 .ops = &i9xx_always_on_power_well_ops,
2391                 .id = DISP_PW_ID_NONE,
2392         },
2393 };
2394
2395 static const struct i915_power_well_ops i830_pipes_power_well_ops = {
2396         .sync_hw = i830_pipes_power_well_sync_hw,
2397         .enable = i830_pipes_power_well_enable,
2398         .disable = i830_pipes_power_well_disable,
2399         .is_enabled = i830_pipes_power_well_enabled,
2400 };
2401
2402 static const struct i915_power_well_desc i830_power_wells[] = {
2403         {
2404                 .name = "always-on",
2405                 .always_on = true,
2406                 .domains = POWER_DOMAIN_MASK,
2407                 .ops = &i9xx_always_on_power_well_ops,
2408                 .id = DISP_PW_ID_NONE,
2409         },
2410         {
2411                 .name = "pipes",
2412                 .domains = I830_PIPES_POWER_DOMAINS,
2413                 .ops = &i830_pipes_power_well_ops,
2414                 .id = DISP_PW_ID_NONE,
2415         },
2416 };
2417
2418 static const struct i915_power_well_ops hsw_power_well_ops = {
2419         .sync_hw = hsw_power_well_sync_hw,
2420         .enable = hsw_power_well_enable,
2421         .disable = hsw_power_well_disable,
2422         .is_enabled = hsw_power_well_enabled,
2423 };
2424
2425 static const struct i915_power_well_ops gen9_dc_off_power_well_ops = {
2426         .sync_hw = i9xx_power_well_sync_hw_noop,
2427         .enable = gen9_dc_off_power_well_enable,
2428         .disable = gen9_dc_off_power_well_disable,
2429         .is_enabled = gen9_dc_off_power_well_enabled,
2430 };
2431
2432 static const struct i915_power_well_ops bxt_dpio_cmn_power_well_ops = {
2433         .sync_hw = i9xx_power_well_sync_hw_noop,
2434         .enable = bxt_dpio_cmn_power_well_enable,
2435         .disable = bxt_dpio_cmn_power_well_disable,
2436         .is_enabled = bxt_dpio_cmn_power_well_enabled,
2437 };
2438
2439 static const struct i915_power_well_regs hsw_power_well_regs = {
2440         .bios   = HSW_PWR_WELL_CTL1,
2441         .driver = HSW_PWR_WELL_CTL2,
2442         .kvmr   = HSW_PWR_WELL_CTL3,
2443         .debug  = HSW_PWR_WELL_CTL4,
2444 };
2445
2446 static const struct i915_power_well_desc hsw_power_wells[] = {
2447         {
2448                 .name = "always-on",
2449                 .always_on = true,
2450                 .domains = POWER_DOMAIN_MASK,
2451                 .ops = &i9xx_always_on_power_well_ops,
2452                 .id = DISP_PW_ID_NONE,
2453         },
2454         {
2455                 .name = "display",
2456                 .domains = HSW_DISPLAY_POWER_DOMAINS,
2457                 .ops = &hsw_power_well_ops,
2458                 .id = HSW_DISP_PW_GLOBAL,
2459                 {
2460                         .hsw.regs = &hsw_power_well_regs,
2461                         .hsw.idx = HSW_PW_CTL_IDX_GLOBAL,
2462                         .hsw.has_vga = true,
2463                 },
2464         },
2465 };
2466
2467 static const struct i915_power_well_desc bdw_power_wells[] = {
2468         {
2469                 .name = "always-on",
2470                 .always_on = true,
2471                 .domains = POWER_DOMAIN_MASK,
2472                 .ops = &i9xx_always_on_power_well_ops,
2473                 .id = DISP_PW_ID_NONE,
2474         },
2475         {
2476                 .name = "display",
2477                 .domains = BDW_DISPLAY_POWER_DOMAINS,
2478                 .ops = &hsw_power_well_ops,
2479                 .id = HSW_DISP_PW_GLOBAL,
2480                 {
2481                         .hsw.regs = &hsw_power_well_regs,
2482                         .hsw.idx = HSW_PW_CTL_IDX_GLOBAL,
2483                         .hsw.irq_pipe_mask = BIT(PIPE_B) | BIT(PIPE_C),
2484                         .hsw.has_vga = true,
2485                 },
2486         },
2487 };
2488
2489 static const struct i915_power_well_ops vlv_display_power_well_ops = {
2490         .sync_hw = i9xx_power_well_sync_hw_noop,
2491         .enable = vlv_display_power_well_enable,
2492         .disable = vlv_display_power_well_disable,
2493         .is_enabled = vlv_power_well_enabled,
2494 };
2495
2496 static const struct i915_power_well_ops vlv_dpio_cmn_power_well_ops = {
2497         .sync_hw = i9xx_power_well_sync_hw_noop,
2498         .enable = vlv_dpio_cmn_power_well_enable,
2499         .disable = vlv_dpio_cmn_power_well_disable,
2500         .is_enabled = vlv_power_well_enabled,
2501 };
2502
2503 static const struct i915_power_well_ops vlv_dpio_power_well_ops = {
2504         .sync_hw = i9xx_power_well_sync_hw_noop,
2505         .enable = vlv_power_well_enable,
2506         .disable = vlv_power_well_disable,
2507         .is_enabled = vlv_power_well_enabled,
2508 };
2509
2510 static const struct i915_power_well_desc vlv_power_wells[] = {
2511         {
2512                 .name = "always-on",
2513                 .always_on = true,
2514                 .domains = POWER_DOMAIN_MASK,
2515                 .ops = &i9xx_always_on_power_well_ops,
2516                 .id = DISP_PW_ID_NONE,
2517         },
2518         {
2519                 .name = "display",
2520                 .domains = VLV_DISPLAY_POWER_DOMAINS,
2521                 .ops = &vlv_display_power_well_ops,
2522                 .id = VLV_DISP_PW_DISP2D,
2523                 {
2524                         .vlv.idx = PUNIT_PWGT_IDX_DISP2D,
2525                 },
2526         },
2527         {
2528                 .name = "dpio-tx-b-01",
2529                 .domains = VLV_DPIO_TX_B_LANES_01_POWER_DOMAINS |
2530                            VLV_DPIO_TX_B_LANES_23_POWER_DOMAINS |
2531                            VLV_DPIO_TX_C_LANES_01_POWER_DOMAINS |
2532                            VLV_DPIO_TX_C_LANES_23_POWER_DOMAINS,
2533                 .ops = &vlv_dpio_power_well_ops,
2534                 .id = DISP_PW_ID_NONE,
2535                 {
2536                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_B_LANES_01,
2537                 },
2538         },
2539         {
2540                 .name = "dpio-tx-b-23",
2541                 .domains = VLV_DPIO_TX_B_LANES_01_POWER_DOMAINS |
2542                            VLV_DPIO_TX_B_LANES_23_POWER_DOMAINS |
2543                            VLV_DPIO_TX_C_LANES_01_POWER_DOMAINS |
2544                            VLV_DPIO_TX_C_LANES_23_POWER_DOMAINS,
2545                 .ops = &vlv_dpio_power_well_ops,
2546                 .id = DISP_PW_ID_NONE,
2547                 {
2548                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_B_LANES_23,
2549                 },
2550         },
2551         {
2552                 .name = "dpio-tx-c-01",
2553                 .domains = VLV_DPIO_TX_B_LANES_01_POWER_DOMAINS |
2554                            VLV_DPIO_TX_B_LANES_23_POWER_DOMAINS |
2555                            VLV_DPIO_TX_C_LANES_01_POWER_DOMAINS |
2556                            VLV_DPIO_TX_C_LANES_23_POWER_DOMAINS,
2557                 .ops = &vlv_dpio_power_well_ops,
2558                 .id = DISP_PW_ID_NONE,
2559                 {
2560                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_C_LANES_01,
2561                 },
2562         },
2563         {
2564                 .name = "dpio-tx-c-23",
2565                 .domains = VLV_DPIO_TX_B_LANES_01_POWER_DOMAINS |
2566                            VLV_DPIO_TX_B_LANES_23_POWER_DOMAINS |
2567                            VLV_DPIO_TX_C_LANES_01_POWER_DOMAINS |
2568                            VLV_DPIO_TX_C_LANES_23_POWER_DOMAINS,
2569                 .ops = &vlv_dpio_power_well_ops,
2570                 .id = DISP_PW_ID_NONE,
2571                 {
2572                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_C_LANES_23,
2573                 },
2574         },
2575         {
2576                 .name = "dpio-common",
2577                 .domains = VLV_DPIO_CMN_BC_POWER_DOMAINS,
2578                 .ops = &vlv_dpio_cmn_power_well_ops,
2579                 .id = VLV_DISP_PW_DPIO_CMN_BC,
2580                 {
2581                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_CMN_BC,
2582                 },
2583         },
2584 };
2585
2586 static const struct i915_power_well_desc chv_power_wells[] = {
2587         {
2588                 .name = "always-on",
2589                 .always_on = true,
2590                 .domains = POWER_DOMAIN_MASK,
2591                 .ops = &i9xx_always_on_power_well_ops,
2592                 .id = DISP_PW_ID_NONE,
2593         },
2594         {
2595                 .name = "display",
2596                 /*
2597                  * Pipe A power well is the new disp2d well. Pipe B and C
2598                  * power wells don't actually exist. Pipe A power well is
2599                  * required for any pipe to work.
2600                  */
2601                 .domains = CHV_DISPLAY_POWER_DOMAINS,
2602                 .ops = &chv_pipe_power_well_ops,
2603                 .id = DISP_PW_ID_NONE,
2604         },
2605         {
2606                 .name = "dpio-common-bc",
2607                 .domains = CHV_DPIO_CMN_BC_POWER_DOMAINS,
2608                 .ops = &chv_dpio_cmn_power_well_ops,
2609                 .id = VLV_DISP_PW_DPIO_CMN_BC,
2610                 {
2611                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_CMN_BC,
2612                 },
2613         },
2614         {
2615                 .name = "dpio-common-d",
2616                 .domains = CHV_DPIO_CMN_D_POWER_DOMAINS,
2617                 .ops = &chv_dpio_cmn_power_well_ops,
2618                 .id = CHV_DISP_PW_DPIO_CMN_D,
2619                 {
2620                         .vlv.idx = PUNIT_PWGT_IDX_DPIO_CMN_D,
2621                 },
2622         },
2623 };
2624
2625 bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv,
2626                                          enum i915_power_well_id power_well_id)
2627 {
2628         struct i915_power_well *power_well;
2629         bool ret;
2630
2631         power_well = lookup_power_well(dev_priv, power_well_id);
2632         ret = power_well->desc->ops->is_enabled(dev_priv, power_well);
2633
2634         return ret;
2635 }
2636
2637 static const struct i915_power_well_desc skl_power_wells[] = {
2638         {
2639                 .name = "always-on",
2640                 .always_on = true,
2641                 .domains = POWER_DOMAIN_MASK,
2642                 .ops = &i9xx_always_on_power_well_ops,
2643                 .id = DISP_PW_ID_NONE,
2644         },
2645         {
2646                 .name = "power well 1",
2647                 /* Handled by the DMC firmware */
2648                 .always_on = true,
2649                 .domains = 0,
2650                 .ops = &hsw_power_well_ops,
2651                 .id = SKL_DISP_PW_1,
2652                 {
2653                         .hsw.regs = &hsw_power_well_regs,
2654                         .hsw.idx = SKL_PW_CTL_IDX_PW_1,
2655                         .hsw.has_fuses = true,
2656                 },
2657         },
2658         {
2659                 .name = "MISC IO power well",
2660                 /* Handled by the DMC firmware */
2661                 .always_on = true,
2662                 .domains = 0,
2663                 .ops = &hsw_power_well_ops,
2664                 .id = SKL_DISP_PW_MISC_IO,
2665                 {
2666                         .hsw.regs = &hsw_power_well_regs,
2667                         .hsw.idx = SKL_PW_CTL_IDX_MISC_IO,
2668                 },
2669         },
2670         {
2671                 .name = "DC off",
2672                 .domains = SKL_DISPLAY_DC_OFF_POWER_DOMAINS,
2673                 .ops = &gen9_dc_off_power_well_ops,
2674                 .id = DISP_PW_ID_NONE,
2675         },
2676         {
2677                 .name = "power well 2",
2678                 .domains = SKL_DISPLAY_POWERWELL_2_POWER_DOMAINS,
2679                 .ops = &hsw_power_well_ops,
2680                 .id = SKL_DISP_PW_2,
2681                 {
2682                         .hsw.regs = &hsw_power_well_regs,
2683                         .hsw.idx = SKL_PW_CTL_IDX_PW_2,
2684                         .hsw.irq_pipe_mask = BIT(PIPE_B) | BIT(PIPE_C),
2685                         .hsw.has_vga = true,
2686                         .hsw.has_fuses = true,
2687                 },
2688         },
2689         {
2690                 .name = "DDI A/E IO power well",
2691                 .domains = SKL_DISPLAY_DDI_IO_A_E_POWER_DOMAINS,
2692                 .ops = &hsw_power_well_ops,
2693                 .id = DISP_PW_ID_NONE,
2694                 {
2695                         .hsw.regs = &hsw_power_well_regs,
2696                         .hsw.idx = SKL_PW_CTL_IDX_DDI_A_E,
2697                 },
2698         },
2699         {
2700                 .name = "DDI B IO power well",
2701                 .domains = SKL_DISPLAY_DDI_IO_B_POWER_DOMAINS,
2702                 .ops = &hsw_power_well_ops,
2703                 .id = DISP_PW_ID_NONE,
2704                 {
2705                         .hsw.regs = &hsw_power_well_regs,
2706                         .hsw.idx = SKL_PW_CTL_IDX_DDI_B,
2707                 },
2708         },
2709         {
2710                 .name = "DDI C IO power well",
2711                 .domains = SKL_DISPLAY_DDI_IO_C_POWER_DOMAINS,
2712                 .ops = &hsw_power_well_ops,
2713                 .id = DISP_PW_ID_NONE,
2714                 {
2715                         .hsw.regs = &hsw_power_well_regs,
2716                         .hsw.idx = SKL_PW_CTL_IDX_DDI_C,
2717                 },
2718         },
2719         {
2720                 .name = "DDI D IO power well",
2721                 .domains = SKL_DISPLAY_DDI_IO_D_POWER_DOMAINS,
2722                 .ops = &hsw_power_well_ops,
2723                 .id = DISP_PW_ID_NONE,
2724                 {
2725                         .hsw.regs = &hsw_power_well_regs,
2726                         .hsw.idx = SKL_PW_CTL_IDX_DDI_D,
2727                 },
2728         },
2729 };
2730
2731 static const struct i915_power_well_desc bxt_power_wells[] = {
2732         {
2733                 .name = "always-on",
2734                 .always_on = true,
2735                 .domains = POWER_DOMAIN_MASK,
2736                 .ops = &i9xx_always_on_power_well_ops,
2737                 .id = DISP_PW_ID_NONE,
2738         },
2739         {
2740                 .name = "power well 1",
2741                 /* Handled by the DMC firmware */
2742                 .always_on = true,
2743                 .domains = 0,
2744                 .ops = &hsw_power_well_ops,
2745                 .id = SKL_DISP_PW_1,
2746                 {
2747                         .hsw.regs = &hsw_power_well_regs,
2748                         .hsw.idx = SKL_PW_CTL_IDX_PW_1,
2749                         .hsw.has_fuses = true,
2750                 },
2751         },
2752         {
2753                 .name = "DC off",
2754                 .domains = BXT_DISPLAY_DC_OFF_POWER_DOMAINS,
2755                 .ops = &gen9_dc_off_power_well_ops,
2756                 .id = DISP_PW_ID_NONE,
2757         },
2758         {
2759                 .name = "power well 2",
2760                 .domains = BXT_DISPLAY_POWERWELL_2_POWER_DOMAINS,
2761                 .ops = &hsw_power_well_ops,
2762                 .id = SKL_DISP_PW_2,
2763                 {
2764                         .hsw.regs = &hsw_power_well_regs,
2765                         .hsw.idx = SKL_PW_CTL_IDX_PW_2,
2766                         .hsw.irq_pipe_mask = BIT(PIPE_B) | BIT(PIPE_C),
2767                         .hsw.has_vga = true,
2768                         .hsw.has_fuses = true,
2769                 },
2770         },
2771         {
2772                 .name = "dpio-common-a",
2773                 .domains = BXT_DPIO_CMN_A_POWER_DOMAINS,
2774                 .ops = &bxt_dpio_cmn_power_well_ops,
2775                 .id = BXT_DISP_PW_DPIO_CMN_A,
2776                 {
2777                         .bxt.phy = DPIO_PHY1,
2778                 },
2779         },
2780         {
2781                 .name = "dpio-common-bc",
2782                 .domains = BXT_DPIO_CMN_BC_POWER_DOMAINS,
2783                 .ops = &bxt_dpio_cmn_power_well_ops,
2784                 .id = VLV_DISP_PW_DPIO_CMN_BC,
2785                 {
2786                         .bxt.phy = DPIO_PHY0,
2787                 },
2788         },
2789 };
2790
2791 static const struct i915_power_well_desc glk_power_wells[] = {
2792         {
2793                 .name = "always-on",
2794                 .always_on = true,
2795                 .domains = POWER_DOMAIN_MASK,
2796                 .ops = &i9xx_always_on_power_well_ops,
2797                 .id = DISP_PW_ID_NONE,
2798         },
2799         {
2800                 .name = "power well 1",
2801                 /* Handled by the DMC firmware */
2802                 .always_on = true,
2803                 .domains = 0,
2804                 .ops = &hsw_power_well_ops,
2805                 .id = SKL_DISP_PW_1,
2806                 {
2807                         .hsw.regs = &hsw_power_well_regs,
2808                         .hsw.idx = SKL_PW_CTL_IDX_PW_1,
2809                         .hsw.has_fuses = true,
2810                 },
2811         },
2812         {
2813                 .name = "DC off",
2814                 .domains = GLK_DISPLAY_DC_OFF_POWER_DOMAINS,
2815                 .ops = &gen9_dc_off_power_well_ops,
2816                 .id = DISP_PW_ID_NONE,
2817         },
2818         {
2819                 .name = "power well 2",
2820                 .domains = GLK_DISPLAY_POWERWELL_2_POWER_DOMAINS,
2821                 .ops = &hsw_power_well_ops,
2822                 .id = SKL_DISP_PW_2,
2823                 {
2824                         .hsw.regs = &hsw_power_well_regs,
2825                         .hsw.idx = SKL_PW_CTL_IDX_PW_2,
2826                         .hsw.irq_pipe_mask = BIT(PIPE_B) | BIT(PIPE_C),
2827                         .hsw.has_vga = true,
2828                         .hsw.has_fuses = true,
2829                 },
2830         },
2831         {
2832                 .name = "dpio-common-a",
2833                 .domains = GLK_DPIO_CMN_A_POWER_DOMAINS,
2834                 .ops = &bxt_dpio_cmn_power_well_ops,
2835                 .id = BXT_DISP_PW_DPIO_CMN_A,
2836                 {
2837                         .bxt.phy = DPIO_PHY1,
2838                 },
2839         },
2840         {
2841                 .name = "dpio-common-b",
2842                 .domains = GLK_DPIO_CMN_B_POWER_DOMAINS,
2843                 .ops = &bxt_dpio_cmn_power_well_ops,
2844                 .id = VLV_DISP_PW_DPIO_CMN_BC,
2845                 {
2846                         .bxt.phy = DPIO_PHY0,
2847                 },
2848         },
2849         {
2850                 .name = "dpio-common-c",
2851                 .domains = GLK_DPIO_CMN_C_POWER_DOMAINS,
2852                 .ops = &bxt_dpio_cmn_power_well_ops,
2853                 .id = GLK_DISP_PW_DPIO_CMN_C,
2854                 {
2855                         .bxt.phy = DPIO_PHY2,
2856                 },
2857         },
2858         {
2859                 .name = "AUX A",
2860                 .domains = GLK_DISPLAY_AUX_A_POWER_DOMAINS,
2861                 .ops = &hsw_power_well_ops,
2862                 .id = DISP_PW_ID_NONE,
2863                 {
2864                         .hsw.regs = &hsw_power_well_regs,
2865                         .hsw.idx = GLK_PW_CTL_IDX_AUX_A,
2866                 },
2867         },
2868         {
2869                 .name = "AUX B",
2870                 .domains = GLK_DISPLAY_AUX_B_POWER_DOMAINS,
2871                 .ops = &hsw_power_well_ops,
2872                 .id = DISP_PW_ID_NONE,
2873                 {
2874                         .hsw.regs = &hsw_power_well_regs,
2875                         .hsw.idx = GLK_PW_CTL_IDX_AUX_B,
2876                 },
2877         },
2878         {
2879                 .name = "AUX C",
2880                 .domains = GLK_DISPLAY_AUX_C_POWER_DOMAINS,
2881                 .ops = &hsw_power_well_ops,
2882                 .id = DISP_PW_ID_NONE,
2883                 {
2884                         .hsw.regs = &hsw_power_well_regs,
2885                         .hsw.idx = GLK_PW_CTL_IDX_AUX_C,
2886                 },
2887         },
2888         {
2889                 .name = "DDI A IO power well",
2890                 .domains = GLK_DISPLAY_DDI_IO_A_POWER_DOMAINS,
2891                 .ops = &hsw_power_well_ops,
2892                 .id = DISP_PW_ID_NONE,
2893                 {
2894                         .hsw.regs = &hsw_power_well_regs,
2895                         .hsw.idx = GLK_PW_CTL_IDX_DDI_A,
2896                 },
2897         },
2898         {
2899                 .name = "DDI B IO power well",
2900                 .domains = GLK_DISPLAY_DDI_IO_B_POWER_DOMAINS,
2901                 .ops = &hsw_power_well_ops,
2902                 .id = DISP_PW_ID_NONE,
2903                 {
2904                         .hsw.regs = &hsw_power_well_regs,
2905                         .hsw.idx = SKL_PW_CTL_IDX_DDI_B,
2906                 },
2907         },
2908         {
2909                 .name = "DDI C IO power well",
2910                 .domains = GLK_DISPLAY_DDI_IO_C_POWER_DOMAINS,
2911                 .ops = &hsw_power_well_ops,
2912                 .id = DISP_PW_ID_NONE,
2913                 {
2914                         .hsw.regs = &hsw_power_well_regs,
2915                         .hsw.idx = SKL_PW_CTL_IDX_DDI_C,
2916                 },
2917         },
2918 };
2919
2920 static const struct i915_power_well_desc cnl_power_wells[] = {
2921         {
2922                 .name = "always-on",
2923                 .always_on = true,
2924                 .domains = POWER_DOMAIN_MASK,
2925                 .ops = &i9xx_always_on_power_well_ops,
2926                 .id = DISP_PW_ID_NONE,
2927         },
2928         {
2929                 .name = "power well 1",
2930                 /* Handled by the DMC firmware */
2931                 .always_on = true,
2932                 .domains = 0,
2933                 .ops = &hsw_power_well_ops,
2934                 .id = SKL_DISP_PW_1,
2935                 {
2936                         .hsw.regs = &hsw_power_well_regs,
2937                         .hsw.idx = SKL_PW_CTL_IDX_PW_1,
2938                         .hsw.has_fuses = true,
2939                 },
2940         },
2941         {
2942                 .name = "AUX A",
2943                 .domains = CNL_DISPLAY_AUX_A_POWER_DOMAINS,
2944                 .ops = &hsw_power_well_ops,
2945                 .id = DISP_PW_ID_NONE,
2946                 {
2947                         .hsw.regs = &hsw_power_well_regs,
2948                         .hsw.idx = GLK_PW_CTL_IDX_AUX_A,
2949                 },
2950         },
2951         {
2952                 .name = "AUX B",
2953                 .domains = CNL_DISPLAY_AUX_B_POWER_DOMAINS,
2954                 .ops = &hsw_power_well_ops,
2955                 .id = DISP_PW_ID_NONE,
2956                 {
2957                         .hsw.regs = &hsw_power_well_regs,
2958                         .hsw.idx = GLK_PW_CTL_IDX_AUX_B,
2959                 },
2960         },
2961         {
2962                 .name = "AUX C",
2963                 .domains = CNL_DISPLAY_AUX_C_POWER_DOMAINS,
2964                 .ops = &hsw_power_well_ops,
2965                 .id = DISP_PW_ID_NONE,
2966                 {
2967                         .hsw.regs = &hsw_power_well_regs,
2968                         .hsw.idx = GLK_PW_CTL_IDX_AUX_C,
2969                 },
2970         },
2971         {
2972                 .name = "AUX D",
2973                 .domains = CNL_DISPLAY_AUX_D_POWER_DOMAINS,
2974                 .ops = &hsw_power_well_ops,
2975                 .id = DISP_PW_ID_NONE,
2976                 {
2977                         .hsw.regs = &hsw_power_well_regs,
2978                         .hsw.idx = CNL_PW_CTL_IDX_AUX_D,
2979                 },
2980         },
2981         {
2982                 .name = "DC off",
2983                 .domains = CNL_DISPLAY_DC_OFF_POWER_DOMAINS,
2984                 .ops = &gen9_dc_off_power_well_ops,
2985                 .id = DISP_PW_ID_NONE,
2986         },
2987         {
2988                 .name = "power well 2",
2989                 .domains = CNL_DISPLAY_POWERWELL_2_POWER_DOMAINS,
2990                 .ops = &hsw_power_well_ops,
2991                 .id = SKL_DISP_PW_2,
2992                 {
2993                         .hsw.regs = &hsw_power_well_regs,
2994                         .hsw.idx = SKL_PW_CTL_IDX_PW_2,
2995                         .hsw.irq_pipe_mask = BIT(PIPE_B) | BIT(PIPE_C),
2996                         .hsw.has_vga = true,
2997                         .hsw.has_fuses = true,
2998                 },
2999         },
3000         {
3001                 .name = "DDI A IO power well",
3002                 .domains = CNL_DISPLAY_DDI_A_IO_POWER_DOMAINS,
3003                 .ops = &hsw_power_well_ops,
3004                 .id = DISP_PW_ID_NONE,
3005                 {
3006                         .hsw.regs = &hsw_power_well_regs,
3007                         .hsw.idx = GLK_PW_CTL_IDX_DDI_A,
3008                 },
3009         },
3010         {
3011                 .name = "DDI B IO power well",
3012                 .domains = CNL_DISPLAY_DDI_B_IO_POWER_DOMAINS,
3013                 .ops = &hsw_power_well_ops,
3014                 .id = DISP_PW_ID_NONE,
3015                 {
3016                         .hsw.regs = &hsw_power_well_regs,
3017                         .hsw.idx = SKL_PW_CTL_IDX_DDI_B,
3018                 },
3019         },
3020         {
3021                 .name = "DDI C IO power well",
3022                 .domains = CNL_DISPLAY_DDI_C_IO_POWER_DOMAINS,
3023                 .ops = &hsw_power_well_ops,
3024                 .id = DISP_PW_ID_NONE,
3025                 {
3026                         .hsw.regs = &hsw_power_well_regs,
3027                         .hsw.idx = SKL_PW_CTL_IDX_DDI_C,
3028                 },
3029         },
3030         {
3031                 .name = "DDI D IO power well",
3032                 .domains = CNL_DISPLAY_DDI_D_IO_POWER_DOMAINS,
3033                 .ops = &hsw_power_well_ops,
3034                 .id = DISP_PW_ID_NONE,
3035                 {
3036                         .hsw.regs = &hsw_power_well_regs,
3037                         .hsw.idx = SKL_PW_CTL_IDX_DDI_D,
3038                 },
3039         },
3040         {
3041                 .name = "DDI F IO power well",
3042                 .domains = CNL_DISPLAY_DDI_F_IO_POWER_DOMAINS,
3043                 .ops = &hsw_power_well_ops,
3044                 .id = DISP_PW_ID_NONE,
3045                 {
3046                         .hsw.regs = &hsw_power_well_regs,
3047                         .hsw.idx = CNL_PW_CTL_IDX_DDI_F,
3048                 },
3049         },
3050         {
3051                 .name = "AUX F",
3052                 .domains = CNL_DISPLAY_AUX_F_POWER_DOMAINS,
3053                 .ops = &hsw_power_well_ops,
3054                 .id = DISP_PW_ID_NONE,
3055                 {
3056                         .hsw.regs = &hsw_power_well_regs,
3057                         .hsw.idx = CNL_PW_CTL_IDX_AUX_F,
3058                 },
3059         },
3060 };
3061
3062 static const struct i915_power_well_ops icl_combo_phy_aux_power_well_ops = {
3063         .sync_hw = hsw_power_well_sync_hw,
3064         .enable = icl_combo_phy_aux_power_well_enable,
3065         .disable = icl_combo_phy_aux_power_well_disable,
3066         .is_enabled = hsw_power_well_enabled,
3067 };
3068
3069 static const struct i915_power_well_ops icl_tc_phy_aux_power_well_ops = {
3070         .sync_hw = hsw_power_well_sync_hw,
3071         .enable = icl_tc_phy_aux_power_well_enable,
3072         .disable = hsw_power_well_disable,
3073         .is_enabled = hsw_power_well_enabled,
3074 };
3075
3076 static const struct i915_power_well_regs icl_aux_power_well_regs = {
3077         .bios   = ICL_PWR_WELL_CTL_AUX1,
3078         .driver = ICL_PWR_WELL_CTL_AUX2,
3079         .debug  = ICL_PWR_WELL_CTL_AUX4,
3080 };
3081
3082 static const struct i915_power_well_regs icl_ddi_power_well_regs = {
3083         .bios   = ICL_PWR_WELL_CTL_DDI1,
3084         .driver = ICL_PWR_WELL_CTL_DDI2,
3085         .debug  = ICL_PWR_WELL_CTL_DDI4,
3086 };
3087
3088 static const struct i915_power_well_desc icl_power_wells[] = {
3089         {
3090                 .name = "always-on",
3091                 .always_on = true,
3092                 .domains = POWER_DOMAIN_MASK,
3093                 .ops = &i9xx_always_on_power_well_ops,
3094                 .id = DISP_PW_ID_NONE,
3095         },
3096         {
3097                 .name = "power well 1",
3098                 /* Handled by the DMC firmware */
3099                 .always_on = true,
3100                 .domains = 0,
3101                 .ops = &hsw_power_well_ops,
3102                 .id = SKL_DISP_PW_1,
3103                 {
3104                         .hsw.regs = &hsw_power_well_regs,
3105                         .hsw.idx = ICL_PW_CTL_IDX_PW_1,
3106                         .hsw.has_fuses = true,
3107                 },
3108         },
3109         {
3110                 .name = "DC off",
3111                 .domains = ICL_DISPLAY_DC_OFF_POWER_DOMAINS,
3112                 .ops = &gen9_dc_off_power_well_ops,
3113                 .id = DISP_PW_ID_NONE,
3114         },
3115         {
3116                 .name = "power well 2",
3117                 .domains = ICL_PW_2_POWER_DOMAINS,
3118                 .ops = &hsw_power_well_ops,
3119                 .id = SKL_DISP_PW_2,
3120                 {
3121                         .hsw.regs = &hsw_power_well_regs,
3122                         .hsw.idx = ICL_PW_CTL_IDX_PW_2,
3123                         .hsw.has_fuses = true,
3124                 },
3125         },
3126         {
3127                 .name = "power well 3",
3128                 .domains = ICL_PW_3_POWER_DOMAINS,
3129                 .ops = &hsw_power_well_ops,
3130                 .id = DISP_PW_ID_NONE,
3131                 {
3132                         .hsw.regs = &hsw_power_well_regs,
3133                         .hsw.idx = ICL_PW_CTL_IDX_PW_3,
3134                         .hsw.irq_pipe_mask = BIT(PIPE_B),
3135                         .hsw.has_vga = true,
3136                         .hsw.has_fuses = true,
3137                 },
3138         },
3139         {
3140                 .name = "DDI A IO",
3141                 .domains = ICL_DDI_IO_A_POWER_DOMAINS,
3142                 .ops = &hsw_power_well_ops,
3143                 .id = DISP_PW_ID_NONE,
3144                 {
3145                         .hsw.regs = &icl_ddi_power_well_regs,
3146                         .hsw.idx = ICL_PW_CTL_IDX_DDI_A,
3147                 },
3148         },
3149         {
3150                 .name = "DDI B IO",
3151                 .domains = ICL_DDI_IO_B_POWER_DOMAINS,
3152                 .ops = &hsw_power_well_ops,
3153                 .id = DISP_PW_ID_NONE,
3154                 {
3155                         .hsw.regs = &icl_ddi_power_well_regs,
3156                         .hsw.idx = ICL_PW_CTL_IDX_DDI_B,
3157                 },
3158         },
3159         {
3160                 .name = "DDI C IO",
3161                 .domains = ICL_DDI_IO_C_POWER_DOMAINS,
3162                 .ops = &hsw_power_well_ops,
3163                 .id = DISP_PW_ID_NONE,
3164                 {
3165                         .hsw.regs = &icl_ddi_power_well_regs,
3166                         .hsw.idx = ICL_PW_CTL_IDX_DDI_C,
3167                 },
3168         },
3169         {
3170                 .name = "DDI D IO",
3171                 .domains = ICL_DDI_IO_D_POWER_DOMAINS,
3172                 .ops = &hsw_power_well_ops,
3173                 .id = DISP_PW_ID_NONE,
3174                 {
3175                         .hsw.regs = &icl_ddi_power_well_regs,
3176                         .hsw.idx = ICL_PW_CTL_IDX_DDI_D,
3177                 },
3178         },
3179         {
3180                 .name = "DDI E IO",
3181                 .domains = ICL_DDI_IO_E_POWER_DOMAINS,
3182                 .ops = &hsw_power_well_ops,
3183                 .id = DISP_PW_ID_NONE,
3184                 {
3185                         .hsw.regs = &icl_ddi_power_well_regs,
3186                         .hsw.idx = ICL_PW_CTL_IDX_DDI_E,
3187                 },
3188         },
3189         {
3190                 .name = "DDI F IO",
3191                 .domains = ICL_DDI_IO_F_POWER_DOMAINS,
3192                 .ops = &hsw_power_well_ops,
3193                 .id = DISP_PW_ID_NONE,
3194                 {
3195                         .hsw.regs = &icl_ddi_power_well_regs,
3196                         .hsw.idx = ICL_PW_CTL_IDX_DDI_F,
3197                 },
3198         },
3199         {
3200                 .name = "AUX A",
3201                 .domains = ICL_AUX_A_IO_POWER_DOMAINS,
3202                 .ops = &icl_combo_phy_aux_power_well_ops,
3203                 .id = DISP_PW_ID_NONE,
3204                 {
3205                         .hsw.regs = &icl_aux_power_well_regs,
3206                         .hsw.idx = ICL_PW_CTL_IDX_AUX_A,
3207                 },
3208         },
3209         {
3210                 .name = "AUX B",
3211                 .domains = ICL_AUX_B_IO_POWER_DOMAINS,
3212                 .ops = &icl_combo_phy_aux_power_well_ops,
3213                 .id = DISP_PW_ID_NONE,
3214                 {
3215                         .hsw.regs = &icl_aux_power_well_regs,
3216                         .hsw.idx = ICL_PW_CTL_IDX_AUX_B,
3217                 },
3218         },
3219         {
3220                 .name = "AUX C",
3221                 .domains = ICL_AUX_C_IO_POWER_DOMAINS,
3222                 .ops = &icl_tc_phy_aux_power_well_ops,
3223                 .id = DISP_PW_ID_NONE,
3224                 {
3225                         .hsw.regs = &icl_aux_power_well_regs,
3226                         .hsw.idx = ICL_PW_CTL_IDX_AUX_C,
3227                         .hsw.is_tc_tbt = false,
3228                 },
3229         },
3230         {
3231                 .name = "AUX D",
3232                 .domains = ICL_AUX_D_IO_POWER_DOMAINS,
3233                 .ops = &icl_tc_phy_aux_power_well_ops,
3234                 .id = DISP_PW_ID_NONE,
3235                 {
3236                         .hsw.regs = &icl_aux_power_well_regs,
3237                         .hsw.idx = ICL_PW_CTL_IDX_AUX_D,
3238                         .hsw.is_tc_tbt = false,
3239                 },
3240         },
3241         {
3242                 .name = "AUX E",
3243                 .domains = ICL_AUX_E_IO_POWER_DOMAINS,
3244                 .ops = &icl_tc_phy_aux_power_well_ops,
3245                 .id = DISP_PW_ID_NONE,
3246                 {
3247                         .hsw.regs = &icl_aux_power_well_regs,
3248                         .hsw.idx = ICL_PW_CTL_IDX_AUX_E,
3249                         .hsw.is_tc_tbt = false,
3250                 },
3251         },
3252         {
3253                 .name = "AUX F",
3254                 .domains = ICL_AUX_F_IO_POWER_DOMAINS,
3255                 .ops = &icl_tc_phy_aux_power_well_ops,
3256                 .id = DISP_PW_ID_NONE,
3257                 {
3258                         .hsw.regs = &icl_aux_power_well_regs,
3259                         .hsw.idx = ICL_PW_CTL_IDX_AUX_F,
3260                         .hsw.is_tc_tbt = false,
3261                 },
3262         },
3263         {
3264                 .name = "AUX TBT1",
3265                 .domains = ICL_AUX_TBT1_IO_POWER_DOMAINS,
3266                 .ops = &icl_tc_phy_aux_power_well_ops,
3267                 .id = DISP_PW_ID_NONE,
3268                 {
3269                         .hsw.regs = &icl_aux_power_well_regs,
3270                         .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT1,
3271                         .hsw.is_tc_tbt = true,
3272                 },
3273         },
3274         {
3275                 .name = "AUX TBT2",
3276                 .domains = ICL_AUX_TBT2_IO_POWER_DOMAINS,
3277                 .ops = &icl_tc_phy_aux_power_well_ops,
3278                 .id = DISP_PW_ID_NONE,
3279                 {
3280                         .hsw.regs = &icl_aux_power_well_regs,
3281                         .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT2,
3282                         .hsw.is_tc_tbt = true,
3283                 },
3284         },
3285         {
3286                 .name = "AUX TBT3",
3287                 .domains = ICL_AUX_TBT3_IO_POWER_DOMAINS,
3288                 .ops = &icl_tc_phy_aux_power_well_ops,
3289                 .id = DISP_PW_ID_NONE,
3290                 {
3291                         .hsw.regs = &icl_aux_power_well_regs,
3292                         .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT3,
3293                         .hsw.is_tc_tbt = true,
3294                 },
3295         },
3296         {
3297                 .name = "AUX TBT4",
3298                 .domains = ICL_AUX_TBT4_IO_POWER_DOMAINS,
3299                 .ops = &icl_tc_phy_aux_power_well_ops,
3300                 .id = DISP_PW_ID_NONE,
3301                 {
3302                         .hsw.regs = &icl_aux_power_well_regs,
3303                         .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT4,
3304                         .hsw.is_tc_tbt = true,
3305                 },
3306         },
3307         {
3308                 .name = "power well 4",
3309                 .domains = ICL_PW_4_POWER_DOMAINS,
3310                 .ops = &hsw_power_well_ops,
3311                 .id = DISP_PW_ID_NONE,
3312                 {
3313                         .hsw.regs = &hsw_power_well_regs,
3314                         .hsw.idx = ICL_PW_CTL_IDX_PW_4,
3315                         .hsw.has_fuses = true,
3316                         .hsw.irq_pipe_mask = BIT(PIPE_C),
3317                 },
3318         },
3319 };
3320
3321 static int
3322 sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv,
3323                                    int disable_power_well)
3324 {
3325         if (disable_power_well >= 0)
3326                 return !!disable_power_well;
3327
3328         return 1;
3329 }
3330
3331 static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv,
3332                                int enable_dc)
3333 {
3334         u32 mask;
3335         int requested_dc;
3336         int max_dc;
3337
3338         if (INTEL_GEN(dev_priv) >= 11) {
3339                 max_dc = 2;
3340                 /*
3341                  * DC9 has a separate HW flow from the rest of the DC states,
3342                  * not depending on the DMC firmware. It's needed by system
3343                  * suspend/resume, so allow it unconditionally.
3344                  */
3345                 mask = DC_STATE_EN_DC9;
3346         } else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv)) {
3347                 max_dc = 2;
3348                 mask = 0;
3349         } else if (IS_GEN9_LP(dev_priv)) {
3350                 max_dc = 1;
3351                 mask = DC_STATE_EN_DC9;
3352         } else {
3353                 max_dc = 0;
3354                 mask = 0;
3355         }
3356
3357         if (!i915_modparams.disable_power_well)
3358                 max_dc = 0;
3359
3360         if (enable_dc >= 0 && enable_dc <= max_dc) {
3361                 requested_dc = enable_dc;
3362         } else if (enable_dc == -1) {
3363                 requested_dc = max_dc;
3364         } else if (enable_dc > max_dc && enable_dc <= 2) {
3365                 DRM_DEBUG_KMS("Adjusting requested max DC state (%d->%d)\n",
3366                               enable_dc, max_dc);
3367                 requested_dc = max_dc;
3368         } else {
3369                 DRM_ERROR("Unexpected value for enable_dc (%d)\n", enable_dc);
3370                 requested_dc = max_dc;
3371         }
3372
3373         if (requested_dc > 1)
3374                 mask |= DC_STATE_EN_UPTO_DC6;
3375         if (requested_dc > 0)
3376                 mask |= DC_STATE_EN_UPTO_DC5;
3377
3378         DRM_DEBUG_KMS("Allowed DC state mask %02x\n", mask);
3379
3380         return mask;
3381 }
3382
3383 static int
3384 __set_power_wells(struct i915_power_domains *power_domains,
3385                   const struct i915_power_well_desc *power_well_descs,
3386                   int power_well_count)
3387 {
3388         u64 power_well_ids = 0;
3389         int i;
3390
3391         power_domains->power_well_count = power_well_count;
3392         power_domains->power_wells =
3393                                 kcalloc(power_well_count,
3394                                         sizeof(*power_domains->power_wells),
3395                                         GFP_KERNEL);
3396         if (!power_domains->power_wells)
3397                 return -ENOMEM;
3398
3399         for (i = 0; i < power_well_count; i++) {
3400                 enum i915_power_well_id id = power_well_descs[i].id;
3401
3402                 power_domains->power_wells[i].desc = &power_well_descs[i];
3403
3404                 if (id == DISP_PW_ID_NONE)
3405                         continue;
3406
3407                 WARN_ON(id >= sizeof(power_well_ids) * 8);
3408                 WARN_ON(power_well_ids & BIT_ULL(id));
3409                 power_well_ids |= BIT_ULL(id);
3410         }
3411
3412         return 0;
3413 }
3414
3415 #define set_power_wells(power_domains, __power_well_descs) \
3416         __set_power_wells(power_domains, __power_well_descs, \
3417                           ARRAY_SIZE(__power_well_descs))
3418
3419 /**
3420  * intel_power_domains_init - initializes the power domain structures
3421  * @dev_priv: i915 device instance
3422  *
3423  * Initializes the power domain structures for @dev_priv depending upon the
3424  * supported platform.
3425  */
3426 int intel_power_domains_init(struct drm_i915_private *dev_priv)
3427 {
3428         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3429         int err;
3430
3431         i915_modparams.disable_power_well =
3432                 sanitize_disable_power_well_option(dev_priv,
3433                                                    i915_modparams.disable_power_well);
3434         dev_priv->csr.allowed_dc_mask =
3435                 get_allowed_dc_mask(dev_priv, i915_modparams.enable_dc);
3436
3437         BUILD_BUG_ON(POWER_DOMAIN_NUM > 64);
3438
3439         mutex_init(&power_domains->lock);
3440
3441         /*
3442          * The enabling order will be from lower to higher indexed wells,
3443          * the disabling order is reversed.
3444          */
3445         if (IS_ICELAKE(dev_priv)) {
3446                 err = set_power_wells(power_domains, icl_power_wells);
3447         } else if (IS_CANNONLAKE(dev_priv)) {
3448                 err = set_power_wells(power_domains, cnl_power_wells);
3449
3450                 /*
3451                  * DDI and Aux IO are getting enabled for all ports
3452                  * regardless the presence or use. So, in order to avoid
3453                  * timeouts, lets remove them from the list
3454                  * for the SKUs without port F.
3455                  */
3456                 if (!IS_CNL_WITH_PORT_F(dev_priv))
3457                         power_domains->power_well_count -= 2;
3458         } else if (IS_GEMINILAKE(dev_priv)) {
3459                 err = set_power_wells(power_domains, glk_power_wells);
3460         } else if (IS_BROXTON(dev_priv)) {
3461                 err = set_power_wells(power_domains, bxt_power_wells);
3462         } else if (IS_GEN9_BC(dev_priv)) {
3463                 err = set_power_wells(power_domains, skl_power_wells);
3464         } else if (IS_CHERRYVIEW(dev_priv)) {
3465                 err = set_power_wells(power_domains, chv_power_wells);
3466         } else if (IS_BROADWELL(dev_priv)) {
3467                 err = set_power_wells(power_domains, bdw_power_wells);
3468         } else if (IS_HASWELL(dev_priv)) {
3469                 err = set_power_wells(power_domains, hsw_power_wells);
3470         } else if (IS_VALLEYVIEW(dev_priv)) {
3471                 err = set_power_wells(power_domains, vlv_power_wells);
3472         } else if (IS_I830(dev_priv)) {
3473                 err = set_power_wells(power_domains, i830_power_wells);
3474         } else {
3475                 err = set_power_wells(power_domains, i9xx_always_on_power_well);
3476         }
3477
3478         return err;
3479 }
3480
3481 /**
3482  * intel_power_domains_cleanup - clean up power domains resources
3483  * @dev_priv: i915 device instance
3484  *
3485  * Release any resources acquired by intel_power_domains_init()
3486  */
3487 void intel_power_domains_cleanup(struct drm_i915_private *dev_priv)
3488 {
3489         kfree(dev_priv->power_domains.power_wells);
3490 }
3491
3492 static void intel_power_domains_sync_hw(struct drm_i915_private *dev_priv)
3493 {
3494         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3495         struct i915_power_well *power_well;
3496
3497         mutex_lock(&power_domains->lock);
3498         for_each_power_well(dev_priv, power_well) {
3499                 power_well->desc->ops->sync_hw(dev_priv, power_well);
3500                 power_well->hw_enabled =
3501                         power_well->desc->ops->is_enabled(dev_priv, power_well);
3502         }
3503         mutex_unlock(&power_domains->lock);
3504 }
3505
3506 static inline
3507 bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv,
3508                           i915_reg_t reg, bool enable)
3509 {
3510         u32 val, status;
3511
3512         val = I915_READ(reg);
3513         val = enable ? (val | DBUF_POWER_REQUEST) : (val & ~DBUF_POWER_REQUEST);
3514         I915_WRITE(reg, val);
3515         POSTING_READ(reg);
3516         udelay(10);
3517
3518         status = I915_READ(reg) & DBUF_POWER_STATE;
3519         if ((enable && !status) || (!enable && status)) {
3520                 DRM_ERROR("DBus power %s timeout!\n",
3521                           enable ? "enable" : "disable");
3522                 return false;
3523         }
3524         return true;
3525 }
3526
3527 static void gen9_dbuf_enable(struct drm_i915_private *dev_priv)
3528 {
3529         intel_dbuf_slice_set(dev_priv, DBUF_CTL, true);
3530 }
3531
3532 static void gen9_dbuf_disable(struct drm_i915_private *dev_priv)
3533 {
3534         intel_dbuf_slice_set(dev_priv, DBUF_CTL, false);
3535 }
3536
3537 static u8 intel_dbuf_max_slices(struct drm_i915_private *dev_priv)
3538 {
3539         if (INTEL_GEN(dev_priv) < 11)
3540                 return 1;
3541         return 2;
3542 }
3543
3544 void icl_dbuf_slices_update(struct drm_i915_private *dev_priv,
3545                             u8 req_slices)
3546 {
3547         const u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices;
3548         bool ret;
3549
3550         if (req_slices > intel_dbuf_max_slices(dev_priv)) {
3551                 DRM_ERROR("Invalid number of dbuf slices requested\n");
3552                 return;
3553         }
3554
3555         if (req_slices == hw_enabled_slices || req_slices == 0)
3556                 return;
3557
3558         if (req_slices > hw_enabled_slices)
3559                 ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, true);
3560         else
3561                 ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, false);
3562
3563         if (ret)
3564                 dev_priv->wm.skl_hw.ddb.enabled_slices = req_slices;
3565 }
3566
3567 static void icl_dbuf_enable(struct drm_i915_private *dev_priv)
3568 {
3569         I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) | DBUF_POWER_REQUEST);
3570         I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) | DBUF_POWER_REQUEST);
3571         POSTING_READ(DBUF_CTL_S2);
3572
3573         udelay(10);
3574
3575         if (!(I915_READ(DBUF_CTL_S1) & DBUF_POWER_STATE) ||
3576             !(I915_READ(DBUF_CTL_S2) & DBUF_POWER_STATE))
3577                 DRM_ERROR("DBuf power enable timeout\n");
3578         else
3579                 dev_priv->wm.skl_hw.ddb.enabled_slices = 2;
3580 }
3581
3582 static void icl_dbuf_disable(struct drm_i915_private *dev_priv)
3583 {
3584         I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) & ~DBUF_POWER_REQUEST);
3585         I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) & ~DBUF_POWER_REQUEST);
3586         POSTING_READ(DBUF_CTL_S2);
3587
3588         udelay(10);
3589
3590         if ((I915_READ(DBUF_CTL_S1) & DBUF_POWER_STATE) ||
3591             (I915_READ(DBUF_CTL_S2) & DBUF_POWER_STATE))
3592                 DRM_ERROR("DBuf power disable timeout!\n");
3593         else
3594                 dev_priv->wm.skl_hw.ddb.enabled_slices = 0;
3595 }
3596
3597 static void icl_mbus_init(struct drm_i915_private *dev_priv)
3598 {
3599         u32 val;
3600
3601         val = MBUS_ABOX_BT_CREDIT_POOL1(16) |
3602               MBUS_ABOX_BT_CREDIT_POOL2(16) |
3603               MBUS_ABOX_B_CREDIT(1) |
3604               MBUS_ABOX_BW_CREDIT(1);
3605
3606         I915_WRITE(MBUS_ABOX_CTL, val);
3607 }
3608
3609 static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv,
3610                                       bool enable)
3611 {
3612         i915_reg_t reg;
3613         u32 reset_bits, val;
3614
3615         if (IS_IVYBRIDGE(dev_priv)) {
3616                 reg = GEN7_MSG_CTL;
3617                 reset_bits = WAIT_FOR_PCH_FLR_ACK | WAIT_FOR_PCH_RESET_ACK;
3618         } else {
3619                 reg = HSW_NDE_RSTWRN_OPT;
3620                 reset_bits = RESET_PCH_HANDSHAKE_ENABLE;
3621         }
3622
3623         val = I915_READ(reg);
3624
3625         if (enable)
3626                 val |= reset_bits;
3627         else
3628                 val &= ~reset_bits;
3629
3630         I915_WRITE(reg, val);
3631 }
3632
3633 static void skl_display_core_init(struct drm_i915_private *dev_priv,
3634                                    bool resume)
3635 {
3636         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3637         struct i915_power_well *well;
3638
3639         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3640
3641         /* enable PCH reset handshake */
3642         intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
3643
3644         /* enable PG1 and Misc I/O */
3645         mutex_lock(&power_domains->lock);
3646
3647         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3648         intel_power_well_enable(dev_priv, well);
3649
3650         well = lookup_power_well(dev_priv, SKL_DISP_PW_MISC_IO);
3651         intel_power_well_enable(dev_priv, well);
3652
3653         mutex_unlock(&power_domains->lock);
3654
3655         skl_init_cdclk(dev_priv);
3656
3657         gen9_dbuf_enable(dev_priv);
3658
3659         if (resume && dev_priv->csr.dmc_payload)
3660                 intel_csr_load_program(dev_priv);
3661 }
3662
3663 static void skl_display_core_uninit(struct drm_i915_private *dev_priv)
3664 {
3665         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3666         struct i915_power_well *well;
3667
3668         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3669
3670         gen9_dbuf_disable(dev_priv);
3671
3672         skl_uninit_cdclk(dev_priv);
3673
3674         /* The spec doesn't call for removing the reset handshake flag */
3675         /* disable PG1 and Misc I/O */
3676
3677         mutex_lock(&power_domains->lock);
3678
3679         /*
3680          * BSpec says to keep the MISC IO power well enabled here, only
3681          * remove our request for power well 1.
3682          * Note that even though the driver's request is removed power well 1
3683          * may stay enabled after this due to DMC's own request on it.
3684          */
3685         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3686         intel_power_well_disable(dev_priv, well);
3687
3688         mutex_unlock(&power_domains->lock);
3689
3690         usleep_range(10, 30);           /* 10 us delay per Bspec */
3691 }
3692
3693 void bxt_display_core_init(struct drm_i915_private *dev_priv,
3694                            bool resume)
3695 {
3696         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3697         struct i915_power_well *well;
3698
3699         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3700
3701         /*
3702          * NDE_RSTWRN_OPT RST PCH Handshake En must always be 0b on BXT
3703          * or else the reset will hang because there is no PCH to respond.
3704          * Move the handshake programming to initialization sequence.
3705          * Previously was left up to BIOS.
3706          */
3707         intel_pch_reset_handshake(dev_priv, false);
3708
3709         /* Enable PG1 */
3710         mutex_lock(&power_domains->lock);
3711
3712         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3713         intel_power_well_enable(dev_priv, well);
3714
3715         mutex_unlock(&power_domains->lock);
3716
3717         bxt_init_cdclk(dev_priv);
3718
3719         gen9_dbuf_enable(dev_priv);
3720
3721         if (resume && dev_priv->csr.dmc_payload)
3722                 intel_csr_load_program(dev_priv);
3723 }
3724
3725 void bxt_display_core_uninit(struct drm_i915_private *dev_priv)
3726 {
3727         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3728         struct i915_power_well *well;
3729
3730         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3731
3732         gen9_dbuf_disable(dev_priv);
3733
3734         bxt_uninit_cdclk(dev_priv);
3735
3736         /* The spec doesn't call for removing the reset handshake flag */
3737
3738         /*
3739          * Disable PW1 (PG1).
3740          * Note that even though the driver's request is removed power well 1
3741          * may stay enabled after this due to DMC's own request on it.
3742          */
3743         mutex_lock(&power_domains->lock);
3744
3745         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3746         intel_power_well_disable(dev_priv, well);
3747
3748         mutex_unlock(&power_domains->lock);
3749
3750         usleep_range(10, 30);           /* 10 us delay per Bspec */
3751 }
3752
3753 static void cnl_display_core_init(struct drm_i915_private *dev_priv, bool resume)
3754 {
3755         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3756         struct i915_power_well *well;
3757
3758         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3759
3760         /* 1. Enable PCH Reset Handshake */
3761         intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
3762
3763         /* 2-3. */
3764         cnl_combo_phys_init(dev_priv);
3765
3766         /*
3767          * 4. Enable Power Well 1 (PG1).
3768          *    The AUX IO power wells will be enabled on demand.
3769          */
3770         mutex_lock(&power_domains->lock);
3771         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3772         intel_power_well_enable(dev_priv, well);
3773         mutex_unlock(&power_domains->lock);
3774
3775         /* 5. Enable CD clock */
3776         cnl_init_cdclk(dev_priv);
3777
3778         /* 6. Enable DBUF */
3779         gen9_dbuf_enable(dev_priv);
3780
3781         if (resume && dev_priv->csr.dmc_payload)
3782                 intel_csr_load_program(dev_priv);
3783 }
3784
3785 static void cnl_display_core_uninit(struct drm_i915_private *dev_priv)
3786 {
3787         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3788         struct i915_power_well *well;
3789
3790         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3791
3792         /* 1. Disable all display engine functions -> aready done */
3793
3794         /* 2. Disable DBUF */
3795         gen9_dbuf_disable(dev_priv);
3796
3797         /* 3. Disable CD clock */
3798         cnl_uninit_cdclk(dev_priv);
3799
3800         /*
3801          * 4. Disable Power Well 1 (PG1).
3802          *    The AUX IO power wells are toggled on demand, so they are already
3803          *    disabled at this point.
3804          */
3805         mutex_lock(&power_domains->lock);
3806         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3807         intel_power_well_disable(dev_priv, well);
3808         mutex_unlock(&power_domains->lock);
3809
3810         usleep_range(10, 30);           /* 10 us delay per Bspec */
3811
3812         /* 5. */
3813         cnl_combo_phys_uninit(dev_priv);
3814 }
3815
3816 void icl_display_core_init(struct drm_i915_private *dev_priv,
3817                            bool resume)
3818 {
3819         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3820         struct i915_power_well *well;
3821
3822         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3823
3824         /* 1. Enable PCH reset handshake. */
3825         intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv));
3826
3827         /* 2-3. */
3828         icl_combo_phys_init(dev_priv);
3829
3830         /*
3831          * 4. Enable Power Well 1 (PG1).
3832          *    The AUX IO power wells will be enabled on demand.
3833          */
3834         mutex_lock(&power_domains->lock);
3835         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3836         intel_power_well_enable(dev_priv, well);
3837         mutex_unlock(&power_domains->lock);
3838
3839         /* 5. Enable CDCLK. */
3840         icl_init_cdclk(dev_priv);
3841
3842         /* 6. Enable DBUF. */
3843         icl_dbuf_enable(dev_priv);
3844
3845         /* 7. Setup MBUS. */
3846         icl_mbus_init(dev_priv);
3847
3848         if (resume && dev_priv->csr.dmc_payload)
3849                 intel_csr_load_program(dev_priv);
3850 }
3851
3852 void icl_display_core_uninit(struct drm_i915_private *dev_priv)
3853 {
3854         struct i915_power_domains *power_domains = &dev_priv->power_domains;
3855         struct i915_power_well *well;
3856
3857         gen9_set_dc_state(dev_priv, DC_STATE_DISABLE);
3858
3859         /* 1. Disable all display engine functions -> aready done */
3860
3861         /* 2. Disable DBUF */
3862         icl_dbuf_disable(dev_priv);
3863
3864         /* 3. Disable CD clock */
3865         icl_uninit_cdclk(dev_priv);
3866
3867         /*
3868          * 4. Disable Power Well 1 (PG1).
3869          *    The AUX IO power wells are toggled on demand, so they are already
3870          *    disabled at this point.
3871          */
3872         mutex_lock(&power_domains->lock);
3873         well = lookup_power_well(dev_priv, SKL_DISP_PW_1);
3874         intel_power_well_disable(dev_priv, well);
3875         mutex_unlock(&power_domains->lock);
3876
3877         /* 5. */
3878         icl_combo_phys_uninit(dev_priv);
3879 }
3880
3881 static void chv_phy_control_init(struct drm_i915_private *dev_priv)
3882 {
3883         struct i915_power_well *cmn_bc =
3884                 lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
3885         struct i915_power_well *cmn_d =
3886                 lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D);
3887
3888         /*
3889          * DISPLAY_PHY_CONTROL can get corrupted if read. As a
3890          * workaround never ever read DISPLAY_PHY_CONTROL, and
3891          * instead maintain a shadow copy ourselves. Use the actual
3892          * power well state and lane status to reconstruct the
3893          * expected initial value.
3894          */
3895         dev_priv->chv_phy_control =
3896                 PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY0) |
3897                 PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY1) |
3898                 PHY_CH_POWER_MODE(PHY_CH_DEEP_PSR, DPIO_PHY0, DPIO_CH0) |
3899                 PHY_CH_POWER_MODE(PHY_CH_DEEP_PSR, DPIO_PHY0, DPIO_CH1) |
3900                 PHY_CH_POWER_MODE(PHY_CH_DEEP_PSR, DPIO_PHY1, DPIO_CH0);
3901
3902         /*
3903          * If all lanes are disabled we leave the override disabled
3904          * with all power down bits cleared to match the state we
3905          * would use after disabling the port. Otherwise enable the
3906          * override and set the lane powerdown bits accding to the
3907          * current lane status.
3908          */
3909         if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) {
3910                 u32 status = I915_READ(DPLL(PIPE_A));
3911                 unsigned int mask;
3912
3913                 mask = status & DPLL_PORTB_READY_MASK;
3914                 if (mask == 0xf)
3915                         mask = 0x0;
3916                 else
3917                         dev_priv->chv_phy_control |=
3918                                 PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH0);
3919
3920                 dev_priv->chv_phy_control |=
3921                         PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH0);
3922
3923                 mask = (status & DPLL_PORTC_READY_MASK) >> 4;
3924                 if (mask == 0xf)
3925                         mask = 0x0;
3926                 else
3927                         dev_priv->chv_phy_control |=
3928                                 PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH1);
3929
3930                 dev_priv->chv_phy_control |=
3931                         PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH1);
3932
3933                 dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0);
3934
3935                 dev_priv->chv_phy_assert[DPIO_PHY0] = false;
3936         } else {
3937                 dev_priv->chv_phy_assert[DPIO_PHY0] = true;
3938         }
3939
3940         if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) {
3941                 u32 status = I915_READ(DPIO_PHY_STATUS);
3942                 unsigned int mask;
3943
3944                 mask = status & DPLL_PORTD_READY_MASK;
3945
3946                 if (mask == 0xf)
3947                         mask = 0x0;
3948                 else
3949                         dev_priv->chv_phy_control |=
3950                                 PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY1, DPIO_CH0);
3951
3952                 dev_priv->chv_phy_control |=
3953                         PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY1, DPIO_CH0);
3954
3955                 dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1);
3956
3957                 dev_priv->chv_phy_assert[DPIO_PHY1] = false;
3958         } else {
3959                 dev_priv->chv_phy_assert[DPIO_PHY1] = true;
3960         }
3961
3962         I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control);
3963
3964         DRM_DEBUG_KMS("Initial PHY_CONTROL=0x%08x\n",
3965                       dev_priv->chv_phy_control);
3966 }
3967
3968 static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv)
3969 {
3970         struct i915_power_well *cmn =
3971                 lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC);
3972         struct i915_power_well *disp2d =
3973                 lookup_power_well(dev_priv, VLV_DISP_PW_DISP2D);
3974
3975         /* If the display might be already active skip this */
3976         if (cmn->desc->ops->is_enabled(dev_priv, cmn) &&
3977             disp2d->desc->ops->is_enabled(dev_priv, disp2d) &&
3978             I915_READ(DPIO_CTL) & DPIO_CMNRST)
3979                 return;
3980
3981         DRM_DEBUG_KMS("toggling display PHY side reset\n");
3982
3983         /* cmnlane needs DPLL registers */
3984         disp2d->desc->ops->enable(dev_priv, disp2d);
3985
3986         /*
3987          * From VLV2A0_DP_eDP_HDMI_DPIO_driver_vbios_notes_11.docx:
3988          * Need to assert and de-assert PHY SB reset by gating the
3989          * common lane power, then un-gating it.
3990          * Simply ungating isn't enough to reset the PHY enough to get
3991          * ports and lanes running.
3992          */
3993         cmn->desc->ops->disable(dev_priv, cmn);
3994 }
3995
3996 static void intel_power_domains_verify_state(struct drm_i915_private *dev_priv);
3997
3998 /**
3999  * intel_power_domains_init_hw - initialize hardware power domain state
4000  * @i915: i915 device instance
4001  * @resume: Called from resume code paths or not
4002  *
4003  * This function initializes the hardware power domain state and enables all
4004  * power wells belonging to the INIT power domain. Power wells in other
4005  * domains (and not in the INIT domain) are referenced or disabled by
4006  * intel_modeset_readout_hw_state(). After that the reference count of each
4007  * power well must match its HW enabled state, see
4008  * intel_power_domains_verify_state().
4009  *
4010  * It will return with power domains disabled (to be enabled later by
4011  * intel_power_domains_enable()) and must be paired with
4012  * intel_power_domains_fini_hw().
4013  */
4014 void intel_power_domains_init_hw(struct drm_i915_private *i915, bool resume)
4015 {
4016         struct i915_power_domains *power_domains = &i915->power_domains;
4017
4018         power_domains->initializing = true;
4019
4020         if (IS_ICELAKE(i915)) {
4021                 icl_display_core_init(i915, resume);
4022         } else if (IS_CANNONLAKE(i915)) {
4023                 cnl_display_core_init(i915, resume);
4024         } else if (IS_GEN9_BC(i915)) {
4025                 skl_display_core_init(i915, resume);
4026         } else if (IS_GEN9_LP(i915)) {
4027                 bxt_display_core_init(i915, resume);
4028         } else if (IS_CHERRYVIEW(i915)) {
4029                 mutex_lock(&power_domains->lock);
4030                 chv_phy_control_init(i915);
4031                 mutex_unlock(&power_domains->lock);
4032         } else if (IS_VALLEYVIEW(i915)) {
4033                 mutex_lock(&power_domains->lock);
4034                 vlv_cmnlane_wa(i915);
4035                 mutex_unlock(&power_domains->lock);
4036         } else if (IS_IVYBRIDGE(i915) || INTEL_GEN(i915) >= 7) {
4037                 intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915));
4038         }
4039
4040         /*
4041          * Keep all power wells enabled for any dependent HW access during
4042          * initialization and to make sure we keep BIOS enabled display HW
4043          * resources powered until display HW readout is complete. We drop
4044          * this reference in intel_power_domains_enable().
4045          */
4046         power_domains->wakeref =
4047                 intel_display_power_get(i915, POWER_DOMAIN_INIT);
4048
4049         /* Disable power support if the user asked so. */
4050         if (!i915_modparams.disable_power_well)
4051                 intel_display_power_get(i915, POWER_DOMAIN_INIT);
4052         intel_power_domains_sync_hw(i915);
4053
4054         power_domains->initializing = false;
4055 }
4056
4057 /**
4058  * intel_power_domains_fini_hw - deinitialize hw power domain state
4059  * @i915: i915 device instance
4060  *
4061  * De-initializes the display power domain HW state. It also ensures that the
4062  * device stays powered up so that the driver can be reloaded.
4063  *
4064  * It must be called with power domains already disabled (after a call to
4065  * intel_power_domains_disable()) and must be paired with
4066  * intel_power_domains_init_hw().
4067  */
4068 void intel_power_domains_fini_hw(struct drm_i915_private *i915)
4069 {
4070         intel_wakeref_t wakeref __maybe_unused =
4071                 fetch_and_zero(&i915->power_domains.wakeref);
4072
4073         /* Remove the refcount we took to keep power well support disabled. */
4074         if (!i915_modparams.disable_power_well)
4075                 intel_display_power_put_unchecked(i915, POWER_DOMAIN_INIT);
4076
4077         intel_power_domains_verify_state(i915);
4078
4079         /* Keep the power well enabled, but cancel its rpm wakeref. */
4080         intel_runtime_pm_put(i915, wakeref);
4081 }
4082
4083 /**
4084  * intel_power_domains_enable - enable toggling of display power wells
4085  * @i915: i915 device instance
4086  *
4087  * Enable the ondemand enabling/disabling of the display power wells. Note that
4088  * power wells not belonging to POWER_DOMAIN_INIT are allowed to be toggled
4089  * only at specific points of the display modeset sequence, thus they are not
4090  * affected by the intel_power_domains_enable()/disable() calls. The purpose
4091  * of these function is to keep the rest of power wells enabled until the end
4092  * of display HW readout (which will acquire the power references reflecting
4093  * the current HW state).
4094  */
4095 void intel_power_domains_enable(struct drm_i915_private *i915)
4096 {
4097         intel_wakeref_t wakeref __maybe_unused =
4098                 fetch_and_zero(&i915->power_domains.wakeref);
4099
4100         intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
4101         intel_power_domains_verify_state(i915);
4102 }
4103
4104 /**
4105  * intel_power_domains_disable - disable toggling of display power wells
4106  * @i915: i915 device instance
4107  *
4108  * Disable the ondemand enabling/disabling of the display power wells. See
4109  * intel_power_domains_enable() for which power wells this call controls.
4110  */
4111 void intel_power_domains_disable(struct drm_i915_private *i915)
4112 {
4113         struct i915_power_domains *power_domains = &i915->power_domains;
4114
4115         WARN_ON(power_domains->wakeref);
4116         power_domains->wakeref =
4117                 intel_display_power_get(i915, POWER_DOMAIN_INIT);
4118
4119         intel_power_domains_verify_state(i915);
4120 }
4121
4122 /**
4123  * intel_power_domains_suspend - suspend power domain state
4124  * @i915: i915 device instance
4125  * @suspend_mode: specifies the target suspend state (idle, mem, hibernation)
4126  *
4127  * This function prepares the hardware power domain state before entering
4128  * system suspend.
4129  *
4130  * It must be called with power domains already disabled (after a call to
4131  * intel_power_domains_disable()) and paired with intel_power_domains_resume().
4132  */
4133 void intel_power_domains_suspend(struct drm_i915_private *i915,
4134                                  enum i915_drm_suspend_mode suspend_mode)
4135 {
4136         struct i915_power_domains *power_domains = &i915->power_domains;
4137         intel_wakeref_t wakeref __maybe_unused =
4138                 fetch_and_zero(&power_domains->wakeref);
4139
4140         intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref);
4141
4142         /*
4143          * In case of suspend-to-idle (aka S0ix) on a DMC platform without DC9
4144          * support don't manually deinit the power domains. This also means the
4145          * CSR/DMC firmware will stay active, it will power down any HW
4146          * resources as required and also enable deeper system power states
4147          * that would be blocked if the firmware was inactive.
4148          */
4149         if (!(i915->csr.allowed_dc_mask & DC_STATE_EN_DC9) &&
4150             suspend_mode == I915_DRM_SUSPEND_IDLE &&
4151             i915->csr.dmc_payload) {
4152                 intel_power_domains_verify_state(i915);
4153                 return;
4154         }
4155
4156         /*
4157          * Even if power well support was disabled we still want to disable
4158          * power wells if power domains must be deinitialized for suspend.
4159          */
4160         if (!i915_modparams.disable_power_well) {
4161                 intel_display_power_put_unchecked(i915, POWER_DOMAIN_INIT);
4162                 intel_power_domains_verify_state(i915);
4163         }
4164
4165         if (IS_ICELAKE(i915))
4166                 icl_display_core_uninit(i915);
4167         else if (IS_CANNONLAKE(i915))
4168                 cnl_display_core_uninit(i915);
4169         else if (IS_GEN9_BC(i915))
4170                 skl_display_core_uninit(i915);
4171         else if (IS_GEN9_LP(i915))
4172                 bxt_display_core_uninit(i915);
4173
4174         power_domains->display_core_suspended = true;
4175 }
4176
4177 /**
4178  * intel_power_domains_resume - resume power domain state
4179  * @i915: i915 device instance
4180  *
4181  * This function resume the hardware power domain state during system resume.
4182  *
4183  * It will return with power domain support disabled (to be enabled later by
4184  * intel_power_domains_enable()) and must be paired with
4185  * intel_power_domains_suspend().
4186  */
4187 void intel_power_domains_resume(struct drm_i915_private *i915)
4188 {
4189         struct i915_power_domains *power_domains = &i915->power_domains;
4190
4191         if (power_domains->display_core_suspended) {
4192                 intel_power_domains_init_hw(i915, true);
4193                 power_domains->display_core_suspended = false;
4194         } else {
4195                 WARN_ON(power_domains->wakeref);
4196                 power_domains->wakeref =
4197                         intel_display_power_get(i915, POWER_DOMAIN_INIT);
4198         }
4199
4200         intel_power_domains_verify_state(i915);
4201 }
4202
4203 #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
4204
4205 static void intel_power_domains_dump_info(struct drm_i915_private *i915)
4206 {
4207         struct i915_power_domains *power_domains = &i915->power_domains;
4208         struct i915_power_well *power_well;
4209
4210         for_each_power_well(i915, power_well) {
4211                 enum intel_display_power_domain domain;
4212
4213                 DRM_DEBUG_DRIVER("%-25s %d\n",
4214                                  power_well->desc->name, power_well->count);
4215
4216                 for_each_power_domain(domain, power_well->desc->domains)
4217                         DRM_DEBUG_DRIVER("  %-23s %d\n",
4218                                          intel_display_power_domain_str(domain),
4219                                          power_domains->domain_use_count[domain]);
4220         }
4221 }
4222
4223 /**
4224  * intel_power_domains_verify_state - verify the HW/SW state for all power wells
4225  * @i915: i915 device instance
4226  *
4227  * Verify if the reference count of each power well matches its HW enabled
4228  * state and the total refcount of the domains it belongs to. This must be
4229  * called after modeset HW state sanitization, which is responsible for
4230  * acquiring reference counts for any power wells in use and disabling the
4231  * ones left on by BIOS but not required by any active output.
4232  */
4233 static void intel_power_domains_verify_state(struct drm_i915_private *i915)
4234 {
4235         struct i915_power_domains *power_domains = &i915->power_domains;
4236         struct i915_power_well *power_well;
4237         bool dump_domain_info;
4238
4239         mutex_lock(&power_domains->lock);
4240
4241         dump_domain_info = false;
4242         for_each_power_well(i915, power_well) {
4243                 enum intel_display_power_domain domain;
4244                 int domains_count;
4245                 bool enabled;
4246
4247                 enabled = power_well->desc->ops->is_enabled(i915, power_well);
4248                 if ((power_well->count || power_well->desc->always_on) !=
4249                     enabled)
4250                         DRM_ERROR("power well %s state mismatch (refcount %d/enabled %d)",
4251                                   power_well->desc->name,
4252                                   power_well->count, enabled);
4253
4254                 domains_count = 0;
4255                 for_each_power_domain(domain, power_well->desc->domains)
4256                         domains_count += power_domains->domain_use_count[domain];
4257
4258                 if (power_well->count != domains_count) {
4259                         DRM_ERROR("power well %s refcount/domain refcount mismatch "
4260                                   "(refcount %d/domains refcount %d)\n",
4261                                   power_well->desc->name, power_well->count,
4262                                   domains_count);
4263                         dump_domain_info = true;
4264                 }
4265         }
4266
4267         if (dump_domain_info) {
4268                 static bool dumped;
4269
4270                 if (!dumped) {
4271                         intel_power_domains_dump_info(i915);
4272                         dumped = true;
4273                 }
4274         }
4275
4276         mutex_unlock(&power_domains->lock);
4277 }
4278
4279 #else
4280
4281 static void intel_power_domains_verify_state(struct drm_i915_private *i915)
4282 {
4283 }
4284
4285 #endif
4286
4287 /**
4288  * intel_runtime_pm_get - grab a runtime pm reference
4289  * @i915: i915 device instance
4290  *
4291  * This function grabs a device-level runtime pm reference (mostly used for GEM
4292  * code to ensure the GTT or GT is on) and ensures that it is powered up.
4293  *
4294  * Any runtime pm reference obtained by this function must have a symmetric
4295  * call to intel_runtime_pm_put() to release the reference again.
4296  *
4297  * Returns: the wakeref cookie to pass to intel_runtime_pm_put()
4298  */
4299 intel_wakeref_t intel_runtime_pm_get(struct drm_i915_private *i915)
4300 {
4301         struct pci_dev *pdev = i915->drm.pdev;
4302         struct device *kdev = &pdev->dev;
4303         int ret;
4304
4305         ret = pm_runtime_get_sync(kdev);
4306         WARN_ONCE(ret < 0, "pm_runtime_get_sync() failed: %d\n", ret);
4307
4308         return track_intel_runtime_pm_wakeref(i915);
4309 }
4310
4311 /**
4312  * intel_runtime_pm_get_if_in_use - grab a runtime pm reference if device in use
4313  * @i915: i915 device instance
4314  *
4315  * This function grabs a device-level runtime pm reference if the device is
4316  * already in use and ensures that it is powered up. It is illegal to try
4317  * and access the HW should intel_runtime_pm_get_if_in_use() report failure.
4318  *
4319  * Any runtime pm reference obtained by this function must have a symmetric
4320  * call to intel_runtime_pm_put() to release the reference again.
4321  *
4322  * Returns: the wakeref cookie to pass to intel_runtime_pm_put(), evaluates
4323  * as True if the wakeref was acquired, or False otherwise.
4324  */
4325 intel_wakeref_t intel_runtime_pm_get_if_in_use(struct drm_i915_private *i915)
4326 {
4327         if (IS_ENABLED(CONFIG_PM)) {
4328                 struct pci_dev *pdev = i915->drm.pdev;
4329                 struct device *kdev = &pdev->dev;
4330
4331                 /*
4332                  * In cases runtime PM is disabled by the RPM core and we get
4333                  * an -EINVAL return value we are not supposed to call this
4334                  * function, since the power state is undefined. This applies
4335                  * atm to the late/early system suspend/resume handlers.
4336                  */
4337                 if (pm_runtime_get_if_in_use(kdev) <= 0)
4338                         return 0;
4339         }
4340
4341         return track_intel_runtime_pm_wakeref(i915);
4342 }
4343
4344 /**
4345  * intel_runtime_pm_get_noresume - grab a runtime pm reference
4346  * @i915: i915 device instance
4347  *
4348  * This function grabs a device-level runtime pm reference (mostly used for GEM
4349  * code to ensure the GTT or GT is on).
4350  *
4351  * It will _not_ power up the device but instead only check that it's powered
4352  * on.  Therefore it is only valid to call this functions from contexts where
4353  * the device is known to be powered up and where trying to power it up would
4354  * result in hilarity and deadlocks. That pretty much means only the system
4355  * suspend/resume code where this is used to grab runtime pm references for
4356  * delayed setup down in work items.
4357  *
4358  * Any runtime pm reference obtained by this function must have a symmetric
4359  * call to intel_runtime_pm_put() to release the reference again.
4360  *
4361  * Returns: the wakeref cookie to pass to intel_runtime_pm_put()
4362  */
4363 intel_wakeref_t intel_runtime_pm_get_noresume(struct drm_i915_private *i915)
4364 {
4365         struct pci_dev *pdev = i915->drm.pdev;
4366         struct device *kdev = &pdev->dev;
4367
4368         assert_rpm_wakelock_held(i915);
4369         pm_runtime_get_noresume(kdev);
4370
4371         return track_intel_runtime_pm_wakeref(i915);
4372 }
4373
4374 /**
4375  * intel_runtime_pm_put - release a runtime pm reference
4376  * @i915: i915 device instance
4377  *
4378  * This function drops the device-level runtime pm reference obtained by
4379  * intel_runtime_pm_get() and might power down the corresponding
4380  * hardware block right away if this is the last reference.
4381  */
4382 void intel_runtime_pm_put_unchecked(struct drm_i915_private *i915)
4383 {
4384         struct pci_dev *pdev = i915->drm.pdev;
4385         struct device *kdev = &pdev->dev;
4386
4387         untrack_intel_runtime_pm_wakeref(i915);
4388
4389         pm_runtime_mark_last_busy(kdev);
4390         pm_runtime_put_autosuspend(kdev);
4391 }
4392
4393 #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)
4394 void intel_runtime_pm_put(struct drm_i915_private *i915, intel_wakeref_t wref)
4395 {
4396         cancel_intel_runtime_pm_wakeref(i915, wref);
4397         intel_runtime_pm_put_unchecked(i915);
4398 }
4399 #endif
4400
4401 /**
4402  * intel_runtime_pm_enable - enable runtime pm
4403  * @i915: i915 device instance
4404  *
4405  * This function enables runtime pm at the end of the driver load sequence.
4406  *
4407  * Note that this function does currently not enable runtime pm for the
4408  * subordinate display power domains. That is done by
4409  * intel_power_domains_enable().
4410  */
4411 void intel_runtime_pm_enable(struct drm_i915_private *i915)
4412 {
4413         struct pci_dev *pdev = i915->drm.pdev;
4414         struct device *kdev = &pdev->dev;
4415
4416         /*
4417          * Disable the system suspend direct complete optimization, which can
4418          * leave the device suspended skipping the driver's suspend handlers
4419          * if the device was already runtime suspended. This is needed due to
4420          * the difference in our runtime and system suspend sequence and
4421          * becaue the HDA driver may require us to enable the audio power
4422          * domain during system suspend.
4423          */
4424         dev_pm_set_driver_flags(kdev, DPM_FLAG_NEVER_SKIP);
4425
4426         pm_runtime_set_autosuspend_delay(kdev, 10000); /* 10s */
4427         pm_runtime_mark_last_busy(kdev);
4428
4429         /*
4430          * Take a permanent reference to disable the RPM functionality and drop
4431          * it only when unloading the driver. Use the low level get/put helpers,
4432          * so the driver's own RPM reference tracking asserts also work on
4433          * platforms without RPM support.
4434          */
4435         if (!HAS_RUNTIME_PM(i915)) {
4436                 int ret;
4437
4438                 pm_runtime_dont_use_autosuspend(kdev);
4439                 ret = pm_runtime_get_sync(kdev);
4440                 WARN(ret < 0, "pm_runtime_get_sync() failed: %d\n", ret);
4441         } else {
4442                 pm_runtime_use_autosuspend(kdev);
4443         }
4444
4445         /*
4446          * The core calls the driver load handler with an RPM reference held.
4447          * We drop that here and will reacquire it during unloading in
4448          * intel_power_domains_fini().
4449          */
4450         pm_runtime_put_autosuspend(kdev);
4451 }
4452
4453 void intel_runtime_pm_disable(struct drm_i915_private *i915)
4454 {
4455         struct pci_dev *pdev = i915->drm.pdev;
4456         struct device *kdev = &pdev->dev;
4457
4458         /* Transfer rpm ownership back to core */
4459         WARN(pm_runtime_get_sync(kdev) < 0,
4460              "Failed to pass rpm ownership back to core\n");
4461
4462         pm_runtime_dont_use_autosuspend(kdev);
4463
4464         if (!HAS_RUNTIME_PM(i915))
4465                 pm_runtime_put(kdev);
4466 }
4467
4468 void intel_runtime_pm_cleanup(struct drm_i915_private *i915)
4469 {
4470         struct i915_runtime_pm *rpm = &i915->runtime_pm;
4471         int count;
4472
4473         count = atomic_fetch_inc(&rpm->wakeref_count); /* balance untrack */
4474         WARN(count,
4475              "i915->runtime_pm.wakeref_count=%d on cleanup\n",
4476              count);
4477
4478         untrack_intel_runtime_pm_wakeref(i915);
4479 }
4480
4481 void intel_runtime_pm_init_early(struct drm_i915_private *i915)
4482 {
4483         init_intel_runtime_pm_wakeref(i915);
4484 }