2 * SPDX-License-Identifier: GPL-2.0
3 * Copyright (c) 2018, The Linux Foundation
7 #include <linux/clk-provider.h>
8 #include <linux/iopoll.h>
14 * DSI PLL 10nm - clock diagram (eg: DSI0):
16 * dsi0_pll_out_div_clk dsi0_pll_bit_clk
19 * +---------+ | +----------+ | +----+
20 * dsi0vco_clk ---| out_div |--o--| divl_3_0 |--o--| /8 |-- dsi0_phy_pll_out_byteclk
21 * +---------+ | +----------+ | +----+
23 * | | dsi0_pll_by_2_bit_clk
25 * | | +----+ | |\ dsi0_pclk_mux
26 * | |--| /2 |--o--| \ |
27 * | | +----+ | \ | +---------+
28 * | --------------| |--o--| div_7_4 |-- dsi0_phy_pll_out_dsiclk
29 * |------------------------------| / +---------+
31 * -----------| /4? |--o----------|/
35 * dsi0_pll_post_out_div_clk
38 #define VCO_REF_CLK_RATE 19200000
41 u32 pll_prop_gain_rate;
43 u32 decimal_div_start;
44 u32 frac_div_start_low;
45 u32 frac_div_start_mid;
46 u32 frac_div_start_high;
47 u32 pll_clock_inverters;
49 u32 ssc_stepsize_high;
57 /* v3.0.0 10nm implementation that requires the old timings settings */
58 #define DSI_PHY_10NM_QUIRK_OLD_TIMINGS BIT(0)
60 struct dsi_pll_config {
65 bool disable_prescaler;
78 struct pll_10nm_cached_state {
79 unsigned long vco_rate;
89 struct msm_dsi_phy *phy;
94 /* protects REG_DSI_10nm_PHY_CMN_CLK_CFG0 register */
95 spinlock_t postdiv_lock;
97 struct dsi_pll_config pll_configuration;
98 struct dsi_pll_regs reg_setup;
100 struct pll_10nm_cached_state cached_state;
102 struct dsi_pll_10nm *slave;
105 #define to_pll_10nm(x) container_of(x, struct dsi_pll_10nm, clk_hw)
108 * Global list of private DSI PLL struct pointers. We need this for Dual DSI
109 * mode, where the master PLL's clk_ops needs access the slave's private data
111 static struct dsi_pll_10nm *pll_10nm_list[DSI_MAX];
113 static void dsi_pll_setup_config(struct dsi_pll_10nm *pll)
115 struct dsi_pll_config *config = &pll->pll_configuration;
117 config->ref_freq = pll->vco_ref_clk_rate;
118 config->output_div = 1;
119 config->dec_bits = 8;
120 config->frac_bits = 18;
121 config->lock_timer = 64;
122 config->ssc_freq = 31500;
123 config->ssc_offset = 5000;
124 config->ssc_adj_per = 2;
125 config->thresh_cycles = 32;
126 config->refclk_cycles = 256;
128 config->div_override = false;
129 config->ignore_frac = false;
130 config->disable_prescaler = false;
132 config->enable_ssc = false;
133 config->ssc_center = 0;
136 static void dsi_pll_calc_dec_frac(struct dsi_pll_10nm *pll)
138 struct dsi_pll_config *config = &pll->pll_configuration;
139 struct dsi_pll_regs *regs = &pll->reg_setup;
140 u64 fref = pll->vco_ref_clk_rate;
143 u64 dec, dec_multiple;
147 pll_freq = pll->vco_current_rate;
149 if (config->disable_prescaler)
154 multiplier = 1 << config->frac_bits;
155 dec_multiple = div_u64(pll_freq * multiplier, divider);
156 dec = div_u64_rem(dec_multiple, multiplier, &frac);
158 if (pll_freq <= 1900000000UL)
159 regs->pll_prop_gain_rate = 8;
160 else if (pll_freq <= 3000000000UL)
161 regs->pll_prop_gain_rate = 10;
163 regs->pll_prop_gain_rate = 12;
164 if (pll_freq < 1100000000UL)
165 regs->pll_clock_inverters = 8;
167 regs->pll_clock_inverters = 0;
169 regs->pll_lockdet_rate = config->lock_timer;
170 regs->decimal_div_start = dec;
171 regs->frac_div_start_low = (frac & 0xff);
172 regs->frac_div_start_mid = (frac & 0xff00) >> 8;
173 regs->frac_div_start_high = (frac & 0x30000) >> 16;
176 #define SSC_CENTER BIT(0)
177 #define SSC_EN BIT(1)
179 static void dsi_pll_calc_ssc(struct dsi_pll_10nm *pll)
181 struct dsi_pll_config *config = &pll->pll_configuration;
182 struct dsi_pll_regs *regs = &pll->reg_setup;
188 if (!config->enable_ssc) {
189 DBG("SSC not enabled\n");
193 ssc_per = DIV_ROUND_CLOSEST(config->ref_freq, config->ssc_freq) / 2 - 1;
194 ssc_mod = (ssc_per + 1) % (config->ssc_adj_per + 1);
197 frac = regs->frac_div_start_low |
198 (regs->frac_div_start_mid << 8) |
199 (regs->frac_div_start_high << 16);
200 ssc_step_size = regs->decimal_div_start;
201 ssc_step_size *= (1 << config->frac_bits);
202 ssc_step_size += frac;
203 ssc_step_size *= config->ssc_offset;
204 ssc_step_size *= (config->ssc_adj_per + 1);
205 ssc_step_size = div_u64(ssc_step_size, (ssc_per + 1));
206 ssc_step_size = DIV_ROUND_CLOSEST_ULL(ssc_step_size, 1000000);
208 regs->ssc_div_per_low = ssc_per & 0xFF;
209 regs->ssc_div_per_high = (ssc_per & 0xFF00) >> 8;
210 regs->ssc_stepsize_low = (u32)(ssc_step_size & 0xFF);
211 regs->ssc_stepsize_high = (u32)((ssc_step_size & 0xFF00) >> 8);
212 regs->ssc_adjper_low = config->ssc_adj_per & 0xFF;
213 regs->ssc_adjper_high = (config->ssc_adj_per & 0xFF00) >> 8;
215 regs->ssc_control = config->ssc_center ? SSC_CENTER : 0;
217 pr_debug("SCC: Dec:%d, frac:%llu, frac_bits:%d\n",
218 regs->decimal_div_start, frac, config->frac_bits);
219 pr_debug("SSC: div_per:0x%X, stepsize:0x%X, adjper:0x%X\n",
220 ssc_per, (u32)ssc_step_size, config->ssc_adj_per);
223 static void dsi_pll_ssc_commit(struct dsi_pll_10nm *pll)
225 void __iomem *base = pll->phy->pll_base;
226 struct dsi_pll_regs *regs = &pll->reg_setup;
228 if (pll->pll_configuration.enable_ssc) {
229 pr_debug("SSC is enabled\n");
231 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_STEPSIZE_LOW_1,
232 regs->ssc_stepsize_low);
233 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_STEPSIZE_HIGH_1,
234 regs->ssc_stepsize_high);
235 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_DIV_PER_LOW_1,
236 regs->ssc_div_per_low);
237 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_DIV_PER_HIGH_1,
238 regs->ssc_div_per_high);
239 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_DIV_ADJPER_LOW_1,
240 regs->ssc_adjper_low);
241 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_DIV_ADJPER_HIGH_1,
242 regs->ssc_adjper_high);
243 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_SSC_CONTROL,
244 SSC_EN | regs->ssc_control);
248 static void dsi_pll_config_hzindep_reg(struct dsi_pll_10nm *pll)
250 void __iomem *base = pll->phy->pll_base;
252 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_ANALOG_CONTROLS_ONE, 0x80);
253 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_ANALOG_CONTROLS_TWO, 0x03);
254 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_ANALOG_CONTROLS_THREE, 0x00);
255 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_DSM_DIVIDER, 0x00);
256 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_FEEDBACK_DIVIDER, 0x4e);
257 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_CALIBRATION_SETTINGS, 0x40);
258 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_BAND_SEL_CAL_SETTINGS_THREE,
260 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_FREQ_DETECT_SETTINGS_ONE, 0x0c);
261 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_OUTDIV, 0x00);
262 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_CORE_OVERRIDE, 0x00);
263 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_DIGITAL_TIMERS_TWO, 0x08);
264 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_PROP_GAIN_RATE_1, 0x08);
265 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_BAND_SET_RATE_1, 0xc0);
266 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_INT_GAIN_IFILT_BAND_1, 0xfa);
267 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_FL_INT_GAIN_PFILT_BAND_1,
269 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_LOCK_OVERRIDE, 0x80);
270 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PFILT, 0x29);
271 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_IFILT, 0x3f);
274 static void dsi_pll_commit(struct dsi_pll_10nm *pll)
276 void __iomem *base = pll->phy->pll_base;
277 struct dsi_pll_regs *reg = &pll->reg_setup;
279 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_CORE_INPUT_OVERRIDE, 0x12);
280 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_DECIMAL_DIV_START_1,
281 reg->decimal_div_start);
282 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_FRAC_DIV_START_LOW_1,
283 reg->frac_div_start_low);
284 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_FRAC_DIV_START_MID_1,
285 reg->frac_div_start_mid);
286 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_FRAC_DIV_START_HIGH_1,
287 reg->frac_div_start_high);
288 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_LOCKDET_RATE_1,
289 reg->pll_lockdet_rate);
290 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_PLL_LOCK_DELAY, 0x06);
291 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_CMODE, 0x10);
292 dsi_phy_write(base + REG_DSI_10nm_PHY_PLL_CLOCK_INVERTERS,
293 reg->pll_clock_inverters);
296 static int dsi_pll_10nm_vco_set_rate(struct clk_hw *hw, unsigned long rate,
297 unsigned long parent_rate)
299 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(hw);
301 DBG("DSI PLL%d rate=%lu, parent's=%lu", pll_10nm->phy->id, rate,
304 pll_10nm->vco_current_rate = rate;
305 pll_10nm->vco_ref_clk_rate = VCO_REF_CLK_RATE;
307 dsi_pll_setup_config(pll_10nm);
309 dsi_pll_calc_dec_frac(pll_10nm);
311 dsi_pll_calc_ssc(pll_10nm);
313 dsi_pll_commit(pll_10nm);
315 dsi_pll_config_hzindep_reg(pll_10nm);
317 dsi_pll_ssc_commit(pll_10nm);
319 /* flush, ensure all register writes are done*/
325 static int dsi_pll_10nm_lock_status(struct dsi_pll_10nm *pll)
327 struct device *dev = &pll->phy->pdev->dev;
330 u32 const delay_us = 100;
331 u32 const timeout_us = 5000;
333 rc = readl_poll_timeout_atomic(pll->phy->pll_base +
334 REG_DSI_10nm_PHY_PLL_COMMON_STATUS_ONE,
336 ((status & BIT(0)) > 0),
340 DRM_DEV_ERROR(dev, "DSI PLL(%d) lock failed, status=0x%08x\n",
341 pll->phy->id, status);
346 static void dsi_pll_disable_pll_bias(struct dsi_pll_10nm *pll)
348 u32 data = dsi_phy_read(pll->phy->base + REG_DSI_10nm_PHY_CMN_CTRL_0);
350 dsi_phy_write(pll->phy->pll_base + REG_DSI_10nm_PHY_PLL_SYSTEM_MUXES, 0);
351 dsi_phy_write(pll->phy->base + REG_DSI_10nm_PHY_CMN_CTRL_0,
356 static void dsi_pll_enable_pll_bias(struct dsi_pll_10nm *pll)
358 u32 data = dsi_phy_read(pll->phy->base + REG_DSI_10nm_PHY_CMN_CTRL_0);
360 dsi_phy_write(pll->phy->base + REG_DSI_10nm_PHY_CMN_CTRL_0,
362 dsi_phy_write(pll->phy->pll_base + REG_DSI_10nm_PHY_PLL_SYSTEM_MUXES, 0xc0);
366 static void dsi_pll_disable_global_clk(struct dsi_pll_10nm *pll)
370 data = dsi_phy_read(pll->phy->base + REG_DSI_10nm_PHY_CMN_CLK_CFG1);
371 dsi_phy_write(pll->phy->base + REG_DSI_10nm_PHY_CMN_CLK_CFG1,
375 static void dsi_pll_enable_global_clk(struct dsi_pll_10nm *pll)
379 data = dsi_phy_read(pll->phy->base + REG_DSI_10nm_PHY_CMN_CLK_CFG1);
380 dsi_phy_write(pll->phy->base + REG_DSI_10nm_PHY_CMN_CLK_CFG1,
384 static int dsi_pll_10nm_vco_prepare(struct clk_hw *hw)
386 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(hw);
387 struct device *dev = &pll_10nm->phy->pdev->dev;
390 dsi_pll_enable_pll_bias(pll_10nm);
392 dsi_pll_enable_pll_bias(pll_10nm->slave);
394 rc = dsi_pll_10nm_vco_set_rate(hw,pll_10nm->vco_current_rate, 0);
396 DRM_DEV_ERROR(dev, "vco_set_rate failed, rc=%d\n", rc);
401 dsi_phy_write(pll_10nm->phy->base + REG_DSI_10nm_PHY_CMN_PLL_CNTRL,
405 * ensure all PLL configurations are written prior to checking
410 /* Check for PLL lock */
411 rc = dsi_pll_10nm_lock_status(pll_10nm);
413 DRM_DEV_ERROR(dev, "PLL(%d) lock failed\n", pll_10nm->phy->id);
417 pll_10nm->phy->pll_on = true;
419 dsi_pll_enable_global_clk(pll_10nm);
421 dsi_pll_enable_global_clk(pll_10nm->slave);
423 dsi_phy_write(pll_10nm->phy->base + REG_DSI_10nm_PHY_CMN_RBUF_CTRL,
426 dsi_phy_write(pll_10nm->slave->phy->base +
427 REG_DSI_10nm_PHY_CMN_RBUF_CTRL, 0x01);
433 static void dsi_pll_disable_sub(struct dsi_pll_10nm *pll)
435 dsi_phy_write(pll->phy->base + REG_DSI_10nm_PHY_CMN_RBUF_CTRL, 0);
436 dsi_pll_disable_pll_bias(pll);
439 static void dsi_pll_10nm_vco_unprepare(struct clk_hw *hw)
441 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(hw);
444 * To avoid any stray glitches while abruptly powering down the PLL
445 * make sure to gate the clock using the clock enable bit before
446 * powering down the PLL
448 dsi_pll_disable_global_clk(pll_10nm);
449 dsi_phy_write(pll_10nm->phy->base + REG_DSI_10nm_PHY_CMN_PLL_CNTRL, 0);
450 dsi_pll_disable_sub(pll_10nm);
451 if (pll_10nm->slave) {
452 dsi_pll_disable_global_clk(pll_10nm->slave);
453 dsi_pll_disable_sub(pll_10nm->slave);
455 /* flush, ensure all register writes are done */
457 pll_10nm->phy->pll_on = false;
460 static unsigned long dsi_pll_10nm_vco_recalc_rate(struct clk_hw *hw,
461 unsigned long parent_rate)
463 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(hw);
464 struct dsi_pll_config *config = &pll_10nm->pll_configuration;
465 void __iomem *base = pll_10nm->phy->pll_base;
466 u64 ref_clk = pll_10nm->vco_ref_clk_rate;
473 dec = dsi_phy_read(base + REG_DSI_10nm_PHY_PLL_DECIMAL_DIV_START_1);
476 frac = dsi_phy_read(base + REG_DSI_10nm_PHY_PLL_FRAC_DIV_START_LOW_1);
477 frac |= ((dsi_phy_read(base + REG_DSI_10nm_PHY_PLL_FRAC_DIV_START_MID_1) &
479 frac |= ((dsi_phy_read(base + REG_DSI_10nm_PHY_PLL_FRAC_DIV_START_HIGH_1) &
484 * 1. Assumes prescaler is disabled
486 multiplier = 1 << config->frac_bits;
487 pll_freq = dec * (ref_clk * 2);
488 tmp64 = (ref_clk * 2 * frac);
489 pll_freq += div_u64(tmp64, multiplier);
493 DBG("DSI PLL%d returning vco rate = %lu, dec = %x, frac = %x",
494 pll_10nm->phy->id, (unsigned long)vco_rate, dec, frac);
496 return (unsigned long)vco_rate;
499 static long dsi_pll_10nm_clk_round_rate(struct clk_hw *hw,
500 unsigned long rate, unsigned long *parent_rate)
502 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(hw);
504 if (rate < pll_10nm->phy->cfg->min_pll_rate)
505 return pll_10nm->phy->cfg->min_pll_rate;
506 else if (rate > pll_10nm->phy->cfg->max_pll_rate)
507 return pll_10nm->phy->cfg->max_pll_rate;
512 static const struct clk_ops clk_ops_dsi_pll_10nm_vco = {
513 .round_rate = dsi_pll_10nm_clk_round_rate,
514 .set_rate = dsi_pll_10nm_vco_set_rate,
515 .recalc_rate = dsi_pll_10nm_vco_recalc_rate,
516 .prepare = dsi_pll_10nm_vco_prepare,
517 .unprepare = dsi_pll_10nm_vco_unprepare,
524 static void dsi_10nm_pll_save_state(struct msm_dsi_phy *phy)
526 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(phy->vco_hw);
527 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state;
528 void __iomem *phy_base = pll_10nm->phy->base;
529 u32 cmn_clk_cfg0, cmn_clk_cfg1;
531 cached->pll_out_div = dsi_phy_read(pll_10nm->phy->pll_base +
532 REG_DSI_10nm_PHY_PLL_PLL_OUTDIV_RATE);
533 cached->pll_out_div &= 0x3;
535 cmn_clk_cfg0 = dsi_phy_read(phy_base + REG_DSI_10nm_PHY_CMN_CLK_CFG0);
536 cached->bit_clk_div = cmn_clk_cfg0 & 0xf;
537 cached->pix_clk_div = (cmn_clk_cfg0 & 0xf0) >> 4;
539 cmn_clk_cfg1 = dsi_phy_read(phy_base + REG_DSI_10nm_PHY_CMN_CLK_CFG1);
540 cached->pll_mux = cmn_clk_cfg1 & 0x3;
542 DBG("DSI PLL%d outdiv %x bit_clk_div %x pix_clk_div %x pll_mux %x",
543 pll_10nm->phy->id, cached->pll_out_div, cached->bit_clk_div,
544 cached->pix_clk_div, cached->pll_mux);
547 static int dsi_10nm_pll_restore_state(struct msm_dsi_phy *phy)
549 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(phy->vco_hw);
550 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state;
551 void __iomem *phy_base = pll_10nm->phy->base;
555 val = dsi_phy_read(pll_10nm->phy->pll_base + REG_DSI_10nm_PHY_PLL_PLL_OUTDIV_RATE);
557 val |= cached->pll_out_div;
558 dsi_phy_write(pll_10nm->phy->pll_base + REG_DSI_10nm_PHY_PLL_PLL_OUTDIV_RATE, val);
560 dsi_phy_write(phy_base + REG_DSI_10nm_PHY_CMN_CLK_CFG0,
561 cached->bit_clk_div | (cached->pix_clk_div << 4));
563 val = dsi_phy_read(phy_base + REG_DSI_10nm_PHY_CMN_CLK_CFG1);
565 val |= cached->pll_mux;
566 dsi_phy_write(phy_base + REG_DSI_10nm_PHY_CMN_CLK_CFG1, val);
568 ret = dsi_pll_10nm_vco_set_rate(phy->vco_hw,
569 pll_10nm->vco_current_rate,
570 pll_10nm->vco_ref_clk_rate);
572 DRM_DEV_ERROR(&pll_10nm->phy->pdev->dev,
573 "restore vco rate failed. ret=%d\n", ret);
577 DBG("DSI PLL%d", pll_10nm->phy->id);
582 static int dsi_10nm_set_usecase(struct msm_dsi_phy *phy)
584 struct dsi_pll_10nm *pll_10nm = to_pll_10nm(phy->vco_hw);
585 void __iomem *base = phy->base;
586 u32 data = 0x0; /* internal PLL */
588 DBG("DSI PLL%d", pll_10nm->phy->id);
590 switch (phy->usecase) {
591 case MSM_DSI_PHY_STANDALONE:
593 case MSM_DSI_PHY_MASTER:
594 pll_10nm->slave = pll_10nm_list[(pll_10nm->phy->id + 1) % DSI_MAX];
596 case MSM_DSI_PHY_SLAVE:
597 data = 0x1; /* external PLL */
604 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CLK_CFG1, (data << 2));
610 * The post dividers and mux clocks are created using the standard divider and
611 * mux API. Unlike the 14nm PHY, the slave PLL doesn't need its dividers/mux
612 * state to follow the master PLL's divider/mux state. Therefore, we don't
613 * require special clock ops that also configure the slave PLL registers
615 static int pll_10nm_register(struct dsi_pll_10nm *pll_10nm, struct clk_hw **provided_clocks)
617 char clk_name[32], parent[32], vco_name[32];
618 char parent2[32], parent3[32], parent4[32];
619 struct clk_init_data vco_init = {
620 .parent_names = (const char *[]){ "xo" },
623 .flags = CLK_IGNORE_UNUSED,
624 .ops = &clk_ops_dsi_pll_10nm_vco,
626 struct device *dev = &pll_10nm->phy->pdev->dev;
630 DBG("DSI%d", pll_10nm->phy->id);
632 snprintf(vco_name, 32, "dsi%dvco_clk", pll_10nm->phy->id);
633 pll_10nm->clk_hw.init = &vco_init;
635 ret = devm_clk_hw_register(dev, &pll_10nm->clk_hw);
639 snprintf(clk_name, 32, "dsi%d_pll_out_div_clk", pll_10nm->phy->id);
640 snprintf(parent, 32, "dsi%dvco_clk", pll_10nm->phy->id);
642 hw = devm_clk_hw_register_divider(dev, clk_name,
643 parent, CLK_SET_RATE_PARENT,
644 pll_10nm->phy->pll_base +
645 REG_DSI_10nm_PHY_PLL_PLL_OUTDIV_RATE,
646 0, 2, CLK_DIVIDER_POWER_OF_TWO, NULL);
652 snprintf(clk_name, 32, "dsi%d_pll_bit_clk", pll_10nm->phy->id);
653 snprintf(parent, 32, "dsi%d_pll_out_div_clk", pll_10nm->phy->id);
655 /* BIT CLK: DIV_CTRL_3_0 */
656 hw = devm_clk_hw_register_divider(dev, clk_name, parent,
658 pll_10nm->phy->base +
659 REG_DSI_10nm_PHY_CMN_CLK_CFG0,
660 0, 4, CLK_DIVIDER_ONE_BASED,
661 &pll_10nm->postdiv_lock);
667 snprintf(clk_name, 32, "dsi%d_phy_pll_out_byteclk", pll_10nm->phy->id);
668 snprintf(parent, 32, "dsi%d_pll_bit_clk", pll_10nm->phy->id);
670 /* DSI Byte clock = VCO_CLK / OUT_DIV / BIT_DIV / 8 */
671 hw = devm_clk_hw_register_fixed_factor(dev, clk_name, parent,
672 CLK_SET_RATE_PARENT, 1, 8);
678 provided_clocks[DSI_BYTE_PLL_CLK] = hw;
680 snprintf(clk_name, 32, "dsi%d_pll_by_2_bit_clk", pll_10nm->phy->id);
681 snprintf(parent, 32, "dsi%d_pll_bit_clk", pll_10nm->phy->id);
683 hw = devm_clk_hw_register_fixed_factor(dev, clk_name, parent,
690 snprintf(clk_name, 32, "dsi%d_pll_post_out_div_clk", pll_10nm->phy->id);
691 snprintf(parent, 32, "dsi%d_pll_out_div_clk", pll_10nm->phy->id);
693 hw = devm_clk_hw_register_fixed_factor(dev, clk_name, parent,
700 snprintf(clk_name, 32, "dsi%d_pclk_mux", pll_10nm->phy->id);
701 snprintf(parent, 32, "dsi%d_pll_bit_clk", pll_10nm->phy->id);
702 snprintf(parent2, 32, "dsi%d_pll_by_2_bit_clk", pll_10nm->phy->id);
703 snprintf(parent3, 32, "dsi%d_pll_out_div_clk", pll_10nm->phy->id);
704 snprintf(parent4, 32, "dsi%d_pll_post_out_div_clk", pll_10nm->phy->id);
706 hw = devm_clk_hw_register_mux(dev, clk_name,
708 parent, parent2, parent3, parent4
709 }), 4, 0, pll_10nm->phy->base +
710 REG_DSI_10nm_PHY_CMN_CLK_CFG1,
717 snprintf(clk_name, 32, "dsi%d_phy_pll_out_dsiclk", pll_10nm->phy->id);
718 snprintf(parent, 32, "dsi%d_pclk_mux", pll_10nm->phy->id);
720 /* PIX CLK DIV : DIV_CTRL_7_4*/
721 hw = devm_clk_hw_register_divider(dev, clk_name, parent,
722 0, pll_10nm->phy->base +
723 REG_DSI_10nm_PHY_CMN_CLK_CFG0,
724 4, 4, CLK_DIVIDER_ONE_BASED,
725 &pll_10nm->postdiv_lock);
731 provided_clocks[DSI_PIXEL_PLL_CLK] = hw;
740 static int dsi_pll_10nm_init(struct msm_dsi_phy *phy)
742 struct platform_device *pdev = phy->pdev;
743 struct dsi_pll_10nm *pll_10nm;
746 pll_10nm = devm_kzalloc(&pdev->dev, sizeof(*pll_10nm), GFP_KERNEL);
750 DBG("DSI PLL%d", phy->id);
752 pll_10nm_list[phy->id] = pll_10nm;
754 spin_lock_init(&pll_10nm->postdiv_lock);
758 ret = pll_10nm_register(pll_10nm, phy->provided_clocks->hws);
760 DRM_DEV_ERROR(&pdev->dev, "failed to register PLL: %d\n", ret);
764 phy->vco_hw = &pll_10nm->clk_hw;
766 /* TODO: Remove this when we have proper display handover support */
767 msm_dsi_phy_pll_save_state(phy);
772 static int dsi_phy_hw_v3_0_is_pll_on(struct msm_dsi_phy *phy)
774 void __iomem *base = phy->base;
777 data = dsi_phy_read(base + REG_DSI_10nm_PHY_CMN_PLL_CNTRL);
778 mb(); /* make sure read happened */
780 return (data & BIT(0));
783 static void dsi_phy_hw_v3_0_config_lpcdrx(struct msm_dsi_phy *phy, bool enable)
785 void __iomem *lane_base = phy->lane_base;
786 int phy_lane_0 = 0; /* TODO: Support all lane swap configs */
789 * LPRX and CDRX need to enabled only for physical data lane
790 * corresponding to the logical data lane 0
793 dsi_phy_write(lane_base +
794 REG_DSI_10nm_PHY_LN_LPRX_CTRL(phy_lane_0), 0x3);
796 dsi_phy_write(lane_base +
797 REG_DSI_10nm_PHY_LN_LPRX_CTRL(phy_lane_0), 0);
800 static void dsi_phy_hw_v3_0_lane_settings(struct msm_dsi_phy *phy)
803 u8 tx_dctrl[] = { 0x00, 0x00, 0x00, 0x04, 0x01 };
804 void __iomem *lane_base = phy->lane_base;
806 if (phy->cfg->quirks & DSI_PHY_10NM_QUIRK_OLD_TIMINGS)
809 /* Strength ctrl settings */
810 for (i = 0; i < 5; i++) {
811 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_LPTX_STR_CTRL(i),
814 * Disable LPRX and CDRX for all lanes. And later on, it will
815 * be only enabled for the physical data lane corresponding
816 * to the logical data lane 0
818 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_LPRX_CTRL(i), 0);
819 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_PIN_SWAP(i), 0x0);
820 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_HSTX_STR_CTRL(i),
824 dsi_phy_hw_v3_0_config_lpcdrx(phy, true);
827 for (i = 0; i < 5; i++) {
828 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_CFG0(i), 0x0);
829 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_CFG1(i), 0x0);
830 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_CFG2(i), 0x0);
831 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_CFG3(i),
832 i == 4 ? 0x80 : 0x0);
833 dsi_phy_write(lane_base +
834 REG_DSI_10nm_PHY_LN_OFFSET_TOP_CTRL(i), 0x0);
835 dsi_phy_write(lane_base +
836 REG_DSI_10nm_PHY_LN_OFFSET_BOT_CTRL(i), 0x0);
837 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_TX_DCTRL(i),
841 if (!(phy->cfg->quirks & DSI_PHY_10NM_QUIRK_OLD_TIMINGS)) {
842 /* Toggle BIT 0 to release freeze I/0 */
843 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_TX_DCTRL(3), 0x05);
844 dsi_phy_write(lane_base + REG_DSI_10nm_PHY_LN_TX_DCTRL(3), 0x04);
848 static int dsi_10nm_phy_enable(struct msm_dsi_phy *phy, int src_pll_id,
849 struct msm_dsi_phy_clk_request *clk_req)
853 u32 const delay_us = 5;
854 u32 const timeout_us = 1000;
855 struct msm_dsi_dphy_timing *timing = &phy->timing;
856 void __iomem *base = phy->base;
861 if (msm_dsi_dphy_timing_calc_v3(timing, clk_req)) {
862 DRM_DEV_ERROR(&phy->pdev->dev,
863 "%s: D-PHY timing calculation failed\n", __func__);
867 if (dsi_phy_hw_v3_0_is_pll_on(phy))
868 pr_warn("PLL turned on before configuring PHY\n");
870 /* wait for REFGEN READY */
871 ret = readl_poll_timeout_atomic(base + REG_DSI_10nm_PHY_CMN_PHY_STATUS,
872 status, (status & BIT(0)),
873 delay_us, timeout_us);
875 pr_err("Ref gen not ready. Aborting\n");
879 /* de-assert digital and pll power down */
880 data = BIT(6) | BIT(5);
881 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CTRL_0, data);
883 /* Assert PLL core reset */
884 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_PLL_CNTRL, 0x00);
886 /* turn off resync FIFO */
887 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_RBUF_CTRL, 0x00);
889 /* Select MS1 byte-clk */
890 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_GLBL_CTRL, 0x10);
893 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_VREG_CTRL, 0x59);
895 /* Configure PHY lane swap (TODO: we need to calculate this) */
896 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_LANE_CFG0, 0x21);
897 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_LANE_CFG1, 0x84);
899 /* DSI PHY timings */
900 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_0,
901 timing->hs_halfbyte_en);
902 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_1,
904 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_2,
905 timing->clk_prepare);
906 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_3,
908 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_4,
910 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_5,
912 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_6,
914 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_7,
916 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_8,
918 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_9,
919 timing->ta_go | (timing->ta_sure << 3));
920 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_10,
922 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_TIMING_CTRL_11,
925 /* Remove power down from all blocks */
926 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CTRL_0, 0x7f);
929 data = dsi_phy_read(base + REG_DSI_10nm_PHY_CMN_CTRL_0);
931 /* TODO: only power up lanes that are used */
933 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CTRL_0, data);
934 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_LANE_CTRL0, 0x1F);
936 /* Select full-rate mode */
937 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CTRL_2, 0x40);
939 ret = dsi_10nm_set_usecase(phy);
941 DRM_DEV_ERROR(&phy->pdev->dev, "%s: set pll usecase failed, %d\n",
946 /* DSI lane settings */
947 dsi_phy_hw_v3_0_lane_settings(phy);
949 DBG("DSI%d PHY enabled", phy->id);
954 static void dsi_10nm_phy_disable(struct msm_dsi_phy *phy)
956 void __iomem *base = phy->base;
961 if (dsi_phy_hw_v3_0_is_pll_on(phy))
962 pr_warn("Turning OFF PHY while PLL is on\n");
964 dsi_phy_hw_v3_0_config_lpcdrx(phy, false);
965 data = dsi_phy_read(base + REG_DSI_10nm_PHY_CMN_CTRL_0);
967 /* disable all lanes */
969 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CTRL_0, data);
970 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_LANE_CTRL0, 0);
972 /* Turn off all PHY blocks */
973 dsi_phy_write(base + REG_DSI_10nm_PHY_CMN_CTRL_0, 0x00);
974 /* make sure phy is turned off */
977 DBG("DSI%d PHY disabled", phy->id);
980 const struct msm_dsi_phy_cfg dsi_phy_10nm_cfgs = {
981 .src_pll_truthtable = { {false, false}, {true, false} },
982 .has_phy_lane = true,
990 .enable = dsi_10nm_phy_enable,
991 .disable = dsi_10nm_phy_disable,
992 .pll_init = dsi_pll_10nm_init,
993 .save_pll_state = dsi_10nm_pll_save_state,
994 .restore_pll_state = dsi_10nm_pll_restore_state,
996 .min_pll_rate = 1000000000UL,
997 .max_pll_rate = 3500000000UL,
998 .io_start = { 0xae94400, 0xae96400 },
1002 const struct msm_dsi_phy_cfg dsi_phy_10nm_8998_cfgs = {
1003 .src_pll_truthtable = { {false, false}, {true, false} },
1004 .has_phy_lane = true,
1008 {"vdds", 36000, 32},
1012 .enable = dsi_10nm_phy_enable,
1013 .disable = dsi_10nm_phy_disable,
1014 .pll_init = dsi_pll_10nm_init,
1015 .save_pll_state = dsi_10nm_pll_save_state,
1016 .restore_pll_state = dsi_10nm_pll_restore_state,
1018 .min_pll_rate = 1000000000UL,
1019 .max_pll_rate = 3500000000UL,
1020 .io_start = { 0xc994400, 0xc996400 },
1022 .quirks = DSI_PHY_10NM_QUIRK_OLD_TIMINGS,