Merge tag 'dt-5.15' of git://git.kernel.org/pub/scm/linux/kernel/git/soc/soc
[linux-2.6-microblaze.git] / drivers / bus / mhi / core / boot.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (c) 2018-2020, The Linux Foundation. All rights reserved.
4  *
5  */
6
7 #include <linux/delay.h>
8 #include <linux/device.h>
9 #include <linux/dma-direction.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/firmware.h>
12 #include <linux/interrupt.h>
13 #include <linux/list.h>
14 #include <linux/mhi.h>
15 #include <linux/module.h>
16 #include <linux/random.h>
17 #include <linux/slab.h>
18 #include <linux/wait.h>
19 #include "internal.h"
20
21 /* Setup RDDM vector table for RDDM transfer and program RXVEC */
22 void mhi_rddm_prepare(struct mhi_controller *mhi_cntrl,
23                       struct image_info *img_info)
24 {
25         struct mhi_buf *mhi_buf = img_info->mhi_buf;
26         struct bhi_vec_entry *bhi_vec = img_info->bhi_vec;
27         void __iomem *base = mhi_cntrl->bhie;
28         struct device *dev = &mhi_cntrl->mhi_dev->dev;
29         u32 sequence_id;
30         unsigned int i;
31
32         for (i = 0; i < img_info->entries - 1; i++, mhi_buf++, bhi_vec++) {
33                 bhi_vec->dma_addr = mhi_buf->dma_addr;
34                 bhi_vec->size = mhi_buf->len;
35         }
36
37         dev_dbg(dev, "BHIe programming for RDDM\n");
38
39         mhi_write_reg(mhi_cntrl, base, BHIE_RXVECADDR_HIGH_OFFS,
40                       upper_32_bits(mhi_buf->dma_addr));
41
42         mhi_write_reg(mhi_cntrl, base, BHIE_RXVECADDR_LOW_OFFS,
43                       lower_32_bits(mhi_buf->dma_addr));
44
45         mhi_write_reg(mhi_cntrl, base, BHIE_RXVECSIZE_OFFS, mhi_buf->len);
46         sequence_id = MHI_RANDOM_U32_NONZERO(BHIE_RXVECSTATUS_SEQNUM_BMSK);
47
48         mhi_write_reg_field(mhi_cntrl, base, BHIE_RXVECDB_OFFS,
49                             BHIE_RXVECDB_SEQNUM_BMSK, BHIE_RXVECDB_SEQNUM_SHFT,
50                             sequence_id);
51
52         dev_dbg(dev, "Address: %p and len: 0x%zx sequence: %u\n",
53                 &mhi_buf->dma_addr, mhi_buf->len, sequence_id);
54 }
55
56 /* Collect RDDM buffer during kernel panic */
57 static int __mhi_download_rddm_in_panic(struct mhi_controller *mhi_cntrl)
58 {
59         int ret;
60         u32 rx_status;
61         enum mhi_ee_type ee;
62         const u32 delayus = 2000;
63         u32 retry = (mhi_cntrl->timeout_ms * 1000) / delayus;
64         const u32 rddm_timeout_us = 200000;
65         int rddm_retry = rddm_timeout_us / delayus;
66         void __iomem *base = mhi_cntrl->bhie;
67         struct device *dev = &mhi_cntrl->mhi_dev->dev;
68
69         dev_dbg(dev, "Entered with pm_state:%s dev_state:%s ee:%s\n",
70                 to_mhi_pm_state_str(mhi_cntrl->pm_state),
71                 TO_MHI_STATE_STR(mhi_cntrl->dev_state),
72                 TO_MHI_EXEC_STR(mhi_cntrl->ee));
73
74         /*
75          * This should only be executing during a kernel panic, we expect all
76          * other cores to shutdown while we're collecting RDDM buffer. After
77          * returning from this function, we expect the device to reset.
78          *
79          * Normaly, we read/write pm_state only after grabbing the
80          * pm_lock, since we're in a panic, skipping it. Also there is no
81          * gurantee that this state change would take effect since
82          * we're setting it w/o grabbing pm_lock
83          */
84         mhi_cntrl->pm_state = MHI_PM_LD_ERR_FATAL_DETECT;
85         /* update should take the effect immediately */
86         smp_wmb();
87
88         /*
89          * Make sure device is not already in RDDM. In case the device asserts
90          * and a kernel panic follows, device will already be in RDDM.
91          * Do not trigger SYS ERR again and proceed with waiting for
92          * image download completion.
93          */
94         ee = mhi_get_exec_env(mhi_cntrl);
95         if (ee == MHI_EE_MAX)
96                 goto error_exit_rddm;
97
98         if (ee != MHI_EE_RDDM) {
99                 dev_dbg(dev, "Trigger device into RDDM mode using SYS ERR\n");
100                 mhi_set_mhi_state(mhi_cntrl, MHI_STATE_SYS_ERR);
101
102                 dev_dbg(dev, "Waiting for device to enter RDDM\n");
103                 while (rddm_retry--) {
104                         ee = mhi_get_exec_env(mhi_cntrl);
105                         if (ee == MHI_EE_RDDM)
106                                 break;
107
108                         udelay(delayus);
109                 }
110
111                 if (rddm_retry <= 0) {
112                         /* Hardware reset so force device to enter RDDM */
113                         dev_dbg(dev,
114                                 "Did not enter RDDM, do a host req reset\n");
115                         mhi_write_reg(mhi_cntrl, mhi_cntrl->regs,
116                                       MHI_SOC_RESET_REQ_OFFSET,
117                                       MHI_SOC_RESET_REQ);
118                         udelay(delayus);
119                 }
120
121                 ee = mhi_get_exec_env(mhi_cntrl);
122         }
123
124         dev_dbg(dev,
125                 "Waiting for RDDM image download via BHIe, current EE:%s\n",
126                 TO_MHI_EXEC_STR(ee));
127
128         while (retry--) {
129                 ret = mhi_read_reg_field(mhi_cntrl, base, BHIE_RXVECSTATUS_OFFS,
130                                          BHIE_RXVECSTATUS_STATUS_BMSK,
131                                          BHIE_RXVECSTATUS_STATUS_SHFT,
132                                          &rx_status);
133                 if (ret)
134                         return -EIO;
135
136                 if (rx_status == BHIE_RXVECSTATUS_STATUS_XFER_COMPL)
137                         return 0;
138
139                 udelay(delayus);
140         }
141
142         ee = mhi_get_exec_env(mhi_cntrl);
143         ret = mhi_read_reg(mhi_cntrl, base, BHIE_RXVECSTATUS_OFFS, &rx_status);
144
145         dev_err(dev, "RXVEC_STATUS: 0x%x\n", rx_status);
146
147 error_exit_rddm:
148         dev_err(dev, "RDDM transfer failed. Current EE: %s\n",
149                 TO_MHI_EXEC_STR(ee));
150
151         return -EIO;
152 }
153
154 /* Download RDDM image from device */
155 int mhi_download_rddm_image(struct mhi_controller *mhi_cntrl, bool in_panic)
156 {
157         void __iomem *base = mhi_cntrl->bhie;
158         struct device *dev = &mhi_cntrl->mhi_dev->dev;
159         u32 rx_status;
160
161         if (in_panic)
162                 return __mhi_download_rddm_in_panic(mhi_cntrl);
163
164         dev_dbg(dev, "Waiting for RDDM image download via BHIe\n");
165
166         /* Wait for the image download to complete */
167         wait_event_timeout(mhi_cntrl->state_event,
168                            mhi_read_reg_field(mhi_cntrl, base,
169                                               BHIE_RXVECSTATUS_OFFS,
170                                               BHIE_RXVECSTATUS_STATUS_BMSK,
171                                               BHIE_RXVECSTATUS_STATUS_SHFT,
172                                               &rx_status) || rx_status,
173                            msecs_to_jiffies(mhi_cntrl->timeout_ms));
174
175         return (rx_status == BHIE_RXVECSTATUS_STATUS_XFER_COMPL) ? 0 : -EIO;
176 }
177 EXPORT_SYMBOL_GPL(mhi_download_rddm_image);
178
179 static int mhi_fw_load_bhie(struct mhi_controller *mhi_cntrl,
180                             const struct mhi_buf *mhi_buf)
181 {
182         void __iomem *base = mhi_cntrl->bhie;
183         struct device *dev = &mhi_cntrl->mhi_dev->dev;
184         rwlock_t *pm_lock = &mhi_cntrl->pm_lock;
185         u32 tx_status, sequence_id;
186         int ret;
187
188         read_lock_bh(pm_lock);
189         if (!MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
190                 read_unlock_bh(pm_lock);
191                 return -EIO;
192         }
193
194         sequence_id = MHI_RANDOM_U32_NONZERO(BHIE_TXVECSTATUS_SEQNUM_BMSK);
195         dev_dbg(dev, "Starting image download via BHIe. Sequence ID: %u\n",
196                 sequence_id);
197         mhi_write_reg(mhi_cntrl, base, BHIE_TXVECADDR_HIGH_OFFS,
198                       upper_32_bits(mhi_buf->dma_addr));
199
200         mhi_write_reg(mhi_cntrl, base, BHIE_TXVECADDR_LOW_OFFS,
201                       lower_32_bits(mhi_buf->dma_addr));
202
203         mhi_write_reg(mhi_cntrl, base, BHIE_TXVECSIZE_OFFS, mhi_buf->len);
204
205         mhi_write_reg_field(mhi_cntrl, base, BHIE_TXVECDB_OFFS,
206                             BHIE_TXVECDB_SEQNUM_BMSK, BHIE_TXVECDB_SEQNUM_SHFT,
207                             sequence_id);
208         read_unlock_bh(pm_lock);
209
210         /* Wait for the image download to complete */
211         ret = wait_event_timeout(mhi_cntrl->state_event,
212                                  MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
213                                  mhi_read_reg_field(mhi_cntrl, base,
214                                                    BHIE_TXVECSTATUS_OFFS,
215                                                    BHIE_TXVECSTATUS_STATUS_BMSK,
216                                                    BHIE_TXVECSTATUS_STATUS_SHFT,
217                                                    &tx_status) || tx_status,
218                                  msecs_to_jiffies(mhi_cntrl->timeout_ms));
219         if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
220             tx_status != BHIE_TXVECSTATUS_STATUS_XFER_COMPL)
221                 return -EIO;
222
223         return (!ret) ? -ETIMEDOUT : 0;
224 }
225
226 static int mhi_fw_load_bhi(struct mhi_controller *mhi_cntrl,
227                            dma_addr_t dma_addr,
228                            size_t size)
229 {
230         u32 tx_status, val, session_id;
231         int i, ret;
232         void __iomem *base = mhi_cntrl->bhi;
233         rwlock_t *pm_lock = &mhi_cntrl->pm_lock;
234         struct device *dev = &mhi_cntrl->mhi_dev->dev;
235         struct {
236                 char *name;
237                 u32 offset;
238         } error_reg[] = {
239                 { "ERROR_CODE", BHI_ERRCODE },
240                 { "ERROR_DBG1", BHI_ERRDBG1 },
241                 { "ERROR_DBG2", BHI_ERRDBG2 },
242                 { "ERROR_DBG3", BHI_ERRDBG3 },
243                 { NULL },
244         };
245
246         read_lock_bh(pm_lock);
247         if (!MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
248                 read_unlock_bh(pm_lock);
249                 goto invalid_pm_state;
250         }
251
252         session_id = MHI_RANDOM_U32_NONZERO(BHI_TXDB_SEQNUM_BMSK);
253         dev_dbg(dev, "Starting image download via BHI. Session ID: %u\n",
254                 session_id);
255         mhi_write_reg(mhi_cntrl, base, BHI_STATUS, 0);
256         mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_HIGH,
257                       upper_32_bits(dma_addr));
258         mhi_write_reg(mhi_cntrl, base, BHI_IMGADDR_LOW,
259                       lower_32_bits(dma_addr));
260         mhi_write_reg(mhi_cntrl, base, BHI_IMGSIZE, size);
261         mhi_write_reg(mhi_cntrl, base, BHI_IMGTXDB, session_id);
262         read_unlock_bh(pm_lock);
263
264         /* Wait for the image download to complete */
265         ret = wait_event_timeout(mhi_cntrl->state_event,
266                            MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state) ||
267                            mhi_read_reg_field(mhi_cntrl, base, BHI_STATUS,
268                                               BHI_STATUS_MASK, BHI_STATUS_SHIFT,
269                                               &tx_status) || tx_status,
270                            msecs_to_jiffies(mhi_cntrl->timeout_ms));
271         if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state))
272                 goto invalid_pm_state;
273
274         if (tx_status == BHI_STATUS_ERROR) {
275                 dev_err(dev, "Image transfer failed\n");
276                 read_lock_bh(pm_lock);
277                 if (MHI_REG_ACCESS_VALID(mhi_cntrl->pm_state)) {
278                         for (i = 0; error_reg[i].name; i++) {
279                                 ret = mhi_read_reg(mhi_cntrl, base,
280                                                    error_reg[i].offset, &val);
281                                 if (ret)
282                                         break;
283                                 dev_err(dev, "Reg: %s value: 0x%x\n",
284                                         error_reg[i].name, val);
285                         }
286                 }
287                 read_unlock_bh(pm_lock);
288                 goto invalid_pm_state;
289         }
290
291         return (!ret) ? -ETIMEDOUT : 0;
292
293 invalid_pm_state:
294
295         return -EIO;
296 }
297
298 void mhi_free_bhie_table(struct mhi_controller *mhi_cntrl,
299                          struct image_info *image_info)
300 {
301         int i;
302         struct mhi_buf *mhi_buf = image_info->mhi_buf;
303
304         for (i = 0; i < image_info->entries; i++, mhi_buf++)
305                 dma_free_coherent(mhi_cntrl->cntrl_dev, mhi_buf->len,
306                                   mhi_buf->buf, mhi_buf->dma_addr);
307
308         kfree(image_info->mhi_buf);
309         kfree(image_info);
310 }
311
312 int mhi_alloc_bhie_table(struct mhi_controller *mhi_cntrl,
313                          struct image_info **image_info,
314                          size_t alloc_size)
315 {
316         size_t seg_size = mhi_cntrl->seg_len;
317         int segments = DIV_ROUND_UP(alloc_size, seg_size) + 1;
318         int i;
319         struct image_info *img_info;
320         struct mhi_buf *mhi_buf;
321
322         img_info = kzalloc(sizeof(*img_info), GFP_KERNEL);
323         if (!img_info)
324                 return -ENOMEM;
325
326         /* Allocate memory for entries */
327         img_info->mhi_buf = kcalloc(segments, sizeof(*img_info->mhi_buf),
328                                     GFP_KERNEL);
329         if (!img_info->mhi_buf)
330                 goto error_alloc_mhi_buf;
331
332         /* Allocate and populate vector table */
333         mhi_buf = img_info->mhi_buf;
334         for (i = 0; i < segments; i++, mhi_buf++) {
335                 size_t vec_size = seg_size;
336
337                 /* Vector table is the last entry */
338                 if (i == segments - 1)
339                         vec_size = sizeof(struct bhi_vec_entry) * i;
340
341                 mhi_buf->len = vec_size;
342                 mhi_buf->buf = dma_alloc_coherent(mhi_cntrl->cntrl_dev,
343                                                   vec_size, &mhi_buf->dma_addr,
344                                                   GFP_KERNEL);
345                 if (!mhi_buf->buf)
346                         goto error_alloc_segment;
347         }
348
349         img_info->bhi_vec = img_info->mhi_buf[segments - 1].buf;
350         img_info->entries = segments;
351         *image_info = img_info;
352
353         return 0;
354
355 error_alloc_segment:
356         for (--i, --mhi_buf; i >= 0; i--, mhi_buf--)
357                 dma_free_coherent(mhi_cntrl->cntrl_dev, mhi_buf->len,
358                                   mhi_buf->buf, mhi_buf->dma_addr);
359
360 error_alloc_mhi_buf:
361         kfree(img_info);
362
363         return -ENOMEM;
364 }
365
366 static void mhi_firmware_copy(struct mhi_controller *mhi_cntrl,
367                               const struct firmware *firmware,
368                               struct image_info *img_info)
369 {
370         size_t remainder = firmware->size;
371         size_t to_cpy;
372         const u8 *buf = firmware->data;
373         struct mhi_buf *mhi_buf = img_info->mhi_buf;
374         struct bhi_vec_entry *bhi_vec = img_info->bhi_vec;
375
376         while (remainder) {
377                 to_cpy = min(remainder, mhi_buf->len);
378                 memcpy(mhi_buf->buf, buf, to_cpy);
379                 bhi_vec->dma_addr = mhi_buf->dma_addr;
380                 bhi_vec->size = to_cpy;
381
382                 buf += to_cpy;
383                 remainder -= to_cpy;
384                 bhi_vec++;
385                 mhi_buf++;
386         }
387 }
388
389 void mhi_fw_load_handler(struct mhi_controller *mhi_cntrl)
390 {
391         const struct firmware *firmware = NULL;
392         struct device *dev = &mhi_cntrl->mhi_dev->dev;
393         const char *fw_name;
394         void *buf;
395         dma_addr_t dma_addr;
396         size_t size;
397         int i, ret;
398
399         if (MHI_PM_IN_ERROR_STATE(mhi_cntrl->pm_state)) {
400                 dev_err(dev, "Device MHI is not in valid state\n");
401                 return;
402         }
403
404         /* save hardware info from BHI */
405         ret = mhi_read_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_SERIALNU,
406                            &mhi_cntrl->serial_number);
407         if (ret)
408                 dev_err(dev, "Could not capture serial number via BHI\n");
409
410         for (i = 0; i < ARRAY_SIZE(mhi_cntrl->oem_pk_hash); i++) {
411                 ret = mhi_read_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_OEMPKHASH(i),
412                                    &mhi_cntrl->oem_pk_hash[i]);
413                 if (ret) {
414                         dev_err(dev, "Could not capture OEM PK HASH via BHI\n");
415                         break;
416                 }
417         }
418
419         /* wait for ready on pass through or any other execution environment */
420         if (mhi_cntrl->ee != MHI_EE_EDL && mhi_cntrl->ee != MHI_EE_PBL)
421                 goto fw_load_ready_state;
422
423         fw_name = (mhi_cntrl->ee == MHI_EE_EDL) ?
424                 mhi_cntrl->edl_image : mhi_cntrl->fw_image;
425
426         if (!fw_name || (mhi_cntrl->fbc_download && (!mhi_cntrl->sbl_size ||
427                                                      !mhi_cntrl->seg_len))) {
428                 dev_err(dev,
429                         "No firmware image defined or !sbl_size || !seg_len\n");
430                 goto error_fw_load;
431         }
432
433         ret = request_firmware(&firmware, fw_name, dev);
434         if (ret) {
435                 dev_err(dev, "Error loading firmware: %d\n", ret);
436                 goto error_fw_load;
437         }
438
439         size = (mhi_cntrl->fbc_download) ? mhi_cntrl->sbl_size : firmware->size;
440
441         /* SBL size provided is maximum size, not necessarily the image size */
442         if (size > firmware->size)
443                 size = firmware->size;
444
445         buf = dma_alloc_coherent(mhi_cntrl->cntrl_dev, size, &dma_addr,
446                                  GFP_KERNEL);
447         if (!buf) {
448                 release_firmware(firmware);
449                 goto error_fw_load;
450         }
451
452         /* Download image using BHI */
453         memcpy(buf, firmware->data, size);
454         ret = mhi_fw_load_bhi(mhi_cntrl, dma_addr, size);
455         dma_free_coherent(mhi_cntrl->cntrl_dev, size, buf, dma_addr);
456
457         /* Error or in EDL mode, we're done */
458         if (ret) {
459                 dev_err(dev, "MHI did not load image over BHI, ret: %d\n", ret);
460                 release_firmware(firmware);
461                 goto error_fw_load;
462         }
463
464         /* Wait for ready since EDL image was loaded */
465         if (fw_name == mhi_cntrl->edl_image) {
466                 release_firmware(firmware);
467                 goto fw_load_ready_state;
468         }
469
470         write_lock_irq(&mhi_cntrl->pm_lock);
471         mhi_cntrl->dev_state = MHI_STATE_RESET;
472         write_unlock_irq(&mhi_cntrl->pm_lock);
473
474         /*
475          * If we're doing fbc, populate vector tables while
476          * device transitioning into MHI READY state
477          */
478         if (mhi_cntrl->fbc_download) {
479                 ret = mhi_alloc_bhie_table(mhi_cntrl, &mhi_cntrl->fbc_image,
480                                            firmware->size);
481                 if (ret) {
482                         release_firmware(firmware);
483                         goto error_fw_load;
484                 }
485
486                 /* Load the firmware into BHIE vec table */
487                 mhi_firmware_copy(mhi_cntrl, firmware, mhi_cntrl->fbc_image);
488         }
489
490         release_firmware(firmware);
491
492 fw_load_ready_state:
493         /* Transitioning into MHI RESET->READY state */
494         ret = mhi_ready_state_transition(mhi_cntrl);
495         if (ret) {
496                 dev_err(dev, "MHI did not enter READY state\n");
497                 goto error_ready_state;
498         }
499
500         dev_info(dev, "Wait for device to enter SBL or Mission mode\n");
501         return;
502
503 error_ready_state:
504         if (mhi_cntrl->fbc_download) {
505                 mhi_free_bhie_table(mhi_cntrl, mhi_cntrl->fbc_image);
506                 mhi_cntrl->fbc_image = NULL;
507         }
508
509 error_fw_load:
510         mhi_cntrl->pm_state = MHI_PM_FW_DL_ERR;
511         wake_up_all(&mhi_cntrl->state_event);
512 }
513
514 int mhi_download_amss_image(struct mhi_controller *mhi_cntrl)
515 {
516         struct image_info *image_info = mhi_cntrl->fbc_image;
517         struct device *dev = &mhi_cntrl->mhi_dev->dev;
518         int ret;
519
520         if (!image_info)
521                 return -EIO;
522
523         ret = mhi_fw_load_bhie(mhi_cntrl,
524                                /* Vector table is the last entry */
525                                &image_info->mhi_buf[image_info->entries - 1]);
526         if (ret) {
527                 dev_err(dev, "MHI did not load AMSS, ret:%d\n", ret);
528                 mhi_cntrl->pm_state = MHI_PM_FW_DL_ERR;
529                 wake_up_all(&mhi_cntrl->state_event);
530         }
531
532         return ret;
533 }