1 // SPDX-License-Identifier: GPL-2.0
3 // Copyright (C) 2022 Microchip Technology Inc.
4 // Authors: Tharun Kumar P <tharunkumar.pasumarthi@microchip.com>
5 // Kumaravel Thiagarajan <Kumaravel.Thiagarajan@microchip.com>
8 #include <linux/dma-mapping.h>
9 #include <linux/iopoll.h>
10 #include <linux/irq.h>
11 #include <linux/module.h>
12 #include <linux/msi.h>
13 #include <linux/pci_regs.h>
14 #include <linux/pci.h>
15 #include <linux/spi/spi.h>
16 #include <linux/delay.h>
18 #define DRV_NAME "spi-pci1xxxx"
20 #define SYS_FREQ_DEFAULT (62500000)
22 #define PCI1XXXX_SPI_MAX_CLOCK_HZ (30000000)
23 #define PCI1XXXX_SPI_CLK_20MHZ (20000000)
24 #define PCI1XXXX_SPI_CLK_15MHZ (15000000)
25 #define PCI1XXXX_SPI_CLK_12MHZ (12000000)
26 #define PCI1XXXX_SPI_CLK_10MHZ (10000000)
27 #define PCI1XXXX_SPI_MIN_CLOCK_HZ (2000000)
29 #define PCI1XXXX_SPI_BUFFER_SIZE (320)
31 #define SPI_MST_CTL_DEVSEL_MASK (GENMASK(27, 25))
32 #define SPI_MST_CTL_CMD_LEN_MASK (GENMASK(16, 8))
33 #define SPI_MST_CTL_SPEED_MASK (GENMASK(7, 5))
34 #define SPI_MSI_VECTOR_SEL_MASK (GENMASK(4, 4))
36 #define SPI_MST_CTL_FORCE_CE (BIT(4))
37 #define SPI_MST_CTL_MODE_SEL (BIT(2))
38 #define SPI_MST_CTL_GO (BIT(0))
40 #define SPI_SYSTEM_ADDR_BASE (0x2000)
41 #define SPI_MST1_ADDR_BASE (0x800)
43 #define DEV_REV_REG (SPI_SYSTEM_ADDR_BASE + 0x00)
44 #define SPI_SYSLOCK_REG (SPI_SYSTEM_ADDR_BASE + 0xA0)
45 #define SPI_CONFIG_PERI_ENABLE_REG (SPI_SYSTEM_ADDR_BASE + 0x108)
47 #define SPI_PERI_ENBLE_PF_MASK (GENMASK(17, 16))
48 #define DEV_REV_MASK (GENMASK(7, 0))
50 #define SPI_SYSLOCK BIT(4)
52 /* DMA Related Registers */
53 #define SPI_DMA_ADDR_BASE (0x1000)
54 #define SPI_DMA_GLOBAL_WR_ENGINE_EN (SPI_DMA_ADDR_BASE + 0x0C)
55 #define SPI_DMA_GLOBAL_RD_ENGINE_EN (SPI_DMA_ADDR_BASE + 0x2C)
56 #define SPI_DMA_INTR_IMWR_WDONE_LOW (SPI_DMA_ADDR_BASE + 0x60)
57 #define SPI_DMA_INTR_IMWR_WDONE_HIGH (SPI_DMA_ADDR_BASE + 0x64)
58 #define SPI_DMA_INTR_IMWR_WABORT_LOW (SPI_DMA_ADDR_BASE + 0x68)
59 #define SPI_DMA_INTR_IMWR_WABORT_HIGH (SPI_DMA_ADDR_BASE + 0x6C)
60 #define SPI_DMA_INTR_WR_IMWR_DATA (SPI_DMA_ADDR_BASE + 0x70)
61 #define SPI_DMA_INTR_IMWR_RDONE_LOW (SPI_DMA_ADDR_BASE + 0xCC)
62 #define SPI_DMA_INTR_IMWR_RDONE_HIGH (SPI_DMA_ADDR_BASE + 0xD0)
63 #define SPI_DMA_INTR_IMWR_RABORT_LOW (SPI_DMA_ADDR_BASE + 0xD4)
64 #define SPI_DMA_INTR_IMWR_RABORT_HIGH (SPI_DMA_ADDR_BASE + 0xD8)
65 #define SPI_DMA_INTR_RD_IMWR_DATA (SPI_DMA_ADDR_BASE + 0xDC)
67 /* x refers to SPI Host Controller HW instance id in the below macros - 0 or 1 */
69 #define SPI_MST_CMD_BUF_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x00)
70 #define SPI_MST_RSP_BUF_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x200)
71 #define SPI_MST_CTL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x400)
72 #define SPI_MST_EVENT_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x420)
73 #define SPI_MST_EVENT_MASK_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x424)
74 #define SPI_MST_PAD_CTL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x460)
75 #define SPIALERT_MST_DB_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x464)
76 #define SPIALERT_MST_VAL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x468)
77 #define SPI_PCI_CTRL_REG_OFFSET(x) (((x) * SPI_MST1_ADDR_BASE) + 0x480)
79 #define PCI1XXXX_IRQ_FLAGS (IRQF_NO_SUSPEND | IRQF_TRIGGER_NONE)
80 #define SPI_MAX_DATA_LEN 320
82 #define PCI1XXXX_SPI_TIMEOUT (msecs_to_jiffies(100))
83 #define SYSLOCK_RETRY_CNT (1000)
84 #define SPI_DMA_ENGINE_EN (0x1)
86 #define SPI_INTR BIT(8)
87 #define SPI_FORCE_CE BIT(4)
89 #define SPI_CHIP_SEL_COUNT 7
90 #define VENDOR_ID_MCHP 0x1055
92 #define SPI_SUSPEND_CONFIG 0x101
93 #define SPI_RESUME_CONFIG 0x203
95 struct pci1xxxx_spi_internal {
97 bool spi_xfer_in_progress;
99 struct completion spi_xfer_done;
100 struct spi_controller *spi_host;
101 struct pci1xxxx_spi *parent;
103 unsigned int dev_sel : 3;
104 unsigned int msi_vector_sel : 1;
108 struct pci1xxxx_spi {
110 u8 total_hw_instances;
112 void __iomem *reg_base;
113 void __iomem *dma_offset_bar;
115 struct pci1xxxx_spi_internal *spi_int[] __counted_by(total_hw_instances);
118 static const struct pci_device_id pci1xxxx_spi_pci_id_table[] = {
119 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
120 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
121 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
122 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
123 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
124 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
125 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
126 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
127 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
128 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
129 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
130 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
131 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
132 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
133 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
134 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
135 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
136 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
137 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
138 { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
142 MODULE_DEVICE_TABLE(pci, pci1xxxx_spi_pci_id_table);
144 static int pci1xxxx_set_sys_lock(struct pci1xxxx_spi *par)
146 writel(SPI_SYSLOCK, par->reg_base + SPI_SYSLOCK_REG);
147 return readl(par->reg_base + SPI_SYSLOCK_REG);
150 static int pci1xxxx_acquire_sys_lock(struct pci1xxxx_spi *par)
154 return readx_poll_timeout(pci1xxxx_set_sys_lock, par, regval,
155 (regval & SPI_SYSLOCK), 100,
156 SYSLOCK_RETRY_CNT * 100);
159 static void pci1xxxx_release_sys_lock(struct pci1xxxx_spi *par)
161 writel(0x0, par->reg_base + SPI_SYSLOCK_REG);
164 static int pci1xxxx_check_spi_can_dma(struct pci1xxxx_spi *spi_bus, int irq)
166 struct pci_dev *pdev = spi_bus->dev;
172 * DEV REV Registers is a system register, HW Syslock bit
173 * should be acquired before accessing the register
175 ret = pci1xxxx_acquire_sys_lock(spi_bus);
177 dev_err(&pdev->dev, "Error failed to acquire syslock\n");
181 regval = readl(spi_bus->reg_base + DEV_REV_REG);
182 spi_bus->dev_rev = regval & DEV_REV_MASK;
183 if (spi_bus->dev_rev >= 0xC0) {
184 regval = readl(spi_bus->reg_base +
185 SPI_CONFIG_PERI_ENABLE_REG);
186 pf_num = regval & SPI_PERI_ENBLE_PF_MASK;
189 pci1xxxx_release_sys_lock(spi_bus);
192 * DMA is supported only from C0 and SPI can use DMA only if
193 * it is mapped to PF0
195 if (spi_bus->dev_rev < 0xC0 || pf_num)
199 * DMA Supported only with MSI Interrupts
200 * One of the SPI instance's MSI vector address and data
201 * is used for DMA Interrupt
203 if (!irq_get_msi_desc(irq)) {
204 dev_warn(&pdev->dev, "Error MSI Interrupt not supported, will operate in PIO mode\n");
208 spi_bus->dma_offset_bar = pcim_iomap(pdev, 2, pci_resource_len(pdev, 2));
209 if (!spi_bus->dma_offset_bar) {
210 dev_warn(&pdev->dev, "Error failed to map dma bar, will operate in PIO mode\n");
214 if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64))) {
215 dev_warn(&pdev->dev, "Error failed to set DMA mask, will operate in PIO mode\n");
216 pcim_iounmap(pdev, spi_bus->dma_offset_bar);
217 spi_bus->dma_offset_bar = NULL;
224 static int pci1xxxx_spi_dma_init(struct pci1xxxx_spi *spi_bus, int irq)
229 ret = pci1xxxx_check_spi_can_dma(spi_bus, irq);
233 get_cached_msi_msg(irq, &msi);
234 writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
235 writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
236 writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_HIGH);
237 writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_HIGH);
238 writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_HIGH);
239 writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_HIGH);
240 writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_LOW);
241 writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_LOW);
242 writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_LOW);
243 writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_LOW);
244 writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
245 writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
246 spi_bus->can_dma = true;
250 static void pci1xxxx_spi_set_cs(struct spi_device *spi, bool enable)
252 struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi->controller);
253 struct pci1xxxx_spi *par = p->parent;
256 /* Set the DEV_SEL bits of the SPI_MST_CTL_REG */
257 regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
259 regval |= SPI_FORCE_CE;
260 regval &= ~SPI_MST_CTL_DEVSEL_MASK;
261 regval |= (spi_get_chipselect(spi, 0) << 25);
263 regval &= ~SPI_FORCE_CE;
265 writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
268 static u8 pci1xxxx_get_clock_div(u32 hz)
272 if (hz >= PCI1XXXX_SPI_MAX_CLOCK_HZ)
274 else if ((hz < PCI1XXXX_SPI_MAX_CLOCK_HZ) && (hz >= PCI1XXXX_SPI_CLK_20MHZ))
276 else if ((hz < PCI1XXXX_SPI_CLK_20MHZ) && (hz >= PCI1XXXX_SPI_CLK_15MHZ))
278 else if ((hz < PCI1XXXX_SPI_CLK_15MHZ) && (hz >= PCI1XXXX_SPI_CLK_12MHZ))
280 else if ((hz < PCI1XXXX_SPI_CLK_12MHZ) && (hz >= PCI1XXXX_SPI_CLK_10MHZ))
282 else if ((hz < PCI1XXXX_SPI_CLK_10MHZ) && (hz >= PCI1XXXX_SPI_MIN_CLOCK_HZ))
290 static int pci1xxxx_spi_transfer_one(struct spi_controller *spi_ctlr,
291 struct spi_device *spi, struct spi_transfer *xfer)
293 struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
294 int mode, len, loop_iter, transfer_len;
295 struct pci1xxxx_spi *par = p->parent;
296 unsigned long bytes_transfered;
297 unsigned long bytes_recvd;
298 unsigned long loop_count;
304 p->spi_xfer_in_progress = true;
306 clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
307 tx_buf = xfer->tx_buf;
308 rx_buf = xfer->rx_buf;
309 transfer_len = xfer->len;
310 regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
311 writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
314 bytes_transfered = 0;
316 loop_count = transfer_len / SPI_MAX_DATA_LEN;
317 if (transfer_len % SPI_MAX_DATA_LEN != 0)
320 for (loop_iter = 0; loop_iter < loop_count; loop_iter++) {
321 len = SPI_MAX_DATA_LEN;
322 if ((transfer_len % SPI_MAX_DATA_LEN != 0) &&
323 (loop_iter == loop_count - 1))
324 len = transfer_len % SPI_MAX_DATA_LEN;
326 reinit_completion(&p->spi_xfer_done);
327 memcpy_toio(par->reg_base + SPI_MST_CMD_BUF_OFFSET(p->hw_inst),
328 &tx_buf[bytes_transfered], len);
329 bytes_transfered += len;
330 regval = readl(par->reg_base +
331 SPI_MST_CTL_REG_OFFSET(p->hw_inst));
332 regval &= ~(SPI_MST_CTL_MODE_SEL | SPI_MST_CTL_CMD_LEN_MASK |
333 SPI_MST_CTL_SPEED_MASK);
335 if (mode == SPI_MODE_3)
336 regval |= SPI_MST_CTL_MODE_SEL;
338 regval &= ~SPI_MST_CTL_MODE_SEL;
340 regval |= (clkdiv << 5);
341 regval &= ~SPI_MST_CTL_CMD_LEN_MASK;
342 regval |= (len << 8);
343 writel(regval, par->reg_base +
344 SPI_MST_CTL_REG_OFFSET(p->hw_inst));
345 regval = readl(par->reg_base +
346 SPI_MST_CTL_REG_OFFSET(p->hw_inst));
347 regval |= SPI_MST_CTL_GO;
348 writel(regval, par->reg_base +
349 SPI_MST_CTL_REG_OFFSET(p->hw_inst));
351 /* Wait for DMA_TERM interrupt */
352 result = wait_for_completion_timeout(&p->spi_xfer_done,
353 PCI1XXXX_SPI_TIMEOUT);
358 memcpy_fromio(&rx_buf[bytes_recvd], par->reg_base +
359 SPI_MST_RSP_BUF_OFFSET(p->hw_inst), len);
364 p->spi_xfer_in_progress = false;
369 static irqreturn_t pci1xxxx_spi_isr(int irq, void *dev)
371 struct pci1xxxx_spi_internal *p = dev;
372 irqreturn_t spi_int_fired = IRQ_NONE;
375 /* Clear the SPI GO_BIT Interrupt */
376 regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
377 if (regval & SPI_INTR) {
378 /* Clear xfer_done */
379 complete(&p->spi_xfer_done);
380 spi_int_fired = IRQ_HANDLED;
383 writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
385 return spi_int_fired;
388 static int pci1xxxx_spi_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
390 u8 hw_inst_cnt, iter, start, only_sec_inst;
391 struct pci1xxxx_spi_internal *spi_sub_ptr;
392 struct device *dev = &pdev->dev;
393 struct pci1xxxx_spi *spi_bus;
394 struct spi_controller *spi_host;
398 hw_inst_cnt = ent->driver_data & 0x0f;
399 start = (ent->driver_data & 0xf0) >> 4;
405 spi_bus = devm_kzalloc(&pdev->dev,
406 struct_size(spi_bus, spi_int, hw_inst_cnt),
412 spi_bus->total_hw_instances = hw_inst_cnt;
413 pci_set_master(pdev);
415 for (iter = 0; iter < hw_inst_cnt; iter++) {
416 spi_bus->spi_int[iter] = devm_kzalloc(&pdev->dev,
417 sizeof(struct pci1xxxx_spi_internal),
419 spi_sub_ptr = spi_bus->spi_int[iter];
420 spi_sub_ptr->spi_host = devm_spi_alloc_host(dev, sizeof(struct spi_controller));
421 if (!spi_sub_ptr->spi_host)
424 spi_sub_ptr->parent = spi_bus;
425 spi_sub_ptr->spi_xfer_in_progress = false;
428 ret = pcim_enable_device(pdev);
432 ret = pci_request_regions(pdev, DRV_NAME);
436 spi_bus->reg_base = pcim_iomap(pdev, 0, pci_resource_len(pdev, 0));
437 if (!spi_bus->reg_base) {
442 ret = pci_alloc_irq_vectors(pdev, hw_inst_cnt, hw_inst_cnt,
445 dev_err(&pdev->dev, "Error allocating MSI vectors\n");
449 init_completion(&spi_sub_ptr->spi_xfer_done);
450 /* Initialize Interrupts - SPI_INT */
451 regval = readl(spi_bus->reg_base +
452 SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
454 writel(regval, spi_bus->reg_base +
455 SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
456 spi_sub_ptr->irq = pci_irq_vector(pdev, 0);
458 ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
459 pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
460 pci_name(pdev), spi_sub_ptr);
462 dev_err(&pdev->dev, "Unable to request irq : %d",
468 ret = pci1xxxx_spi_dma_init(spi_bus, spi_sub_ptr->irq);
469 if (ret && ret != -EOPNOTSUPP)
472 /* This register is only applicable for 1st instance */
473 regval = readl(spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
479 writel(regval, spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
482 spi_sub_ptr->hw_inst = start++;
485 init_completion(&spi_sub_ptr->spi_xfer_done);
486 /* Initialize Interrupts - SPI_INT */
487 regval = readl(spi_bus->reg_base +
488 SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
490 writel(regval, spi_bus->reg_base +
491 SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
492 spi_sub_ptr->irq = pci_irq_vector(pdev, iter);
493 ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
494 pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
495 pci_name(pdev), spi_sub_ptr);
497 dev_err(&pdev->dev, "Unable to request irq : %d",
504 spi_host = spi_sub_ptr->spi_host;
505 spi_host->num_chipselect = SPI_CHIP_SEL_COUNT;
506 spi_host->mode_bits = SPI_MODE_0 | SPI_MODE_3 | SPI_RX_DUAL |
507 SPI_TX_DUAL | SPI_LOOP;
508 spi_host->transfer_one = pci1xxxx_spi_transfer_one;
509 spi_host->set_cs = pci1xxxx_spi_set_cs;
510 spi_host->bits_per_word_mask = SPI_BPW_MASK(8);
511 spi_host->max_speed_hz = PCI1XXXX_SPI_MAX_CLOCK_HZ;
512 spi_host->min_speed_hz = PCI1XXXX_SPI_MIN_CLOCK_HZ;
513 spi_host->flags = SPI_CONTROLLER_MUST_TX;
514 spi_controller_set_devdata(spi_host, spi_sub_ptr);
515 ret = devm_spi_register_controller(dev, spi_host);
519 pci_set_drvdata(pdev, spi_bus);
524 pci_release_regions(pdev);
528 static void store_restore_config(struct pci1xxxx_spi *spi_ptr,
529 struct pci1xxxx_spi_internal *spi_sub_ptr,
535 regval = readl(spi_ptr->reg_base +
536 SPI_MST_CTL_REG_OFFSET(spi_sub_ptr->hw_inst));
537 regval &= SPI_MST_CTL_DEVSEL_MASK;
538 spi_sub_ptr->prev_val.dev_sel = (regval >> 25) & 7;
539 regval = readl(spi_ptr->reg_base +
540 SPI_PCI_CTRL_REG_OFFSET(spi_sub_ptr->hw_inst));
541 regval &= SPI_MSI_VECTOR_SEL_MASK;
542 spi_sub_ptr->prev_val.msi_vector_sel = (regval >> 4) & 1;
544 regval = readl(spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
545 regval &= ~SPI_MST_CTL_DEVSEL_MASK;
546 regval |= (spi_sub_ptr->prev_val.dev_sel << 25);
548 spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
549 writel((spi_sub_ptr->prev_val.msi_vector_sel << 4),
550 spi_ptr->reg_base + SPI_PCI_CTRL_REG_OFFSET(inst));
554 static int pci1xxxx_spi_resume(struct device *dev)
556 struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
557 struct pci1xxxx_spi_internal *spi_sub_ptr;
558 u32 regval = SPI_RESUME_CONFIG;
561 for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
562 spi_sub_ptr = spi_ptr->spi_int[iter];
563 spi_controller_resume(spi_sub_ptr->spi_host);
564 writel(regval, spi_ptr->reg_base +
565 SPI_MST_EVENT_MASK_REG_OFFSET(iter));
567 /* Restore config at resume */
568 store_restore_config(spi_ptr, spi_sub_ptr, iter, 0);
574 static int pci1xxxx_spi_suspend(struct device *dev)
576 struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
577 struct pci1xxxx_spi_internal *spi_sub_ptr;
578 u32 reg1 = SPI_SUSPEND_CONFIG;
581 for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
582 spi_sub_ptr = spi_ptr->spi_int[iter];
584 while (spi_sub_ptr->spi_xfer_in_progress)
587 /* Store existing config before suspend */
588 store_restore_config(spi_ptr, spi_sub_ptr, iter, 1);
589 spi_controller_suspend(spi_sub_ptr->spi_host);
590 writel(reg1, spi_ptr->reg_base +
591 SPI_MST_EVENT_MASK_REG_OFFSET(iter));
597 static DEFINE_SIMPLE_DEV_PM_OPS(spi_pm_ops, pci1xxxx_spi_suspend,
598 pci1xxxx_spi_resume);
600 static struct pci_driver pci1xxxx_spi_driver = {
602 .id_table = pci1xxxx_spi_pci_id_table,
603 .probe = pci1xxxx_spi_probe,
605 .pm = pm_sleep_ptr(&spi_pm_ops),
609 module_pci_driver(pci1xxxx_spi_driver);
611 MODULE_DESCRIPTION("Microchip Technology Inc. pci1xxxx SPI bus driver");
612 MODULE_AUTHOR("Tharun Kumar P<tharunkumar.pasumarthi@microchip.com>");
613 MODULE_AUTHOR("Kumaravel Thiagarajan<kumaravel.thiagarajan@microchip.com>");
614 MODULE_LICENSE("GPL v2");