]> Git Repo - linux.git/blob - drivers/spi/spi-pci1xxxx.c
x86/kaslr: Expose and use the end of the physical memory address space
[linux.git] / drivers / spi / spi-pci1xxxx.c
1 // SPDX-License-Identifier: GPL-2.0
2 // PCI1xxxx SPI driver
3 // Copyright (C) 2022 Microchip Technology Inc.
4 // Authors: Tharun Kumar P <[email protected]>
5 //          Kumaravel Thiagarajan <[email protected]>
6
7
8 #include <linux/bitfield.h>
9 #include <linux/delay.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/iopoll.h>
12 #include <linux/irq.h>
13 #include <linux/module.h>
14 #include <linux/msi.h>
15 #include <linux/pci_regs.h>
16 #include <linux/pci.h>
17 #include <linux/spinlock.h>
18 #include <linux/spi/spi.h>
19 #include "internals.h"
20
21 #define DRV_NAME "spi-pci1xxxx"
22
23 #define SYS_FREQ_DEFAULT                (62500000)
24
25 #define PCI1XXXX_SPI_MAX_CLOCK_HZ       (30000000)
26 #define PCI1XXXX_SPI_CLK_20MHZ          (20000000)
27 #define PCI1XXXX_SPI_CLK_15MHZ          (15000000)
28 #define PCI1XXXX_SPI_CLK_12MHZ          (12000000)
29 #define PCI1XXXX_SPI_CLK_10MHZ          (10000000)
30 #define PCI1XXXX_SPI_MIN_CLOCK_HZ       (2000000)
31
32 #define PCI1XXXX_SPI_BUFFER_SIZE        (320)
33
34 #define SPI_MST_CTL_DEVSEL_MASK         (GENMASK(27, 25))
35 #define SPI_MST_CTL_CMD_LEN_MASK        (GENMASK(16, 8))
36 #define SPI_MST_CTL_SPEED_MASK          (GENMASK(7, 5))
37 #define SPI_MSI_VECTOR_SEL_MASK         (GENMASK(4, 4))
38
39 #define SPI_MST_CTL_FORCE_CE            (BIT(4))
40 #define SPI_MST_CTL_MODE_SEL            (BIT(2))
41 #define SPI_MST_CTL_GO                  (BIT(0))
42
43 #define SPI_PERI_ADDR_BASE              (0x160000)
44 #define SPI_SYSTEM_ADDR_BASE            (0x2000)
45 #define SPI_MST1_ADDR_BASE              (0x800)
46
47 #define DEV_REV_REG                     (SPI_SYSTEM_ADDR_BASE + 0x00)
48 #define SPI_SYSLOCK_REG                 (SPI_SYSTEM_ADDR_BASE + 0xA0)
49 #define SPI_CONFIG_PERI_ENABLE_REG      (SPI_SYSTEM_ADDR_BASE + 0x108)
50
51 #define SPI_PERI_ENBLE_PF_MASK          (GENMASK(17, 16))
52 #define DEV_REV_MASK                    (GENMASK(7, 0))
53
54 #define SPI_SYSLOCK                     BIT(4)
55 #define SPI0                            (0)
56 #define SPI1                            (1)
57
58 /* DMA Related Registers */
59 #define SPI_DMA_ADDR_BASE               (0x1000)
60 #define SPI_DMA_GLOBAL_WR_ENGINE_EN     (SPI_DMA_ADDR_BASE + 0x0C)
61 #define SPI_DMA_WR_DOORBELL_REG         (SPI_DMA_ADDR_BASE + 0x10)
62 #define SPI_DMA_GLOBAL_RD_ENGINE_EN     (SPI_DMA_ADDR_BASE + 0x2C)
63 #define SPI_DMA_RD_DOORBELL_REG         (SPI_DMA_ADDR_BASE + 0x30)
64 #define SPI_DMA_INTR_WR_STS             (SPI_DMA_ADDR_BASE + 0x4C)
65 #define SPI_DMA_WR_INT_MASK             (SPI_DMA_ADDR_BASE + 0x54)
66 #define SPI_DMA_INTR_WR_CLR             (SPI_DMA_ADDR_BASE + 0x58)
67 #define SPI_DMA_ERR_WR_STS              (SPI_DMA_ADDR_BASE + 0x5C)
68 #define SPI_DMA_INTR_IMWR_WDONE_LOW     (SPI_DMA_ADDR_BASE + 0x60)
69 #define SPI_DMA_INTR_IMWR_WDONE_HIGH    (SPI_DMA_ADDR_BASE + 0x64)
70 #define SPI_DMA_INTR_IMWR_WABORT_LOW    (SPI_DMA_ADDR_BASE + 0x68)
71 #define SPI_DMA_INTR_IMWR_WABORT_HIGH   (SPI_DMA_ADDR_BASE + 0x6C)
72 #define SPI_DMA_INTR_WR_IMWR_DATA       (SPI_DMA_ADDR_BASE + 0x70)
73 #define SPI_DMA_INTR_RD_STS             (SPI_DMA_ADDR_BASE + 0xA0)
74 #define SPI_DMA_RD_INT_MASK             (SPI_DMA_ADDR_BASE + 0xA8)
75 #define SPI_DMA_INTR_RD_CLR             (SPI_DMA_ADDR_BASE + 0xAC)
76 #define SPI_DMA_ERR_RD_STS              (SPI_DMA_ADDR_BASE + 0xB8)
77 #define SPI_DMA_INTR_IMWR_RDONE_LOW     (SPI_DMA_ADDR_BASE + 0xCC)
78 #define SPI_DMA_INTR_IMWR_RDONE_HIGH    (SPI_DMA_ADDR_BASE + 0xD0)
79 #define SPI_DMA_INTR_IMWR_RABORT_LOW    (SPI_DMA_ADDR_BASE + 0xD4)
80 #define SPI_DMA_INTR_IMWR_RABORT_HIGH   (SPI_DMA_ADDR_BASE + 0xD8)
81 #define SPI_DMA_INTR_RD_IMWR_DATA       (SPI_DMA_ADDR_BASE + 0xDC)
82
83 #define SPI_DMA_CH0_WR_BASE             (SPI_DMA_ADDR_BASE + 0x200)
84 #define SPI_DMA_CH0_RD_BASE             (SPI_DMA_ADDR_BASE + 0x300)
85 #define SPI_DMA_CH1_WR_BASE             (SPI_DMA_ADDR_BASE + 0x400)
86 #define SPI_DMA_CH1_RD_BASE             (SPI_DMA_ADDR_BASE + 0x500)
87
88 #define SPI_DMA_CH_CTL1_OFFSET          (0x00)
89 #define SPI_DMA_CH_XFER_LEN_OFFSET      (0x08)
90 #define SPI_DMA_CH_SAR_LO_OFFSET        (0x0C)
91 #define SPI_DMA_CH_SAR_HI_OFFSET        (0x10)
92 #define SPI_DMA_CH_DAR_LO_OFFSET        (0x14)
93 #define SPI_DMA_CH_DAR_HI_OFFSET        (0x18)
94
95 #define SPI_DMA_CH0_DONE_INT            BIT(0)
96 #define SPI_DMA_CH1_DONE_INT            BIT(1)
97 #define SPI_DMA_CH0_ABORT_INT           BIT(16)
98 #define SPI_DMA_CH1_ABORT_INT           BIT(17)
99 #define SPI_DMA_DONE_INT_MASK           (SPI_DMA_CH0_DONE_INT | SPI_DMA_CH1_DONE_INT)
100 #define SPI_DMA_ABORT_INT_MASK          (SPI_DMA_CH0_ABORT_INT | SPI_DMA_CH1_ABORT_INT)
101 #define DMA_CH_CONTROL_LIE              BIT(3)
102 #define DMA_CH_CONTROL_RIE              BIT(4)
103 #define DMA_INTR_EN                     (DMA_CH_CONTROL_RIE | DMA_CH_CONTROL_LIE)
104
105 /* x refers to SPI Host Controller HW instance id in the below macros - 0 or 1 */
106
107 #define SPI_MST_CMD_BUF_OFFSET(x)               (((x) * SPI_MST1_ADDR_BASE) + 0x00)
108 #define SPI_MST_RSP_BUF_OFFSET(x)               (((x) * SPI_MST1_ADDR_BASE) + 0x200)
109 #define SPI_MST_CTL_REG_OFFSET(x)               (((x) * SPI_MST1_ADDR_BASE) + 0x400)
110 #define SPI_MST_EVENT_REG_OFFSET(x)             (((x) * SPI_MST1_ADDR_BASE) + 0x420)
111 #define SPI_MST_EVENT_MASK_REG_OFFSET(x)        (((x) * SPI_MST1_ADDR_BASE) + 0x424)
112 #define SPI_MST_PAD_CTL_REG_OFFSET(x)           (((x) * SPI_MST1_ADDR_BASE) + 0x460)
113 #define SPIALERT_MST_DB_REG_OFFSET(x)           (((x) * SPI_MST1_ADDR_BASE) + 0x464)
114 #define SPIALERT_MST_VAL_REG_OFFSET(x)          (((x) * SPI_MST1_ADDR_BASE) + 0x468)
115 #define SPI_PCI_CTRL_REG_OFFSET(x)              (((x) * SPI_MST1_ADDR_BASE) + 0x480)
116
117 #define PCI1XXXX_IRQ_FLAGS                      (IRQF_NO_SUSPEND | IRQF_TRIGGER_NONE)
118 #define SPI_MAX_DATA_LEN                        320
119
120 #define PCI1XXXX_SPI_TIMEOUT                    (msecs_to_jiffies(100))
121 #define SYSLOCK_RETRY_CNT                       (1000)
122 #define SPI_DMA_ENGINE_EN                       (0x1)
123 #define SPI_DMA_ENGINE_DIS                      (0x0)
124
125 #define SPI_INTR                BIT(8)
126 #define SPI_FORCE_CE            BIT(4)
127
128 #define SPI_CHIP_SEL_COUNT 7
129 #define VENDOR_ID_MCHP 0x1055
130
131 #define SPI_SUSPEND_CONFIG 0x101
132 #define SPI_RESUME_CONFIG 0x203
133
134 struct pci1xxxx_spi_internal {
135         u8 hw_inst;
136         u8 clkdiv;
137         int irq;
138         int mode;
139         bool spi_xfer_in_progress;
140         void *rx_buf;
141         bool dma_aborted_rd;
142         u32 bytes_recvd;
143         u32 tx_sgl_len;
144         u32 rx_sgl_len;
145         struct scatterlist *tx_sgl, *rx_sgl;
146         bool dma_aborted_wr;
147         struct completion spi_xfer_done;
148         struct spi_controller *spi_host;
149         struct pci1xxxx_spi *parent;
150         struct spi_transfer *xfer;
151         struct {
152                 unsigned int dev_sel : 3;
153                 unsigned int msi_vector_sel : 1;
154         } prev_val;
155 };
156
157 struct pci1xxxx_spi {
158         struct pci_dev *dev;
159         u8 total_hw_instances;
160         u8 dev_rev;
161         void __iomem *reg_base;
162         void __iomem *dma_offset_bar;
163         /* lock to safely access the DMA registers in isr */
164         spinlock_t dma_reg_lock;
165         bool can_dma;
166         struct pci1xxxx_spi_internal *spi_int[] __counted_by(total_hw_instances);
167 };
168
169 static const struct pci_device_id pci1xxxx_spi_pci_id_table[] = {
170         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
171         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
172         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
173         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa004, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
174         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
175         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
176         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
177         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa014, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
178         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
179         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
180         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
181         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa024, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
182         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
183         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
184         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
185         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa034, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
186         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0001), 0, 0, 0x02},
187         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0002), 0, 0, 0x01},
188         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, 0x0003), 0, 0, 0x11},
189         { PCI_DEVICE_SUB(VENDOR_ID_MCHP, 0xa044, PCI_ANY_ID, PCI_ANY_ID), 0, 0, 0x01},
190         { 0, }
191 };
192
193 MODULE_DEVICE_TABLE(pci, pci1xxxx_spi_pci_id_table);
194
195 static int pci1xxxx_set_sys_lock(struct pci1xxxx_spi *par)
196 {
197         writel(SPI_SYSLOCK, par->reg_base + SPI_SYSLOCK_REG);
198         return readl(par->reg_base + SPI_SYSLOCK_REG);
199 }
200
201 static int pci1xxxx_acquire_sys_lock(struct pci1xxxx_spi *par)
202 {
203         u32 regval;
204
205         return readx_poll_timeout(pci1xxxx_set_sys_lock, par, regval,
206                            (regval & SPI_SYSLOCK), 100,
207                            SYSLOCK_RETRY_CNT * 100);
208 }
209
210 static void pci1xxxx_release_sys_lock(struct pci1xxxx_spi *par)
211 {
212         writel(0x0, par->reg_base + SPI_SYSLOCK_REG);
213 }
214
215 static int pci1xxxx_check_spi_can_dma(struct pci1xxxx_spi *spi_bus, int irq)
216 {
217         struct pci_dev *pdev = spi_bus->dev;
218         u32 pf_num;
219         u32 regval;
220         int ret;
221
222         /*
223          * DEV REV Registers is a system register, HW Syslock bit
224          * should be acquired before accessing the register
225          */
226         ret = pci1xxxx_acquire_sys_lock(spi_bus);
227         if (ret) {
228                 dev_err(&pdev->dev, "Error failed to acquire syslock\n");
229                 return ret;
230         }
231
232         regval = readl(spi_bus->reg_base + DEV_REV_REG);
233         spi_bus->dev_rev = regval & DEV_REV_MASK;
234         if (spi_bus->dev_rev >= 0xC0) {
235                 regval = readl(spi_bus->reg_base +
236                                SPI_CONFIG_PERI_ENABLE_REG);
237                 pf_num = regval & SPI_PERI_ENBLE_PF_MASK;
238         }
239
240         pci1xxxx_release_sys_lock(spi_bus);
241
242         /*
243          * DMA is supported only from C0 and SPI can use DMA only if
244          * it is mapped to PF0
245          */
246         if (spi_bus->dev_rev < 0xC0 || pf_num)
247                 return -EOPNOTSUPP;
248
249         /*
250          * DMA Supported only with MSI Interrupts
251          * One of the SPI instance's MSI vector address and data
252          * is used for DMA Interrupt
253          */
254         if (!irq_get_msi_desc(irq)) {
255                 dev_warn(&pdev->dev, "Error MSI Interrupt not supported, will operate in PIO mode\n");
256                 return -EOPNOTSUPP;
257         }
258
259         spi_bus->dma_offset_bar = pcim_iomap(pdev, 2, pci_resource_len(pdev, 2));
260         if (!spi_bus->dma_offset_bar) {
261                 dev_warn(&pdev->dev, "Error failed to map dma bar, will operate in PIO mode\n");
262                 return -EOPNOTSUPP;
263         }
264
265         if (dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64))) {
266                 dev_warn(&pdev->dev, "Error failed to set DMA mask, will operate in PIO mode\n");
267                 pcim_iounmap(pdev, spi_bus->dma_offset_bar);
268                 spi_bus->dma_offset_bar = NULL;
269                 return -EOPNOTSUPP;
270         }
271
272         return 0;
273 }
274
275 static int pci1xxxx_spi_dma_init(struct pci1xxxx_spi *spi_bus, int irq)
276 {
277         struct msi_msg msi;
278         int ret;
279
280         ret = pci1xxxx_check_spi_can_dma(spi_bus, irq);
281         if (ret)
282                 return ret;
283
284         spin_lock_init(&spi_bus->dma_reg_lock);
285         get_cached_msi_msg(irq, &msi);
286         writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
287         writel(SPI_DMA_ENGINE_EN, spi_bus->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
288         writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_HIGH);
289         writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_HIGH);
290         writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_HIGH);
291         writel(msi.address_hi, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_HIGH);
292         writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WDONE_LOW);
293         writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_WABORT_LOW);
294         writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RDONE_LOW);
295         writel(msi.address_lo, spi_bus->dma_offset_bar + SPI_DMA_INTR_IMWR_RABORT_LOW);
296         writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_WR_IMWR_DATA);
297         writel(msi.data, spi_bus->dma_offset_bar + SPI_DMA_INTR_RD_IMWR_DATA);
298         dma_set_max_seg_size(&spi_bus->dev->dev, PCI1XXXX_SPI_BUFFER_SIZE);
299         spi_bus->can_dma = true;
300         return 0;
301 }
302
303 static void pci1xxxx_spi_set_cs(struct spi_device *spi, bool enable)
304 {
305         struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi->controller);
306         struct pci1xxxx_spi *par = p->parent;
307         u32 regval;
308
309         /* Set the DEV_SEL bits of the SPI_MST_CTL_REG */
310         regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
311         if (!enable) {
312                 regval |= SPI_FORCE_CE;
313                 regval &= ~SPI_MST_CTL_DEVSEL_MASK;
314                 regval |= (spi_get_chipselect(spi, 0) << 25);
315         } else {
316                 regval &= ~SPI_FORCE_CE;
317         }
318         writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(p->hw_inst));
319 }
320
321 static u8 pci1xxxx_get_clock_div(u32 hz)
322 {
323         u8 val = 0;
324
325         if (hz >= PCI1XXXX_SPI_MAX_CLOCK_HZ)
326                 val = 2;
327         else if ((hz < PCI1XXXX_SPI_MAX_CLOCK_HZ) && (hz >= PCI1XXXX_SPI_CLK_20MHZ))
328                 val = 3;
329         else if ((hz < PCI1XXXX_SPI_CLK_20MHZ) && (hz >= PCI1XXXX_SPI_CLK_15MHZ))
330                 val = 4;
331         else if ((hz < PCI1XXXX_SPI_CLK_15MHZ) && (hz >= PCI1XXXX_SPI_CLK_12MHZ))
332                 val = 5;
333         else if ((hz < PCI1XXXX_SPI_CLK_12MHZ) && (hz >= PCI1XXXX_SPI_CLK_10MHZ))
334                 val = 6;
335         else if ((hz < PCI1XXXX_SPI_CLK_10MHZ) && (hz >= PCI1XXXX_SPI_MIN_CLOCK_HZ))
336                 val = 7;
337         else
338                 val = 2;
339
340         return val;
341 }
342
343 static void pci1xxxx_spi_setup_dma_to_io(struct pci1xxxx_spi_internal *p,
344                                          dma_addr_t dma_addr, u32 len)
345 {
346         void __iomem *base;
347
348         if (!p->hw_inst)
349                 base = p->parent->dma_offset_bar + SPI_DMA_CH0_RD_BASE;
350         else
351                 base = p->parent->dma_offset_bar + SPI_DMA_CH1_RD_BASE;
352
353         writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
354         writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
355         writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_SAR_LO_OFFSET);
356         writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_SAR_HI_OFFSET);
357         /* Updated SPI Command Registers */
358         writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
359                base + SPI_DMA_CH_DAR_LO_OFFSET);
360         writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_CMD_BUF_OFFSET(p->hw_inst)),
361                base + SPI_DMA_CH_DAR_HI_OFFSET);
362 }
363
364 static void pci1xxxx_spi_setup_dma_from_io(struct pci1xxxx_spi_internal *p,
365                                            dma_addr_t dma_addr, u32 len)
366 {
367         void *base;
368
369         if (!p->hw_inst)
370                 base = p->parent->dma_offset_bar + SPI_DMA_CH0_WR_BASE;
371         else
372                 base = p->parent->dma_offset_bar + SPI_DMA_CH1_WR_BASE;
373
374         writel(DMA_INTR_EN, base + SPI_DMA_CH_CTL1_OFFSET);
375         writel(len, base + SPI_DMA_CH_XFER_LEN_OFFSET);
376         writel(lower_32_bits(dma_addr), base + SPI_DMA_CH_DAR_LO_OFFSET);
377         writel(upper_32_bits(dma_addr), base + SPI_DMA_CH_DAR_HI_OFFSET);
378         writel(lower_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
379                base + SPI_DMA_CH_SAR_LO_OFFSET);
380         writel(upper_32_bits(SPI_PERI_ADDR_BASE + SPI_MST_RSP_BUF_OFFSET(p->hw_inst)),
381                base + SPI_DMA_CH_SAR_HI_OFFSET);
382 }
383
384 static void pci1xxxx_spi_setup(struct pci1xxxx_spi *par, u8 hw_inst, u32 mode,
385                                u8 clkdiv, u32 len)
386 {
387         u32 regval;
388
389         regval = readl(par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
390         regval &= ~(SPI_MST_CTL_MODE_SEL | SPI_MST_CTL_CMD_LEN_MASK |
391                     SPI_MST_CTL_SPEED_MASK);
392
393         if (mode == SPI_MODE_3)
394                 regval |= SPI_MST_CTL_MODE_SEL;
395
396         regval |= FIELD_PREP(SPI_MST_CTL_CMD_LEN_MASK, len);
397         regval |= FIELD_PREP(SPI_MST_CTL_SPEED_MASK, clkdiv);
398         writel(regval, par->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
399 }
400
401 static void pci1xxxx_start_spi_xfer(struct pci1xxxx_spi_internal *p, u8 hw_inst)
402 {
403         u32 regval;
404
405         regval = readl(p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
406         regval |= SPI_MST_CTL_GO;
407         writel(regval, p->parent->reg_base + SPI_MST_CTL_REG_OFFSET(hw_inst));
408 }
409
410 static int pci1xxxx_spi_transfer_with_io(struct spi_controller *spi_ctlr,
411                                          struct spi_device *spi, struct spi_transfer *xfer)
412 {
413         struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
414         struct pci1xxxx_spi *par = p->parent;
415         int len, loop_iter, transfer_len;
416         unsigned long bytes_transfered;
417         unsigned long bytes_recvd;
418         unsigned long loop_count;
419         u8 *rx_buf, result;
420         const u8 *tx_buf;
421         u32 regval;
422         u8 clkdiv;
423
424         p->spi_xfer_in_progress = true;
425         p->bytes_recvd = 0;
426         clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
427         tx_buf = xfer->tx_buf;
428         rx_buf = xfer->rx_buf;
429         transfer_len = xfer->len;
430         regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
431         writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
432
433         if (tx_buf) {
434                 bytes_transfered = 0;
435                 bytes_recvd = 0;
436                 loop_count = transfer_len / SPI_MAX_DATA_LEN;
437                 if (transfer_len % SPI_MAX_DATA_LEN != 0)
438                         loop_count += 1;
439
440                 for (loop_iter = 0; loop_iter < loop_count; loop_iter++) {
441                         len = SPI_MAX_DATA_LEN;
442                         if ((transfer_len % SPI_MAX_DATA_LEN != 0) &&
443                             (loop_iter == loop_count - 1))
444                                 len = transfer_len % SPI_MAX_DATA_LEN;
445
446                         reinit_completion(&p->spi_xfer_done);
447                         memcpy_toio(par->reg_base + SPI_MST_CMD_BUF_OFFSET(p->hw_inst),
448                                     &tx_buf[bytes_transfered], len);
449                         bytes_transfered += len;
450                         pci1xxxx_spi_setup(par, p->hw_inst, spi->mode, clkdiv, len);
451                         pci1xxxx_start_spi_xfer(p, p->hw_inst);
452
453                         /* Wait for DMA_TERM interrupt */
454                         result = wait_for_completion_timeout(&p->spi_xfer_done,
455                                                              PCI1XXXX_SPI_TIMEOUT);
456                         if (!result)
457                                 return -ETIMEDOUT;
458
459                         if (rx_buf) {
460                                 memcpy_fromio(&rx_buf[bytes_recvd], par->reg_base +
461                                               SPI_MST_RSP_BUF_OFFSET(p->hw_inst), len);
462                                 bytes_recvd += len;
463                         }
464                 }
465         }
466         p->spi_xfer_in_progress = false;
467
468         return 0;
469 }
470
471 static int pci1xxxx_spi_transfer_with_dma(struct spi_controller *spi_ctlr,
472                                           struct spi_device *spi,
473                                           struct spi_transfer *xfer)
474 {
475         struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(spi_ctlr);
476         struct pci1xxxx_spi *par = p->parent;
477         dma_addr_t rx_dma_addr = 0;
478         dma_addr_t tx_dma_addr = 0;
479         int ret = 0;
480         u32 regval;
481
482         p->spi_xfer_in_progress = true;
483         p->tx_sgl = xfer->tx_sg.sgl;
484         p->rx_sgl = xfer->rx_sg.sgl;
485         p->rx_buf = xfer->rx_buf;
486         regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
487         writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
488
489         if (!xfer->tx_buf || !p->tx_sgl) {
490                 ret = -EINVAL;
491                 goto error;
492         }
493         p->xfer = xfer;
494         p->mode = spi->mode;
495         p->clkdiv = pci1xxxx_get_clock_div(xfer->speed_hz);
496         p->bytes_recvd = 0;
497         p->rx_buf = xfer->rx_buf;
498         regval = readl(par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
499         writel(regval, par->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
500
501         tx_dma_addr = sg_dma_address(p->tx_sgl);
502         rx_dma_addr = sg_dma_address(p->rx_sgl);
503         p->tx_sgl_len = sg_dma_len(p->tx_sgl);
504         p->rx_sgl_len = sg_dma_len(p->rx_sgl);
505         pci1xxxx_spi_setup(par, p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
506         pci1xxxx_spi_setup_dma_to_io(p, (tx_dma_addr), p->tx_sgl_len);
507         if (rx_dma_addr)
508                 pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
509         writel(p->hw_inst, par->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
510
511         reinit_completion(&p->spi_xfer_done);
512         /* Wait for DMA_TERM interrupt */
513         ret = wait_for_completion_timeout(&p->spi_xfer_done, PCI1XXXX_SPI_TIMEOUT);
514         if (!ret) {
515                 ret = -ETIMEDOUT;
516                 if (p->dma_aborted_rd) {
517                         writel(SPI_DMA_ENGINE_DIS,
518                                par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
519                         /*
520                          * DMA ENGINE reset takes time if any TLP
521                          * completeion in progress, should wait
522                          * till DMA Engine reset is completed.
523                          */
524                         ret = readl_poll_timeout(par->dma_offset_bar +
525                                                  SPI_DMA_GLOBAL_RD_ENGINE_EN, regval,
526                                                  (regval == 0x0), 0, USEC_PER_MSEC);
527                         if (ret) {
528                                 ret = -ECANCELED;
529                                 goto error;
530                         }
531                         writel(SPI_DMA_ENGINE_EN,
532                                par->dma_offset_bar + SPI_DMA_GLOBAL_RD_ENGINE_EN);
533                         p->dma_aborted_rd = false;
534                         ret = -ECANCELED;
535                 }
536                 if (p->dma_aborted_wr) {
537                         writel(SPI_DMA_ENGINE_DIS,
538                                par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
539
540                         /*
541                          * DMA ENGINE reset takes time if any TLP
542                          * completeion in progress, should wait
543                          * till DMA Engine reset is completed.
544                          */
545                         ret = readl_poll_timeout(par->dma_offset_bar +
546                                                  SPI_DMA_GLOBAL_WR_ENGINE_EN, regval,
547                                                  (regval == 0x0), 0, USEC_PER_MSEC);
548                         if (ret) {
549                                 ret = -ECANCELED;
550                                 goto error;
551                         }
552
553                         writel(SPI_DMA_ENGINE_EN,
554                                par->dma_offset_bar + SPI_DMA_GLOBAL_WR_ENGINE_EN);
555                         p->dma_aborted_wr = false;
556                         ret = -ECANCELED;
557                 }
558                 goto error;
559         }
560         ret = 0;
561
562 error:
563         p->spi_xfer_in_progress = false;
564
565         return ret;
566 }
567
568 static int pci1xxxx_spi_transfer_one(struct spi_controller *spi_ctlr,
569                                      struct spi_device *spi, struct spi_transfer *xfer)
570 {
571         if (spi_xfer_is_dma_mapped(spi_ctlr, spi, xfer))
572                 return pci1xxxx_spi_transfer_with_dma(spi_ctlr, spi, xfer);
573         else
574                 return pci1xxxx_spi_transfer_with_io(spi_ctlr, spi, xfer);
575 }
576
577 static irqreturn_t pci1xxxx_spi_isr_io(int irq, void *dev)
578 {
579         struct pci1xxxx_spi_internal *p = dev;
580         irqreturn_t spi_int_fired = IRQ_NONE;
581         u32 regval;
582
583         /* Clear the SPI GO_BIT Interrupt */
584         regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
585         if (regval & SPI_INTR) {
586                 /* Clear xfer_done */
587                 if (p->parent->can_dma && p->rx_buf)
588                         writel(p->hw_inst, p->parent->dma_offset_bar +
589                                SPI_DMA_WR_DOORBELL_REG);
590                 else
591                         complete(&p->parent->spi_int[p->hw_inst]->spi_xfer_done);
592                 spi_int_fired = IRQ_HANDLED;
593         }
594         writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
595         return spi_int_fired;
596 }
597
598 static void pci1xxxx_spi_setup_next_dma_transfer(struct pci1xxxx_spi_internal *p)
599 {
600         dma_addr_t tx_dma_addr = 0;
601         dma_addr_t rx_dma_addr = 0;
602         u32 prev_len;
603
604         p->tx_sgl = sg_next(p->tx_sgl);
605         if (p->rx_sgl)
606                 p->rx_sgl = sg_next(p->rx_sgl);
607         if (!p->tx_sgl) {
608                 /* Clear xfer_done */
609                 complete(&p->spi_xfer_done);
610         } else {
611                 tx_dma_addr = sg_dma_address(p->tx_sgl);
612                 prev_len = p->tx_sgl_len;
613                 p->tx_sgl_len = sg_dma_len(p->tx_sgl);
614                 if (prev_len != p->tx_sgl_len)
615                         pci1xxxx_spi_setup(p->parent,
616                                            p->hw_inst, p->mode, p->clkdiv, p->tx_sgl_len);
617                 pci1xxxx_spi_setup_dma_to_io(p, tx_dma_addr, p->tx_sgl_len);
618                 if (p->rx_sgl) {
619                         rx_dma_addr = sg_dma_address(p->rx_sgl);
620                         p->rx_sgl_len = sg_dma_len(p->rx_sgl);
621                         pci1xxxx_spi_setup_dma_from_io(p, rx_dma_addr, p->rx_sgl_len);
622                 }
623                 writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_RD_DOORBELL_REG);
624         }
625 }
626
627 static irqreturn_t pci1xxxx_spi_isr_dma(int irq, void *dev)
628 {
629         struct pci1xxxx_spi_internal *p = dev;
630         irqreturn_t spi_int_fired = IRQ_NONE;
631         unsigned long flags;
632         u32 regval;
633
634         spin_lock_irqsave(&p->parent->dma_reg_lock, flags);
635         /* Clear the DMA RD INT and start spi xfer*/
636         regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_RD_STS);
637         if (regval & SPI_DMA_DONE_INT_MASK) {
638                 if (regval & SPI_DMA_CH0_DONE_INT)
639                         pci1xxxx_start_spi_xfer(p, SPI0);
640                 if (regval & SPI_DMA_CH1_DONE_INT)
641                         pci1xxxx_start_spi_xfer(p, SPI1);
642                 spi_int_fired = IRQ_HANDLED;
643         }
644         if (regval & SPI_DMA_ABORT_INT_MASK) {
645                 p->dma_aborted_rd = true;
646                 spi_int_fired = IRQ_HANDLED;
647         }
648         writel(regval, p->parent->dma_offset_bar + SPI_DMA_INTR_RD_CLR);
649
650         /* Clear the DMA WR INT */
651         regval = readl(p->parent->dma_offset_bar + SPI_DMA_INTR_WR_STS);
652         if (regval & SPI_DMA_DONE_INT_MASK) {
653                 if (regval & SPI_DMA_CH0_DONE_INT)
654                         pci1xxxx_spi_setup_next_dma_transfer(p->parent->spi_int[SPI0]);
655
656                 if (regval & SPI_DMA_CH1_DONE_INT)
657                         pci1xxxx_spi_setup_next_dma_transfer(p->parent->spi_int[SPI1]);
658
659                 spi_int_fired = IRQ_HANDLED;
660         }
661         if (regval & SPI_DMA_ABORT_INT_MASK) {
662                 p->dma_aborted_wr = true;
663                 spi_int_fired = IRQ_HANDLED;
664         }
665         writel(regval, p->parent->dma_offset_bar + SPI_DMA_INTR_WR_CLR);
666         spin_unlock_irqrestore(&p->parent->dma_reg_lock, flags);
667
668         /* Clear the SPI GO_BIT Interrupt */
669         regval = readl(p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
670         if (regval & SPI_INTR) {
671                 writel(p->hw_inst, p->parent->dma_offset_bar + SPI_DMA_WR_DOORBELL_REG);
672                 spi_int_fired = IRQ_HANDLED;
673         }
674         writel(regval, p->parent->reg_base + SPI_MST_EVENT_REG_OFFSET(p->hw_inst));
675         return spi_int_fired;
676 }
677
678 static irqreturn_t pci1xxxx_spi_isr(int irq, void *dev)
679 {
680         struct pci1xxxx_spi_internal *p = dev;
681
682         if (p->spi_host->can_dma(p->spi_host, NULL, p->xfer))
683                 return pci1xxxx_spi_isr_dma(irq, dev);
684         else
685                 return pci1xxxx_spi_isr_io(irq, dev);
686 }
687
688 static bool pci1xxxx_spi_can_dma(struct spi_controller *host,
689                                  struct spi_device *spi,
690                                  struct spi_transfer *xfer)
691 {
692         struct pci1xxxx_spi_internal *p = spi_controller_get_devdata(host);
693         struct pci1xxxx_spi *par = p->parent;
694
695         return par->can_dma;
696 }
697
698 static int pci1xxxx_spi_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
699 {
700         u8 hw_inst_cnt, iter, start, only_sec_inst;
701         struct pci1xxxx_spi_internal *spi_sub_ptr;
702         struct device *dev = &pdev->dev;
703         struct pci1xxxx_spi *spi_bus;
704         struct spi_controller *spi_host;
705         u32 regval;
706         int ret;
707
708         hw_inst_cnt = ent->driver_data & 0x0f;
709         start = (ent->driver_data & 0xf0) >> 4;
710         if (start == 1)
711                 only_sec_inst = 1;
712         else
713                 only_sec_inst = 0;
714
715         spi_bus = devm_kzalloc(&pdev->dev,
716                                struct_size(spi_bus, spi_int, hw_inst_cnt),
717                                GFP_KERNEL);
718         if (!spi_bus)
719                 return -ENOMEM;
720
721         spi_bus->dev = pdev;
722         spi_bus->total_hw_instances = hw_inst_cnt;
723         pci_set_master(pdev);
724
725         for (iter = 0; iter < hw_inst_cnt; iter++) {
726                 spi_bus->spi_int[iter] = devm_kzalloc(&pdev->dev,
727                                                       sizeof(struct pci1xxxx_spi_internal),
728                                                       GFP_KERNEL);
729                 if (!spi_bus->spi_int[iter])
730                         return -ENOMEM;
731                 spi_sub_ptr = spi_bus->spi_int[iter];
732                 spi_sub_ptr->spi_host = devm_spi_alloc_host(dev, sizeof(struct spi_controller));
733                 if (!spi_sub_ptr->spi_host)
734                         return -ENOMEM;
735
736                 spi_sub_ptr->parent = spi_bus;
737                 spi_sub_ptr->spi_xfer_in_progress = false;
738
739                 if (!iter) {
740                         ret = pcim_enable_device(pdev);
741                         if (ret)
742                                 return -ENOMEM;
743
744                         ret = pci_request_regions(pdev, DRV_NAME);
745                         if (ret)
746                                 return -ENOMEM;
747
748                         spi_bus->reg_base = pcim_iomap(pdev, 0, pci_resource_len(pdev, 0));
749                         if (!spi_bus->reg_base) {
750                                 ret = -EINVAL;
751                                 goto error;
752                         }
753
754                         ret = pci_alloc_irq_vectors(pdev, hw_inst_cnt, hw_inst_cnt,
755                                                     PCI_IRQ_ALL_TYPES);
756                         if (ret < 0) {
757                                 dev_err(&pdev->dev, "Error allocating MSI vectors\n");
758                                 goto error;
759                         }
760
761                         init_completion(&spi_sub_ptr->spi_xfer_done);
762                         /* Initialize Interrupts - SPI_INT */
763                         regval = readl(spi_bus->reg_base +
764                                        SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
765                         regval &= ~SPI_INTR;
766                         writel(regval, spi_bus->reg_base +
767                                SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
768                         spi_sub_ptr->irq = pci_irq_vector(pdev, 0);
769
770                         ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
771                                                pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
772                                                pci_name(pdev), spi_sub_ptr);
773                         if (ret < 0) {
774                                 dev_err(&pdev->dev, "Unable to request irq : %d",
775                                         spi_sub_ptr->irq);
776                                 ret = -ENODEV;
777                                 goto error;
778                         }
779
780                         ret = pci1xxxx_spi_dma_init(spi_bus, spi_sub_ptr->irq);
781                         if (ret && ret != -EOPNOTSUPP)
782                                 goto error;
783
784                         /* This register is only applicable for 1st instance */
785                         regval = readl(spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
786                         if (!only_sec_inst)
787                                 regval |= (BIT(4));
788                         else
789                                 regval &= ~(BIT(4));
790
791                         writel(regval, spi_bus->reg_base + SPI_PCI_CTRL_REG_OFFSET(0));
792                 }
793
794                 spi_sub_ptr->hw_inst = start++;
795
796                 if (iter == 1) {
797                         init_completion(&spi_sub_ptr->spi_xfer_done);
798                         /* Initialize Interrupts - SPI_INT */
799                         regval = readl(spi_bus->reg_base +
800                                SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
801                         regval &= ~SPI_INTR;
802                         writel(regval, spi_bus->reg_base +
803                                SPI_MST_EVENT_MASK_REG_OFFSET(spi_sub_ptr->hw_inst));
804                         spi_sub_ptr->irq = pci_irq_vector(pdev, iter);
805                         ret = devm_request_irq(&pdev->dev, spi_sub_ptr->irq,
806                                                pci1xxxx_spi_isr, PCI1XXXX_IRQ_FLAGS,
807                                                pci_name(pdev), spi_sub_ptr);
808                         if (ret < 0) {
809                                 dev_err(&pdev->dev, "Unable to request irq : %d",
810                                         spi_sub_ptr->irq);
811                                 ret = -ENODEV;
812                                 goto error;
813                         }
814                 }
815
816                 spi_host = spi_sub_ptr->spi_host;
817                 spi_host->num_chipselect = SPI_CHIP_SEL_COUNT;
818                 spi_host->mode_bits = SPI_MODE_0 | SPI_MODE_3 | SPI_RX_DUAL |
819                                       SPI_TX_DUAL | SPI_LOOP;
820                 spi_host->can_dma = pci1xxxx_spi_can_dma;
821                 spi_host->transfer_one = pci1xxxx_spi_transfer_one;
822
823                 spi_host->set_cs = pci1xxxx_spi_set_cs;
824                 spi_host->bits_per_word_mask = SPI_BPW_MASK(8);
825                 spi_host->max_speed_hz = PCI1XXXX_SPI_MAX_CLOCK_HZ;
826                 spi_host->min_speed_hz = PCI1XXXX_SPI_MIN_CLOCK_HZ;
827                 spi_host->flags = SPI_CONTROLLER_MUST_TX;
828                 spi_controller_set_devdata(spi_host, spi_sub_ptr);
829                 ret = devm_spi_register_controller(dev, spi_host);
830                 if (ret)
831                         goto error;
832         }
833         pci_set_drvdata(pdev, spi_bus);
834
835         return 0;
836
837 error:
838         pci_release_regions(pdev);
839         return ret;
840 }
841
842 static void store_restore_config(struct pci1xxxx_spi *spi_ptr,
843                                  struct pci1xxxx_spi_internal *spi_sub_ptr,
844                                  u8 inst, bool store)
845 {
846         u32 regval;
847
848         if (store) {
849                 regval = readl(spi_ptr->reg_base +
850                                SPI_MST_CTL_REG_OFFSET(spi_sub_ptr->hw_inst));
851                 regval &= SPI_MST_CTL_DEVSEL_MASK;
852                 spi_sub_ptr->prev_val.dev_sel = (regval >> 25) & 7;
853                 regval = readl(spi_ptr->reg_base +
854                                SPI_PCI_CTRL_REG_OFFSET(spi_sub_ptr->hw_inst));
855                 regval &= SPI_MSI_VECTOR_SEL_MASK;
856                 spi_sub_ptr->prev_val.msi_vector_sel = (regval >> 4) & 1;
857         } else {
858                 regval = readl(spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
859                 regval &= ~SPI_MST_CTL_DEVSEL_MASK;
860                 regval |= (spi_sub_ptr->prev_val.dev_sel << 25);
861                 writel(regval,
862                        spi_ptr->reg_base + SPI_MST_CTL_REG_OFFSET(inst));
863                 writel((spi_sub_ptr->prev_val.msi_vector_sel << 4),
864                         spi_ptr->reg_base + SPI_PCI_CTRL_REG_OFFSET(inst));
865         }
866 }
867
868 static int pci1xxxx_spi_resume(struct device *dev)
869 {
870         struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
871         struct pci1xxxx_spi_internal *spi_sub_ptr;
872         u32 regval = SPI_RESUME_CONFIG;
873         u8 iter;
874
875         for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
876                 spi_sub_ptr = spi_ptr->spi_int[iter];
877                 spi_controller_resume(spi_sub_ptr->spi_host);
878                 writel(regval, spi_ptr->reg_base +
879                        SPI_MST_EVENT_MASK_REG_OFFSET(iter));
880
881                 /* Restore config at resume */
882                 store_restore_config(spi_ptr, spi_sub_ptr, iter, 0);
883         }
884
885         return 0;
886 }
887
888 static int pci1xxxx_spi_suspend(struct device *dev)
889 {
890         struct pci1xxxx_spi *spi_ptr = dev_get_drvdata(dev);
891         struct pci1xxxx_spi_internal *spi_sub_ptr;
892         u32 reg1 = SPI_SUSPEND_CONFIG;
893         u8 iter;
894
895         for (iter = 0; iter < spi_ptr->total_hw_instances; iter++) {
896                 spi_sub_ptr = spi_ptr->spi_int[iter];
897
898                 while (spi_sub_ptr->spi_xfer_in_progress)
899                         msleep(20);
900
901                 /* Store existing config before suspend */
902                 store_restore_config(spi_ptr, spi_sub_ptr, iter, 1);
903                 spi_controller_suspend(spi_sub_ptr->spi_host);
904                 writel(reg1, spi_ptr->reg_base +
905                        SPI_MST_EVENT_MASK_REG_OFFSET(iter));
906         }
907
908         return 0;
909 }
910
911 static DEFINE_SIMPLE_DEV_PM_OPS(spi_pm_ops, pci1xxxx_spi_suspend,
912                                 pci1xxxx_spi_resume);
913
914 static struct pci_driver pci1xxxx_spi_driver = {
915         .name           = DRV_NAME,
916         .id_table       = pci1xxxx_spi_pci_id_table,
917         .probe          = pci1xxxx_spi_probe,
918         .driver         =       {
919                 .pm = pm_sleep_ptr(&spi_pm_ops),
920         },
921 };
922
923 module_pci_driver(pci1xxxx_spi_driver);
924
925 MODULE_DESCRIPTION("Microchip Technology Inc. pci1xxxx SPI bus driver");
926 MODULE_AUTHOR("Tharun Kumar P<[email protected]>");
927 MODULE_AUTHOR("Kumaravel Thiagarajan<[email protected]>");
928 MODULE_LICENSE("GPL v2");
This page took 0.083837 seconds and 4 git commands to generate.