]> Git Repo - linux.git/blob - drivers/soundwire/amd_manager.c
Merge patch series "riscv: Extension parsing fixes"
[linux.git] / drivers / soundwire / amd_manager.c
1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2 /*
3  * SoundWire AMD Manager driver
4  *
5  * Copyright 2023-24 Advanced Micro Devices, Inc.
6  */
7
8 #include <linux/completion.h>
9 #include <linux/device.h>
10 #include <linux/io.h>
11 #include <linux/jiffies.h>
12 #include <linux/kernel.h>
13 #include <linux/module.h>
14 #include <linux/slab.h>
15 #include <linux/soundwire/sdw.h>
16 #include <linux/soundwire/sdw_registers.h>
17 #include <linux/pm_runtime.h>
18 #include <linux/wait.h>
19 #include <sound/pcm_params.h>
20 #include <sound/soc.h>
21 #include "bus.h"
22 #include "amd_init.h"
23 #include "amd_manager.h"
24
25 #define DRV_NAME "amd_sdw_manager"
26
27 #define to_amd_sdw(b)   container_of(b, struct amd_sdw_manager, bus)
28
29 static int amd_init_sdw_manager(struct amd_sdw_manager *amd_manager)
30 {
31         u32 val;
32         int ret;
33
34         writel(AMD_SDW_ENABLE, amd_manager->mmio + ACP_SW_EN);
35         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, val, ACP_DELAY_US,
36                                  AMD_SDW_TIMEOUT);
37         if (ret)
38                 return ret;
39
40         /* SoundWire manager bus reset */
41         writel(AMD_SDW_BUS_RESET_REQ, amd_manager->mmio + ACP_SW_BUS_RESET_CTRL);
42         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_BUS_RESET_CTRL, val,
43                                  (val & AMD_SDW_BUS_RESET_DONE), ACP_DELAY_US, AMD_SDW_TIMEOUT);
44         if (ret)
45                 return ret;
46
47         writel(AMD_SDW_BUS_RESET_CLEAR_REQ, amd_manager->mmio + ACP_SW_BUS_RESET_CTRL);
48         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_BUS_RESET_CTRL, val, !val,
49                                  ACP_DELAY_US, AMD_SDW_TIMEOUT);
50         if (ret) {
51                 dev_err(amd_manager->dev, "Failed to reset SoundWire manager instance%d\n",
52                         amd_manager->instance);
53                 return ret;
54         }
55
56         writel(AMD_SDW_DISABLE, amd_manager->mmio + ACP_SW_EN);
57         return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, !val, ACP_DELAY_US,
58                                   AMD_SDW_TIMEOUT);
59 }
60
61 static int amd_enable_sdw_manager(struct amd_sdw_manager *amd_manager)
62 {
63         u32 val;
64
65         writel(AMD_SDW_ENABLE, amd_manager->mmio + ACP_SW_EN);
66         return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, val, ACP_DELAY_US,
67                                   AMD_SDW_TIMEOUT);
68 }
69
70 static int amd_disable_sdw_manager(struct amd_sdw_manager *amd_manager)
71 {
72         u32 val;
73
74         writel(AMD_SDW_DISABLE, amd_manager->mmio + ACP_SW_EN);
75         /*
76          * After invoking manager disable sequence, check whether
77          * manager has executed clock stop sequence. In this case,
78          * manager should ignore checking enable status register.
79          */
80         val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
81         if (val)
82                 return 0;
83         return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, !val, ACP_DELAY_US,
84                                   AMD_SDW_TIMEOUT);
85 }
86
87 static void amd_enable_sdw_interrupts(struct amd_sdw_manager *amd_manager)
88 {
89         u32 val;
90
91         mutex_lock(amd_manager->acp_sdw_lock);
92         val = sdw_manager_reg_mask_array[amd_manager->instance];
93         amd_updatel(amd_manager->acp_mmio, ACP_EXTERNAL_INTR_CNTL(amd_manager->instance), val, val);
94         mutex_unlock(amd_manager->acp_sdw_lock);
95
96         writel(AMD_SDW_IRQ_MASK_0TO7, amd_manager->mmio +
97                        ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
98         writel(AMD_SDW_IRQ_MASK_8TO11, amd_manager->mmio +
99                        ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
100         writel(AMD_SDW_IRQ_ERROR_MASK, amd_manager->mmio + ACP_SW_ERROR_INTR_MASK);
101 }
102
103 static void amd_disable_sdw_interrupts(struct amd_sdw_manager *amd_manager)
104 {
105         u32 irq_mask;
106
107         mutex_lock(amd_manager->acp_sdw_lock);
108         irq_mask = sdw_manager_reg_mask_array[amd_manager->instance];
109         amd_updatel(amd_manager->acp_mmio, ACP_EXTERNAL_INTR_CNTL(amd_manager->instance),
110                     irq_mask, 0);
111         mutex_unlock(amd_manager->acp_sdw_lock);
112
113         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
114         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
115         writel(0x00, amd_manager->mmio + ACP_SW_ERROR_INTR_MASK);
116 }
117
118 static int amd_deinit_sdw_manager(struct amd_sdw_manager *amd_manager)
119 {
120         amd_disable_sdw_interrupts(amd_manager);
121         return amd_disable_sdw_manager(amd_manager);
122 }
123
124 static void amd_sdw_set_frameshape(struct amd_sdw_manager *amd_manager)
125 {
126         u32 frame_size;
127
128         frame_size = (amd_manager->rows_index << 3) | amd_manager->cols_index;
129         writel(frame_size, amd_manager->mmio + ACP_SW_FRAMESIZE);
130 }
131
132 static void amd_sdw_wake_enable(struct amd_sdw_manager *amd_manager, bool enable)
133 {
134         u32 wake_ctrl;
135
136         wake_ctrl = readl(amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
137         if (enable)
138                 wake_ctrl |= AMD_SDW_WAKE_INTR_MASK;
139         else
140                 wake_ctrl &= ~AMD_SDW_WAKE_INTR_MASK;
141
142         writel(wake_ctrl, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
143 }
144
145 static void amd_sdw_ctl_word_prep(u32 *lower_word, u32 *upper_word, struct sdw_msg *msg,
146                                   int cmd_offset)
147 {
148         u32 upper_data;
149         u32 lower_data = 0;
150         u16 addr;
151         u8 upper_addr, lower_addr;
152         u8 data = 0;
153
154         addr = msg->addr + cmd_offset;
155         upper_addr = (addr & 0xFF00) >> 8;
156         lower_addr = addr & 0xFF;
157
158         if (msg->flags == SDW_MSG_FLAG_WRITE)
159                 data = msg->buf[cmd_offset];
160
161         upper_data = FIELD_PREP(AMD_SDW_MCP_CMD_DEV_ADDR, msg->dev_num);
162         upper_data |= FIELD_PREP(AMD_SDW_MCP_CMD_COMMAND, msg->flags + 2);
163         upper_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_ADDR_HIGH, upper_addr);
164         lower_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_ADDR_LOW, lower_addr);
165         lower_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_DATA, data);
166
167         *upper_word = upper_data;
168         *lower_word = lower_data;
169 }
170
171 static u64 amd_sdw_send_cmd_get_resp(struct amd_sdw_manager *amd_manager, u32 lower_data,
172                                      u32 upper_data)
173 {
174         u64 resp;
175         u32 lower_resp, upper_resp;
176         u32 sts;
177         int ret;
178
179         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts,
180                                  !(sts & AMD_SDW_IMM_CMD_BUSY), ACP_DELAY_US, AMD_SDW_TIMEOUT);
181         if (ret) {
182                 dev_err(amd_manager->dev, "SDW%x previous cmd status clear failed\n",
183                         amd_manager->instance);
184                 return ret;
185         }
186
187         if (sts & AMD_SDW_IMM_RES_VALID) {
188                 dev_err(amd_manager->dev, "SDW%x manager is in bad state\n", amd_manager->instance);
189                 writel(0x00, amd_manager->mmio + ACP_SW_IMM_CMD_STS);
190         }
191         writel(upper_data, amd_manager->mmio + ACP_SW_IMM_CMD_UPPER_WORD);
192         writel(lower_data, amd_manager->mmio + ACP_SW_IMM_CMD_LOWER_QWORD);
193
194         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts,
195                                  (sts & AMD_SDW_IMM_RES_VALID), ACP_DELAY_US, AMD_SDW_TIMEOUT);
196         if (ret) {
197                 dev_err(amd_manager->dev, "SDW%x cmd response timeout occurred\n",
198                         amd_manager->instance);
199                 return ret;
200         }
201         upper_resp = readl(amd_manager->mmio + ACP_SW_IMM_RESP_UPPER_WORD);
202         lower_resp = readl(amd_manager->mmio + ACP_SW_IMM_RESP_LOWER_QWORD);
203
204         writel(AMD_SDW_IMM_RES_VALID, amd_manager->mmio + ACP_SW_IMM_CMD_STS);
205         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts,
206                                  !(sts & AMD_SDW_IMM_RES_VALID), ACP_DELAY_US, AMD_SDW_TIMEOUT);
207         if (ret) {
208                 dev_err(amd_manager->dev, "SDW%x cmd status retry failed\n",
209                         amd_manager->instance);
210                 return ret;
211         }
212         resp = upper_resp;
213         resp = (resp << 32) | lower_resp;
214         return resp;
215 }
216
217 static enum sdw_command_response
218 amd_program_scp_addr(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg)
219 {
220         struct sdw_msg scp_msg = {0};
221         u64 response_buf[2] = {0};
222         u32 upper_data = 0, lower_data = 0;
223         int index;
224
225         scp_msg.dev_num = msg->dev_num;
226         scp_msg.addr = SDW_SCP_ADDRPAGE1;
227         scp_msg.buf = &msg->addr_page1;
228         scp_msg.flags = SDW_MSG_FLAG_WRITE;
229         amd_sdw_ctl_word_prep(&lower_data, &upper_data, &scp_msg, 0);
230         response_buf[0] = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data);
231         scp_msg.addr = SDW_SCP_ADDRPAGE2;
232         scp_msg.buf = &msg->addr_page2;
233         amd_sdw_ctl_word_prep(&lower_data, &upper_data, &scp_msg, 0);
234         response_buf[1] = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data);
235
236         for (index = 0; index < 2; index++) {
237                 if (response_buf[index] == -ETIMEDOUT) {
238                         dev_err_ratelimited(amd_manager->dev,
239                                             "SCP_addrpage command timeout for Slave %d\n",
240                                             msg->dev_num);
241                         return SDW_CMD_TIMEOUT;
242                 } else if (!(response_buf[index] & AMD_SDW_MCP_RESP_ACK)) {
243                         if (response_buf[index] & AMD_SDW_MCP_RESP_NACK) {
244                                 dev_err_ratelimited(amd_manager->dev,
245                                                     "SCP_addrpage NACKed for Slave %d\n",
246                                                     msg->dev_num);
247                                 return SDW_CMD_FAIL;
248                         }
249                         dev_dbg_ratelimited(amd_manager->dev, "SCP_addrpage ignored for Slave %d\n",
250                                             msg->dev_num);
251                         return SDW_CMD_IGNORED;
252                 }
253         }
254         return SDW_CMD_OK;
255 }
256
257 static int amd_prep_msg(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg)
258 {
259         int ret;
260
261         if (msg->page) {
262                 ret = amd_program_scp_addr(amd_manager, msg);
263                 if (ret) {
264                         msg->len = 0;
265                         return ret;
266                 }
267         }
268         switch (msg->flags) {
269         case SDW_MSG_FLAG_READ:
270         case SDW_MSG_FLAG_WRITE:
271                 break;
272         default:
273                 dev_err(amd_manager->dev, "Invalid msg cmd: %d\n", msg->flags);
274                 return -EINVAL;
275         }
276         return 0;
277 }
278
279 static enum sdw_command_response amd_sdw_fill_msg_resp(struct amd_sdw_manager *amd_manager,
280                                                        struct sdw_msg *msg, u64 response,
281                                                        int offset)
282 {
283         if (response & AMD_SDW_MCP_RESP_ACK) {
284                 if (msg->flags == SDW_MSG_FLAG_READ)
285                         msg->buf[offset] = FIELD_GET(AMD_SDW_MCP_RESP_RDATA, response);
286         } else {
287                 if (response == -ETIMEDOUT) {
288                         dev_err_ratelimited(amd_manager->dev, "command timeout for Slave %d\n",
289                                             msg->dev_num);
290                         return SDW_CMD_TIMEOUT;
291                 } else if (response & AMD_SDW_MCP_RESP_NACK) {
292                         dev_err_ratelimited(amd_manager->dev,
293                                             "command response NACK received for Slave %d\n",
294                                             msg->dev_num);
295                         return SDW_CMD_FAIL;
296                 }
297                 dev_err_ratelimited(amd_manager->dev, "command is ignored for Slave %d\n",
298                                     msg->dev_num);
299                 return SDW_CMD_IGNORED;
300         }
301         return SDW_CMD_OK;
302 }
303
304 static unsigned int _amd_sdw_xfer_msg(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg,
305                                       int cmd_offset)
306 {
307         u64 response;
308         u32 upper_data = 0, lower_data = 0;
309
310         amd_sdw_ctl_word_prep(&lower_data, &upper_data, msg, cmd_offset);
311         response = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data);
312         return amd_sdw_fill_msg_resp(amd_manager, msg, response, cmd_offset);
313 }
314
315 static enum sdw_command_response amd_sdw_xfer_msg(struct sdw_bus *bus, struct sdw_msg *msg)
316 {
317         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
318         int ret, i;
319
320         ret = amd_prep_msg(amd_manager, msg);
321         if (ret)
322                 return SDW_CMD_FAIL_OTHER;
323         for (i = 0; i < msg->len; i++) {
324                 ret = _amd_sdw_xfer_msg(amd_manager, msg, i);
325                 if (ret)
326                         return ret;
327         }
328         return SDW_CMD_OK;
329 }
330
331 static void amd_sdw_fill_slave_status(struct amd_sdw_manager *amd_manager, u16 index, u32 status)
332 {
333         switch (status) {
334         case SDW_SLAVE_ATTACHED:
335         case SDW_SLAVE_UNATTACHED:
336         case SDW_SLAVE_ALERT:
337                 amd_manager->status[index] = status;
338                 break;
339         default:
340                 amd_manager->status[index] = SDW_SLAVE_RESERVED;
341                 break;
342         }
343 }
344
345 static void amd_sdw_process_ping_status(u64 response, struct amd_sdw_manager *amd_manager)
346 {
347         u64 slave_stat;
348         u32 val;
349         u16 dev_index;
350
351         /* slave status response */
352         slave_stat = FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_0_3, response);
353         slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_4_11, response) << 8;
354         dev_dbg(amd_manager->dev, "slave_stat:0x%llx\n", slave_stat);
355         for (dev_index = 0; dev_index <= SDW_MAX_DEVICES; ++dev_index) {
356                 val = (slave_stat >> (dev_index * 2)) & AMD_SDW_MCP_SLAVE_STATUS_MASK;
357                 dev_dbg(amd_manager->dev, "val:0x%x\n", val);
358                 amd_sdw_fill_slave_status(amd_manager, dev_index, val);
359         }
360 }
361
362 static void amd_sdw_read_and_process_ping_status(struct amd_sdw_manager *amd_manager)
363 {
364         u64 response;
365
366         mutex_lock(&amd_manager->bus.msg_lock);
367         response = amd_sdw_send_cmd_get_resp(amd_manager, 0, 0);
368         mutex_unlock(&amd_manager->bus.msg_lock);
369         amd_sdw_process_ping_status(response, amd_manager);
370 }
371
372 static u32 amd_sdw_read_ping_status(struct sdw_bus *bus)
373 {
374         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
375         u64 response;
376         u32 slave_stat;
377
378         response = amd_sdw_send_cmd_get_resp(amd_manager, 0, 0);
379         /* slave status from ping response */
380         slave_stat = FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_0_3, response);
381         slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_4_11, response) << 8;
382         dev_dbg(amd_manager->dev, "slave_stat:0x%x\n", slave_stat);
383         return slave_stat;
384 }
385
386 static int amd_sdw_compute_params(struct sdw_bus *bus)
387 {
388         struct sdw_transport_data t_data = {0};
389         struct sdw_master_runtime *m_rt;
390         struct sdw_port_runtime *p_rt;
391         struct sdw_bus_params *b_params = &bus->params;
392         int port_bo, hstart, hstop, sample_int;
393         unsigned int rate, bps;
394
395         port_bo = 0;
396         hstart = 1;
397         hstop = bus->params.col - 1;
398         t_data.hstop = hstop;
399         t_data.hstart = hstart;
400
401         list_for_each_entry(m_rt, &bus->m_rt_list, bus_node) {
402                 rate = m_rt->stream->params.rate;
403                 bps = m_rt->stream->params.bps;
404                 sample_int = (bus->params.curr_dr_freq / rate);
405                 list_for_each_entry(p_rt, &m_rt->port_list, port_node) {
406                         port_bo = (p_rt->num * 64) + 1;
407                         dev_dbg(bus->dev, "p_rt->num=%d hstart=%d hstop=%d port_bo=%d\n",
408                                 p_rt->num, hstart, hstop, port_bo);
409                         sdw_fill_xport_params(&p_rt->transport_params, p_rt->num,
410                                               false, SDW_BLK_GRP_CNT_1, sample_int,
411                                               port_bo, port_bo >> 8, hstart, hstop,
412                                               SDW_BLK_PKG_PER_PORT, 0x0);
413
414                         sdw_fill_port_params(&p_rt->port_params,
415                                              p_rt->num, bps,
416                                              SDW_PORT_FLOW_MODE_ISOCH,
417                                              b_params->m_data_mode);
418                         t_data.hstart = hstart;
419                         t_data.hstop = hstop;
420                         t_data.block_offset = port_bo;
421                         t_data.sub_block_offset = 0;
422                 }
423                 sdw_compute_slave_ports(m_rt, &t_data);
424         }
425         return 0;
426 }
427
428 static int amd_sdw_port_params(struct sdw_bus *bus, struct sdw_port_params *p_params,
429                                unsigned int bank)
430 {
431         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
432         u32 frame_fmt_reg, dpn_frame_fmt;
433
434         dev_dbg(amd_manager->dev, "p_params->num:0x%x\n", p_params->num);
435         switch (amd_manager->instance) {
436         case ACP_SDW0:
437                 frame_fmt_reg = sdw0_manager_dp_reg[p_params->num].frame_fmt_reg;
438                 break;
439         case ACP_SDW1:
440                 frame_fmt_reg = sdw1_manager_dp_reg[p_params->num].frame_fmt_reg;
441                 break;
442         default:
443                 return -EINVAL;
444         }
445
446         dpn_frame_fmt = readl(amd_manager->mmio + frame_fmt_reg);
447         u32p_replace_bits(&dpn_frame_fmt, p_params->flow_mode, AMD_DPN_FRAME_FMT_PFM);
448         u32p_replace_bits(&dpn_frame_fmt, p_params->data_mode, AMD_DPN_FRAME_FMT_PDM);
449         u32p_replace_bits(&dpn_frame_fmt, p_params->bps - 1, AMD_DPN_FRAME_FMT_WORD_LEN);
450         writel(dpn_frame_fmt, amd_manager->mmio + frame_fmt_reg);
451         return 0;
452 }
453
454 static int amd_sdw_transport_params(struct sdw_bus *bus,
455                                     struct sdw_transport_params *params,
456                                     enum sdw_reg_bank bank)
457 {
458         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
459         u32 dpn_frame_fmt;
460         u32 dpn_sampleinterval;
461         u32 dpn_hctrl;
462         u32 dpn_offsetctrl;
463         u32 dpn_lanectrl;
464         u32 frame_fmt_reg, sample_int_reg, hctrl_dp0_reg;
465         u32 offset_reg, lane_ctrl_ch_en_reg;
466
467         switch (amd_manager->instance) {
468         case ACP_SDW0:
469                 frame_fmt_reg = sdw0_manager_dp_reg[params->port_num].frame_fmt_reg;
470                 sample_int_reg = sdw0_manager_dp_reg[params->port_num].sample_int_reg;
471                 hctrl_dp0_reg = sdw0_manager_dp_reg[params->port_num].hctrl_dp0_reg;
472                 offset_reg = sdw0_manager_dp_reg[params->port_num].offset_reg;
473                 lane_ctrl_ch_en_reg = sdw0_manager_dp_reg[params->port_num].lane_ctrl_ch_en_reg;
474                 break;
475         case ACP_SDW1:
476                 frame_fmt_reg = sdw1_manager_dp_reg[params->port_num].frame_fmt_reg;
477                 sample_int_reg = sdw1_manager_dp_reg[params->port_num].sample_int_reg;
478                 hctrl_dp0_reg = sdw1_manager_dp_reg[params->port_num].hctrl_dp0_reg;
479                 offset_reg = sdw1_manager_dp_reg[params->port_num].offset_reg;
480                 lane_ctrl_ch_en_reg = sdw1_manager_dp_reg[params->port_num].lane_ctrl_ch_en_reg;
481                 break;
482         default:
483                 return -EINVAL;
484         }
485         writel(AMD_SDW_SSP_COUNTER_VAL, amd_manager->mmio + ACP_SW_SSP_COUNTER);
486
487         dpn_frame_fmt = readl(amd_manager->mmio + frame_fmt_reg);
488         u32p_replace_bits(&dpn_frame_fmt, params->blk_pkg_mode, AMD_DPN_FRAME_FMT_BLK_PKG_MODE);
489         u32p_replace_bits(&dpn_frame_fmt, params->blk_grp_ctrl, AMD_DPN_FRAME_FMT_BLK_GRP_CTRL);
490         u32p_replace_bits(&dpn_frame_fmt, SDW_STREAM_PCM, AMD_DPN_FRAME_FMT_PCM_OR_PDM);
491         writel(dpn_frame_fmt, amd_manager->mmio + frame_fmt_reg);
492
493         dpn_sampleinterval = params->sample_interval - 1;
494         writel(dpn_sampleinterval, amd_manager->mmio + sample_int_reg);
495
496         dpn_hctrl = FIELD_PREP(AMD_DPN_HCTRL_HSTOP, params->hstop);
497         dpn_hctrl |= FIELD_PREP(AMD_DPN_HCTRL_HSTART, params->hstart);
498         writel(dpn_hctrl, amd_manager->mmio + hctrl_dp0_reg);
499
500         dpn_offsetctrl = FIELD_PREP(AMD_DPN_OFFSET_CTRL_1, params->offset1);
501         dpn_offsetctrl |= FIELD_PREP(AMD_DPN_OFFSET_CTRL_2, params->offset2);
502         writel(dpn_offsetctrl, amd_manager->mmio + offset_reg);
503
504         /*
505          * lane_ctrl_ch_en_reg will be used to program lane_ctrl and ch_mask
506          * parameters.
507          */
508         dpn_lanectrl = readl(amd_manager->mmio + lane_ctrl_ch_en_reg);
509         u32p_replace_bits(&dpn_lanectrl, params->lane_ctrl, AMD_DPN_CH_EN_LCTRL);
510         writel(dpn_lanectrl, amd_manager->mmio + lane_ctrl_ch_en_reg);
511         return 0;
512 }
513
514 static int amd_sdw_port_enable(struct sdw_bus *bus,
515                                struct sdw_enable_ch *enable_ch,
516                                unsigned int bank)
517 {
518         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
519         u32 dpn_ch_enable;
520         u32 lane_ctrl_ch_en_reg;
521
522         switch (amd_manager->instance) {
523         case ACP_SDW0:
524                 lane_ctrl_ch_en_reg = sdw0_manager_dp_reg[enable_ch->port_num].lane_ctrl_ch_en_reg;
525                 break;
526         case ACP_SDW1:
527                 lane_ctrl_ch_en_reg = sdw1_manager_dp_reg[enable_ch->port_num].lane_ctrl_ch_en_reg;
528                 break;
529         default:
530                 return -EINVAL;
531         }
532
533         /*
534          * lane_ctrl_ch_en_reg will be used to program lane_ctrl and ch_mask
535          * parameters.
536          */
537         dpn_ch_enable = readl(amd_manager->mmio + lane_ctrl_ch_en_reg);
538         u32p_replace_bits(&dpn_ch_enable, enable_ch->ch_mask, AMD_DPN_CH_EN_CHMASK);
539         if (enable_ch->enable)
540                 writel(dpn_ch_enable, amd_manager->mmio + lane_ctrl_ch_en_reg);
541         else
542                 writel(0, amd_manager->mmio + lane_ctrl_ch_en_reg);
543         return 0;
544 }
545
546 static int sdw_master_read_amd_prop(struct sdw_bus *bus)
547 {
548         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
549         struct fwnode_handle *link;
550         struct sdw_master_prop *prop;
551         u32 quirk_mask = 0;
552         u32 wake_en_mask = 0;
553         u32 power_mode_mask = 0;
554         char name[32];
555
556         prop = &bus->prop;
557         /* Find manager handle */
558         snprintf(name, sizeof(name), "mipi-sdw-link-%d-subproperties", bus->link_id);
559         link = device_get_named_child_node(bus->dev, name);
560         if (!link) {
561                 dev_err(bus->dev, "Manager node %s not found\n", name);
562                 return -EIO;
563         }
564         fwnode_property_read_u32(link, "amd-sdw-enable", &quirk_mask);
565         if (!(quirk_mask & AMD_SDW_QUIRK_MASK_BUS_ENABLE))
566                 prop->hw_disabled = true;
567         prop->quirks = SDW_MASTER_QUIRKS_CLEAR_INITIAL_CLASH |
568                        SDW_MASTER_QUIRKS_CLEAR_INITIAL_PARITY;
569
570         fwnode_property_read_u32(link, "amd-sdw-wakeup-enable", &wake_en_mask);
571         amd_manager->wake_en_mask = wake_en_mask;
572         fwnode_property_read_u32(link, "amd-sdw-power-mode", &power_mode_mask);
573         amd_manager->power_mode_mask = power_mode_mask;
574         return 0;
575 }
576
577 static int amd_prop_read(struct sdw_bus *bus)
578 {
579         sdw_master_read_prop(bus);
580         sdw_master_read_amd_prop(bus);
581         return 0;
582 }
583
584 static const struct sdw_master_port_ops amd_sdw_port_ops = {
585         .dpn_set_port_params = amd_sdw_port_params,
586         .dpn_set_port_transport_params = amd_sdw_transport_params,
587         .dpn_port_enable_ch = amd_sdw_port_enable,
588 };
589
590 static const struct sdw_master_ops amd_sdw_ops = {
591         .read_prop = amd_prop_read,
592         .xfer_msg = amd_sdw_xfer_msg,
593         .read_ping_status = amd_sdw_read_ping_status,
594 };
595
596 static int amd_sdw_hw_params(struct snd_pcm_substream *substream,
597                              struct snd_pcm_hw_params *params,
598                              struct snd_soc_dai *dai)
599 {
600         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
601         struct sdw_amd_dai_runtime *dai_runtime;
602         struct sdw_stream_config sconfig;
603         struct sdw_port_config *pconfig;
604         int ch, dir;
605         int ret;
606
607         dai_runtime = amd_manager->dai_runtime_array[dai->id];
608         if (!dai_runtime)
609                 return -EIO;
610
611         ch = params_channels(params);
612         if (substream->stream == SNDRV_PCM_STREAM_CAPTURE)
613                 dir = SDW_DATA_DIR_RX;
614         else
615                 dir = SDW_DATA_DIR_TX;
616         dev_dbg(amd_manager->dev, "dir:%d dai->id:0x%x\n", dir, dai->id);
617
618         sconfig.direction = dir;
619         sconfig.ch_count = ch;
620         sconfig.frame_rate = params_rate(params);
621         sconfig.type = dai_runtime->stream_type;
622
623         sconfig.bps = snd_pcm_format_width(params_format(params));
624
625         /* Port configuration */
626         pconfig = kzalloc(sizeof(*pconfig), GFP_KERNEL);
627         if (!pconfig) {
628                 ret =  -ENOMEM;
629                 goto error;
630         }
631
632         pconfig->num = dai->id;
633         pconfig->ch_mask = (1 << ch) - 1;
634         ret = sdw_stream_add_master(&amd_manager->bus, &sconfig,
635                                     pconfig, 1, dai_runtime->stream);
636         if (ret)
637                 dev_err(amd_manager->dev, "add manager to stream failed:%d\n", ret);
638
639         kfree(pconfig);
640 error:
641         return ret;
642 }
643
644 static int amd_sdw_hw_free(struct snd_pcm_substream *substream, struct snd_soc_dai *dai)
645 {
646         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
647         struct sdw_amd_dai_runtime *dai_runtime;
648         int ret;
649
650         dai_runtime = amd_manager->dai_runtime_array[dai->id];
651         if (!dai_runtime)
652                 return -EIO;
653
654         ret = sdw_stream_remove_master(&amd_manager->bus, dai_runtime->stream);
655         if (ret < 0)
656                 dev_err(dai->dev, "remove manager from stream %s failed: %d\n",
657                         dai_runtime->stream->name, ret);
658         return ret;
659 }
660
661 static int amd_set_sdw_stream(struct snd_soc_dai *dai, void *stream, int direction)
662 {
663         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
664         struct sdw_amd_dai_runtime *dai_runtime;
665
666         dai_runtime = amd_manager->dai_runtime_array[dai->id];
667         if (stream) {
668                 /* first paranoia check */
669                 if (dai_runtime) {
670                         dev_err(dai->dev, "dai_runtime already allocated for dai %s\n", dai->name);
671                         return -EINVAL;
672                 }
673
674                 /* allocate and set dai_runtime info */
675                 dai_runtime = kzalloc(sizeof(*dai_runtime), GFP_KERNEL);
676                 if (!dai_runtime)
677                         return -ENOMEM;
678
679                 dai_runtime->stream_type = SDW_STREAM_PCM;
680                 dai_runtime->bus = &amd_manager->bus;
681                 dai_runtime->stream = stream;
682                 amd_manager->dai_runtime_array[dai->id] = dai_runtime;
683         } else {
684                 /* second paranoia check */
685                 if (!dai_runtime) {
686                         dev_err(dai->dev, "dai_runtime not allocated for dai %s\n", dai->name);
687                         return -EINVAL;
688                 }
689
690                 /* for NULL stream we release allocated dai_runtime */
691                 kfree(dai_runtime);
692                 amd_manager->dai_runtime_array[dai->id] = NULL;
693         }
694         return 0;
695 }
696
697 static int amd_pcm_set_sdw_stream(struct snd_soc_dai *dai, void *stream, int direction)
698 {
699         return amd_set_sdw_stream(dai, stream, direction);
700 }
701
702 static void *amd_get_sdw_stream(struct snd_soc_dai *dai, int direction)
703 {
704         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
705         struct sdw_amd_dai_runtime *dai_runtime;
706
707         dai_runtime = amd_manager->dai_runtime_array[dai->id];
708         if (!dai_runtime)
709                 return ERR_PTR(-EINVAL);
710
711         return dai_runtime->stream;
712 }
713
714 static const struct snd_soc_dai_ops amd_sdw_dai_ops = {
715         .hw_params = amd_sdw_hw_params,
716         .hw_free = amd_sdw_hw_free,
717         .set_stream = amd_pcm_set_sdw_stream,
718         .get_stream = amd_get_sdw_stream,
719 };
720
721 static const struct snd_soc_component_driver amd_sdw_dai_component = {
722         .name = "soundwire",
723 };
724
725 static int amd_sdw_register_dais(struct amd_sdw_manager *amd_manager)
726 {
727         struct sdw_amd_dai_runtime **dai_runtime_array;
728         struct snd_soc_dai_driver *dais;
729         struct snd_soc_pcm_stream *stream;
730         struct device *dev;
731         int i, num_dais;
732
733         dev = amd_manager->dev;
734         num_dais = amd_manager->num_dout_ports + amd_manager->num_din_ports;
735         dais = devm_kcalloc(dev, num_dais, sizeof(*dais), GFP_KERNEL);
736         if (!dais)
737                 return -ENOMEM;
738
739         dai_runtime_array = devm_kcalloc(dev, num_dais,
740                                          sizeof(struct sdw_amd_dai_runtime *),
741                                          GFP_KERNEL);
742         if (!dai_runtime_array)
743                 return -ENOMEM;
744         amd_manager->dai_runtime_array = dai_runtime_array;
745         for (i = 0; i < num_dais; i++) {
746                 dais[i].name = devm_kasprintf(dev, GFP_KERNEL, "SDW%d Pin%d", amd_manager->instance,
747                                               i);
748                 if (!dais[i].name)
749                         return -ENOMEM;
750                 if (i < amd_manager->num_dout_ports)
751                         stream = &dais[i].playback;
752                 else
753                         stream = &dais[i].capture;
754
755                 stream->channels_min = 2;
756                 stream->channels_max = 2;
757                 stream->rates = SNDRV_PCM_RATE_48000;
758                 stream->formats = SNDRV_PCM_FMTBIT_S16_LE;
759
760                 dais[i].ops = &amd_sdw_dai_ops;
761                 dais[i].id = i;
762         }
763
764         return devm_snd_soc_register_component(dev, &amd_sdw_dai_component,
765                                                dais, num_dais);
766 }
767
768 static void amd_sdw_update_slave_status_work(struct work_struct *work)
769 {
770         struct amd_sdw_manager *amd_manager =
771                 container_of(work, struct amd_sdw_manager, amd_sdw_work);
772         int retry_count = 0;
773
774         if (amd_manager->status[0] == SDW_SLAVE_ATTACHED) {
775                 writel(0, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
776                 writel(0, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
777         }
778
779 update_status:
780         sdw_handle_slave_status(&amd_manager->bus, amd_manager->status);
781         /*
782          * During the peripheral enumeration sequence, the SoundWire manager interrupts
783          * are masked. Once the device number programming is done for all peripherals,
784          * interrupts will be unmasked. Read the peripheral device status from ping command
785          * and process the response. This sequence will ensure all peripheral devices enumerated
786          * and initialized properly.
787          */
788         if (amd_manager->status[0] == SDW_SLAVE_ATTACHED) {
789                 if (retry_count++ < SDW_MAX_DEVICES) {
790                         writel(AMD_SDW_IRQ_MASK_0TO7, amd_manager->mmio +
791                                ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
792                         writel(AMD_SDW_IRQ_MASK_8TO11, amd_manager->mmio +
793                                ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
794                         amd_sdw_read_and_process_ping_status(amd_manager);
795                         goto update_status;
796                 } else {
797                         dev_err_ratelimited(amd_manager->dev,
798                                             "Device0 detected after %d iterations\n",
799                                             retry_count);
800                 }
801         }
802 }
803
804 static void amd_sdw_update_slave_status(u32 status_change_0to7, u32 status_change_8to11,
805                                         struct amd_sdw_manager *amd_manager)
806 {
807         u64 slave_stat;
808         u32 val;
809         int dev_index;
810
811         if (status_change_0to7 == AMD_SDW_SLAVE_0_ATTACHED)
812                 memset(amd_manager->status, 0, sizeof(amd_manager->status));
813         slave_stat = status_change_0to7;
814         slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STATUS_8TO_11, status_change_8to11) << 32;
815         dev_dbg(amd_manager->dev, "status_change_0to7:0x%x status_change_8to11:0x%x\n",
816                 status_change_0to7, status_change_8to11);
817         if (slave_stat) {
818                 for (dev_index = 0; dev_index <= SDW_MAX_DEVICES; ++dev_index) {
819                         if (slave_stat & AMD_SDW_MCP_SLAVE_STATUS_VALID_MASK(dev_index)) {
820                                 val = (slave_stat >> AMD_SDW_MCP_SLAVE_STAT_SHIFT_MASK(dev_index)) &
821                                       AMD_SDW_MCP_SLAVE_STATUS_MASK;
822                                 amd_sdw_fill_slave_status(amd_manager, dev_index, val);
823                         }
824                 }
825         }
826 }
827
828 static void amd_sdw_process_wake_event(struct amd_sdw_manager *amd_manager)
829 {
830         pm_request_resume(amd_manager->dev);
831         writel(0x00, amd_manager->acp_mmio + ACP_SW_WAKE_EN(amd_manager->instance));
832         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_8TO11);
833 }
834
835 static void amd_sdw_irq_thread(struct work_struct *work)
836 {
837         struct amd_sdw_manager *amd_manager =
838                         container_of(work, struct amd_sdw_manager, amd_sdw_irq_thread);
839         u32 status_change_8to11;
840         u32 status_change_0to7;
841
842         status_change_8to11 = readl(amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_8TO11);
843         status_change_0to7 = readl(amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_0TO7);
844         dev_dbg(amd_manager->dev, "[SDW%d] SDW INT: 0to7=0x%x, 8to11=0x%x\n",
845                 amd_manager->instance, status_change_0to7, status_change_8to11);
846         if (status_change_8to11 & AMD_SDW_WAKE_STAT_MASK)
847                 return amd_sdw_process_wake_event(amd_manager);
848
849         if (status_change_8to11 & AMD_SDW_PREQ_INTR_STAT) {
850                 amd_sdw_read_and_process_ping_status(amd_manager);
851         } else {
852                 /* Check for the updated status on peripheral device */
853                 amd_sdw_update_slave_status(status_change_0to7, status_change_8to11, amd_manager);
854         }
855         if (status_change_8to11 || status_change_0to7)
856                 schedule_work(&amd_manager->amd_sdw_work);
857         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_8TO11);
858         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_0TO7);
859 }
860
861 int amd_sdw_manager_start(struct amd_sdw_manager *amd_manager)
862 {
863         struct sdw_master_prop *prop;
864         int ret;
865
866         prop = &amd_manager->bus.prop;
867         if (!prop->hw_disabled) {
868                 ret = amd_init_sdw_manager(amd_manager);
869                 if (ret)
870                         return ret;
871                 amd_enable_sdw_interrupts(amd_manager);
872                 ret = amd_enable_sdw_manager(amd_manager);
873                 if (ret)
874                         return ret;
875                 amd_sdw_set_frameshape(amd_manager);
876         }
877         /* Enable runtime PM */
878         pm_runtime_set_autosuspend_delay(amd_manager->dev, AMD_SDW_MASTER_SUSPEND_DELAY_MS);
879         pm_runtime_use_autosuspend(amd_manager->dev);
880         pm_runtime_mark_last_busy(amd_manager->dev);
881         pm_runtime_set_active(amd_manager->dev);
882         pm_runtime_enable(amd_manager->dev);
883         return 0;
884 }
885
886 static int amd_sdw_manager_probe(struct platform_device *pdev)
887 {
888         const struct acp_sdw_pdata *pdata = pdev->dev.platform_data;
889         struct resource *res;
890         struct device *dev = &pdev->dev;
891         struct sdw_master_prop *prop;
892         struct sdw_bus_params *params;
893         struct amd_sdw_manager *amd_manager;
894         int ret;
895
896         amd_manager = devm_kzalloc(dev, sizeof(struct amd_sdw_manager), GFP_KERNEL);
897         if (!amd_manager)
898                 return -ENOMEM;
899
900         res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
901         if (!res)
902                 return -ENOMEM;
903
904         amd_manager->acp_mmio = devm_ioremap(dev, res->start, resource_size(res));
905         if (!amd_manager->acp_mmio) {
906                 dev_err(dev, "mmio not found\n");
907                 return -ENOMEM;
908         }
909         amd_manager->instance = pdata->instance;
910         amd_manager->mmio = amd_manager->acp_mmio +
911                             (amd_manager->instance * SDW_MANAGER_REG_OFFSET);
912         amd_manager->acp_sdw_lock = pdata->acp_sdw_lock;
913         amd_manager->cols_index = sdw_find_col_index(AMD_SDW_DEFAULT_COLUMNS);
914         amd_manager->rows_index = sdw_find_row_index(AMD_SDW_DEFAULT_ROWS);
915         amd_manager->dev = dev;
916         amd_manager->bus.ops = &amd_sdw_ops;
917         amd_manager->bus.port_ops = &amd_sdw_port_ops;
918         amd_manager->bus.compute_params = &amd_sdw_compute_params;
919         amd_manager->bus.clk_stop_timeout = 200;
920         amd_manager->bus.link_id = amd_manager->instance;
921
922         /*
923          * Due to BIOS compatibility, the two links are exposed within
924          * the scope of a single controller. If this changes, the
925          * controller_id will have to be updated with drv_data
926          * information.
927          */
928         amd_manager->bus.controller_id = 0;
929
930         switch (amd_manager->instance) {
931         case ACP_SDW0:
932                 amd_manager->num_dout_ports = AMD_SDW0_MAX_TX_PORTS;
933                 amd_manager->num_din_ports = AMD_SDW0_MAX_RX_PORTS;
934                 break;
935         case ACP_SDW1:
936                 amd_manager->num_dout_ports = AMD_SDW1_MAX_TX_PORTS;
937                 amd_manager->num_din_ports = AMD_SDW1_MAX_RX_PORTS;
938                 break;
939         default:
940                 return -EINVAL;
941         }
942
943         params = &amd_manager->bus.params;
944
945         params->col = AMD_SDW_DEFAULT_COLUMNS;
946         params->row = AMD_SDW_DEFAULT_ROWS;
947         prop = &amd_manager->bus.prop;
948         prop->clk_freq = &amd_sdw_freq_tbl[0];
949         prop->mclk_freq = AMD_SDW_BUS_BASE_FREQ;
950         prop->max_clk_freq = AMD_SDW_DEFAULT_CLK_FREQ;
951
952         ret = sdw_bus_master_add(&amd_manager->bus, dev, dev->fwnode);
953         if (ret) {
954                 dev_err(dev, "Failed to register SoundWire manager(%d)\n", ret);
955                 return ret;
956         }
957         ret = amd_sdw_register_dais(amd_manager);
958         if (ret) {
959                 dev_err(dev, "CPU DAI registration failed\n");
960                 sdw_bus_master_delete(&amd_manager->bus);
961                 return ret;
962         }
963         dev_set_drvdata(dev, amd_manager);
964         INIT_WORK(&amd_manager->amd_sdw_irq_thread, amd_sdw_irq_thread);
965         INIT_WORK(&amd_manager->amd_sdw_work, amd_sdw_update_slave_status_work);
966         return 0;
967 }
968
969 static void amd_sdw_manager_remove(struct platform_device *pdev)
970 {
971         struct amd_sdw_manager *amd_manager = dev_get_drvdata(&pdev->dev);
972         int ret;
973
974         pm_runtime_disable(&pdev->dev);
975         amd_disable_sdw_interrupts(amd_manager);
976         sdw_bus_master_delete(&amd_manager->bus);
977         ret = amd_disable_sdw_manager(amd_manager);
978         if (ret)
979                 dev_err(&pdev->dev, "Failed to disable device (%pe)\n", ERR_PTR(ret));
980 }
981
982 static int amd_sdw_clock_stop(struct amd_sdw_manager *amd_manager)
983 {
984         u32 val;
985         int ret;
986
987         ret = sdw_bus_prep_clk_stop(&amd_manager->bus);
988         if (ret < 0 && ret != -ENODATA) {
989                 dev_err(amd_manager->dev, "prepare clock stop failed %d", ret);
990                 return 0;
991         }
992         ret = sdw_bus_clk_stop(&amd_manager->bus);
993         if (ret < 0 && ret != -ENODATA) {
994                 dev_err(amd_manager->dev, "bus clock stop failed %d", ret);
995                 return 0;
996         }
997
998         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL, val,
999                                  (val & AMD_SDW_CLK_STOP_DONE), ACP_DELAY_US, AMD_SDW_TIMEOUT);
1000         if (ret) {
1001                 dev_err(amd_manager->dev, "SDW%x clock stop failed\n", amd_manager->instance);
1002                 return 0;
1003         }
1004
1005         amd_manager->clk_stopped = true;
1006         if (amd_manager->wake_en_mask)
1007                 writel(0x01, amd_manager->acp_mmio + ACP_SW_WAKE_EN(amd_manager->instance));
1008
1009         dev_dbg(amd_manager->dev, "SDW%x clock stop successful\n", amd_manager->instance);
1010         return 0;
1011 }
1012
1013 static int amd_sdw_clock_stop_exit(struct amd_sdw_manager *amd_manager)
1014 {
1015         int ret;
1016         u32 val;
1017
1018         if (amd_manager->clk_stopped) {
1019                 val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1020                 val |= AMD_SDW_CLK_RESUME_REQ;
1021                 writel(val, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1022                 ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL, val,
1023                                          (val & AMD_SDW_CLK_RESUME_DONE), ACP_DELAY_US,
1024                                          AMD_SDW_TIMEOUT);
1025                 if (val & AMD_SDW_CLK_RESUME_DONE) {
1026                         writel(0, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1027                         ret = sdw_bus_exit_clk_stop(&amd_manager->bus);
1028                         if (ret < 0)
1029                                 dev_err(amd_manager->dev, "bus failed to exit clock stop %d\n",
1030                                         ret);
1031                         amd_manager->clk_stopped = false;
1032                 }
1033         }
1034         if (amd_manager->clk_stopped) {
1035                 dev_err(amd_manager->dev, "SDW%x clock stop exit failed\n", amd_manager->instance);
1036                 return 0;
1037         }
1038         dev_dbg(amd_manager->dev, "SDW%x clock stop exit successful\n", amd_manager->instance);
1039         return 0;
1040 }
1041
1042 static int amd_resume_child_device(struct device *dev, void *data)
1043 {
1044         struct sdw_slave *slave = dev_to_sdw_dev(dev);
1045         int ret;
1046
1047         if (!slave->probed) {
1048                 dev_dbg(dev, "skipping device, no probed driver\n");
1049                 return 0;
1050         }
1051         if (!slave->dev_num_sticky) {
1052                 dev_dbg(dev, "skipping device, never detected on bus\n");
1053                 return 0;
1054         }
1055         ret = pm_request_resume(dev);
1056         if (ret < 0) {
1057                 dev_err(dev, "pm_request_resume failed: %d\n", ret);
1058                 return ret;
1059         }
1060         return 0;
1061 }
1062
1063 static int __maybe_unused amd_pm_prepare(struct device *dev)
1064 {
1065         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1066         struct sdw_bus *bus = &amd_manager->bus;
1067         int ret;
1068
1069         if (bus->prop.hw_disabled) {
1070                 dev_dbg(bus->dev, "SoundWire manager %d is disabled, ignoring\n",
1071                         bus->link_id);
1072                 return 0;
1073         }
1074         /*
1075          * When multiple peripheral devices connected over the same link, if SoundWire manager
1076          * device is not in runtime suspend state, observed that device alerts are missing
1077          * without pm_prepare on AMD platforms in clockstop mode0.
1078          */
1079         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1080                 ret = pm_request_resume(dev);
1081                 if (ret < 0) {
1082                         dev_err(bus->dev, "pm_request_resume failed: %d\n", ret);
1083                         return 0;
1084                 }
1085         }
1086         /* To force peripheral devices to system level suspend state, resume the devices
1087          * from runtime suspend state first. Without that unable to dispatch the alert
1088          * status to peripheral driver during system level resume as they are in runtime
1089          * suspend state.
1090          */
1091         ret = device_for_each_child(bus->dev, NULL, amd_resume_child_device);
1092         if (ret < 0)
1093                 dev_err(dev, "amd_resume_child_device failed: %d\n", ret);
1094         return 0;
1095 }
1096
1097 static int __maybe_unused amd_suspend(struct device *dev)
1098 {
1099         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1100         struct sdw_bus *bus = &amd_manager->bus;
1101         int ret;
1102
1103         if (bus->prop.hw_disabled) {
1104                 dev_dbg(bus->dev, "SoundWire manager %d is disabled, ignoring\n",
1105                         bus->link_id);
1106                 return 0;
1107         }
1108
1109         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1110                 amd_sdw_wake_enable(amd_manager, false);
1111                 return amd_sdw_clock_stop(amd_manager);
1112         } else if (amd_manager->power_mode_mask & AMD_SDW_POWER_OFF_MODE) {
1113                 /*
1114                  * As per hardware programming sequence on AMD platforms,
1115                  * clock stop should be invoked first before powering-off
1116                  */
1117                 ret = amd_sdw_clock_stop(amd_manager);
1118                 if (ret)
1119                         return ret;
1120                 return amd_deinit_sdw_manager(amd_manager);
1121         }
1122         return 0;
1123 }
1124
1125 static int __maybe_unused amd_suspend_runtime(struct device *dev)
1126 {
1127         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1128         struct sdw_bus *bus = &amd_manager->bus;
1129         int ret;
1130
1131         if (bus->prop.hw_disabled) {
1132                 dev_dbg(bus->dev, "SoundWire manager %d is disabled,\n",
1133                         bus->link_id);
1134                 return 0;
1135         }
1136         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1137                 amd_sdw_wake_enable(amd_manager, true);
1138                 return amd_sdw_clock_stop(amd_manager);
1139         } else if (amd_manager->power_mode_mask & AMD_SDW_POWER_OFF_MODE) {
1140                 ret = amd_sdw_clock_stop(amd_manager);
1141                 if (ret)
1142                         return ret;
1143                 return amd_deinit_sdw_manager(amd_manager);
1144         }
1145         return 0;
1146 }
1147
1148 static int __maybe_unused amd_resume_runtime(struct device *dev)
1149 {
1150         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1151         struct sdw_bus *bus = &amd_manager->bus;
1152         int ret;
1153         u32 val;
1154
1155         if (bus->prop.hw_disabled) {
1156                 dev_dbg(bus->dev, "SoundWire manager %d is disabled, ignoring\n",
1157                         bus->link_id);
1158                 return 0;
1159         }
1160
1161         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1162                 return amd_sdw_clock_stop_exit(amd_manager);
1163         } else if (amd_manager->power_mode_mask & AMD_SDW_POWER_OFF_MODE) {
1164                 val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1165                 if (val) {
1166                         val |= AMD_SDW_CLK_RESUME_REQ;
1167                         writel(val, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1168                         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL, val,
1169                                                  (val & AMD_SDW_CLK_RESUME_DONE), ACP_DELAY_US,
1170                                                  AMD_SDW_TIMEOUT);
1171                         if (val & AMD_SDW_CLK_RESUME_DONE) {
1172                                 writel(0, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1173                                 amd_manager->clk_stopped = false;
1174                         }
1175                 }
1176                 sdw_clear_slave_status(bus, SDW_UNATTACH_REQUEST_MASTER_RESET);
1177                 amd_init_sdw_manager(amd_manager);
1178                 amd_enable_sdw_interrupts(amd_manager);
1179                 ret = amd_enable_sdw_manager(amd_manager);
1180                 if (ret)
1181                         return ret;
1182                 amd_sdw_set_frameshape(amd_manager);
1183         }
1184         return 0;
1185 }
1186
1187 static const struct dev_pm_ops amd_pm = {
1188         .prepare = amd_pm_prepare,
1189         SET_SYSTEM_SLEEP_PM_OPS(amd_suspend, amd_resume_runtime)
1190         SET_RUNTIME_PM_OPS(amd_suspend_runtime, amd_resume_runtime, NULL)
1191 };
1192
1193 static struct platform_driver amd_sdw_driver = {
1194         .probe  = &amd_sdw_manager_probe,
1195         .remove_new = &amd_sdw_manager_remove,
1196         .driver = {
1197                 .name   = "amd_sdw_manager",
1198                 .pm = &amd_pm,
1199         }
1200 };
1201 module_platform_driver(amd_sdw_driver);
1202
1203 MODULE_AUTHOR("[email protected]");
1204 MODULE_DESCRIPTION("AMD SoundWire driver");
1205 MODULE_LICENSE("Dual BSD/GPL");
1206 MODULE_ALIAS("platform:" DRV_NAME);
This page took 0.106009 seconds and 4 git commands to generate.