GNU Linux-libre 6.9.1-gnu
[releases.git] / drivers / soundwire / amd_manager.c
1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2 /*
3  * SoundWire AMD Manager driver
4  *
5  * Copyright 2023-24 Advanced Micro Devices, Inc.
6  */
7
8 #include <linux/completion.h>
9 #include <linux/device.h>
10 #include <linux/io.h>
11 #include <linux/jiffies.h>
12 #include <linux/kernel.h>
13 #include <linux/module.h>
14 #include <linux/slab.h>
15 #include <linux/soundwire/sdw.h>
16 #include <linux/soundwire/sdw_registers.h>
17 #include <linux/pm_runtime.h>
18 #include <linux/wait.h>
19 #include <sound/pcm_params.h>
20 #include <sound/soc.h>
21 #include "bus.h"
22 #include "amd_init.h"
23 #include "amd_manager.h"
24
25 #define DRV_NAME "amd_sdw_manager"
26
27 #define to_amd_sdw(b)   container_of(b, struct amd_sdw_manager, bus)
28
29 static int amd_init_sdw_manager(struct amd_sdw_manager *amd_manager)
30 {
31         u32 val;
32         int ret;
33
34         writel(AMD_SDW_ENABLE, amd_manager->mmio + ACP_SW_EN);
35         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, val, ACP_DELAY_US,
36                                  AMD_SDW_TIMEOUT);
37         if (ret)
38                 return ret;
39
40         /* SoundWire manager bus reset */
41         writel(AMD_SDW_BUS_RESET_REQ, amd_manager->mmio + ACP_SW_BUS_RESET_CTRL);
42         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_BUS_RESET_CTRL, val,
43                                  (val & AMD_SDW_BUS_RESET_DONE), ACP_DELAY_US, AMD_SDW_TIMEOUT);
44         if (ret)
45                 return ret;
46
47         writel(AMD_SDW_BUS_RESET_CLEAR_REQ, amd_manager->mmio + ACP_SW_BUS_RESET_CTRL);
48         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_BUS_RESET_CTRL, val, !val,
49                                  ACP_DELAY_US, AMD_SDW_TIMEOUT);
50         if (ret) {
51                 dev_err(amd_manager->dev, "Failed to reset SoundWire manager instance%d\n",
52                         amd_manager->instance);
53                 return ret;
54         }
55
56         writel(AMD_SDW_DISABLE, amd_manager->mmio + ACP_SW_EN);
57         return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, !val, ACP_DELAY_US,
58                                   AMD_SDW_TIMEOUT);
59 }
60
61 static int amd_enable_sdw_manager(struct amd_sdw_manager *amd_manager)
62 {
63         u32 val;
64
65         writel(AMD_SDW_ENABLE, amd_manager->mmio + ACP_SW_EN);
66         return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, val, ACP_DELAY_US,
67                                   AMD_SDW_TIMEOUT);
68 }
69
70 static int amd_disable_sdw_manager(struct amd_sdw_manager *amd_manager)
71 {
72         u32 val;
73
74         writel(AMD_SDW_DISABLE, amd_manager->mmio + ACP_SW_EN);
75         /*
76          * After invoking manager disable sequence, check whether
77          * manager has executed clock stop sequence. In this case,
78          * manager should ignore checking enable status register.
79          */
80         val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
81         if (val)
82                 return 0;
83         return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, !val, ACP_DELAY_US,
84                                   AMD_SDW_TIMEOUT);
85 }
86
87 static void amd_enable_sdw_interrupts(struct amd_sdw_manager *amd_manager)
88 {
89         u32 val;
90
91         mutex_lock(amd_manager->acp_sdw_lock);
92         val = readl(amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance));
93         val |= sdw_manager_reg_mask_array[amd_manager->instance];
94         writel(val, amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance));
95         mutex_unlock(amd_manager->acp_sdw_lock);
96
97         writel(AMD_SDW_IRQ_MASK_0TO7, amd_manager->mmio +
98                        ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
99         writel(AMD_SDW_IRQ_MASK_8TO11, amd_manager->mmio +
100                        ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
101         writel(AMD_SDW_IRQ_ERROR_MASK, amd_manager->mmio + ACP_SW_ERROR_INTR_MASK);
102 }
103
104 static void amd_disable_sdw_interrupts(struct amd_sdw_manager *amd_manager)
105 {
106         u32 val;
107
108         mutex_lock(amd_manager->acp_sdw_lock);
109         val = readl(amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance));
110         val &= ~sdw_manager_reg_mask_array[amd_manager->instance];
111         writel(val, amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance));
112         mutex_unlock(amd_manager->acp_sdw_lock);
113
114         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
115         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
116         writel(0x00, amd_manager->mmio + ACP_SW_ERROR_INTR_MASK);
117 }
118
119 static int amd_deinit_sdw_manager(struct amd_sdw_manager *amd_manager)
120 {
121         amd_disable_sdw_interrupts(amd_manager);
122         return amd_disable_sdw_manager(amd_manager);
123 }
124
125 static void amd_sdw_set_frameshape(struct amd_sdw_manager *amd_manager)
126 {
127         u32 frame_size;
128
129         frame_size = (amd_manager->rows_index << 3) | amd_manager->cols_index;
130         writel(frame_size, amd_manager->mmio + ACP_SW_FRAMESIZE);
131 }
132
133 static void amd_sdw_wake_enable(struct amd_sdw_manager *amd_manager, bool enable)
134 {
135         u32 wake_ctrl;
136
137         wake_ctrl = readl(amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
138         if (enable)
139                 wake_ctrl |= AMD_SDW_WAKE_INTR_MASK;
140         else
141                 wake_ctrl &= ~AMD_SDW_WAKE_INTR_MASK;
142
143         writel(wake_ctrl, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
144 }
145
146 static void amd_sdw_ctl_word_prep(u32 *lower_word, u32 *upper_word, struct sdw_msg *msg,
147                                   int cmd_offset)
148 {
149         u32 upper_data;
150         u32 lower_data = 0;
151         u16 addr;
152         u8 upper_addr, lower_addr;
153         u8 data = 0;
154
155         addr = msg->addr + cmd_offset;
156         upper_addr = (addr & 0xFF00) >> 8;
157         lower_addr = addr & 0xFF;
158
159         if (msg->flags == SDW_MSG_FLAG_WRITE)
160                 data = msg->buf[cmd_offset];
161
162         upper_data = FIELD_PREP(AMD_SDW_MCP_CMD_DEV_ADDR, msg->dev_num);
163         upper_data |= FIELD_PREP(AMD_SDW_MCP_CMD_COMMAND, msg->flags + 2);
164         upper_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_ADDR_HIGH, upper_addr);
165         lower_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_ADDR_LOW, lower_addr);
166         lower_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_DATA, data);
167
168         *upper_word = upper_data;
169         *lower_word = lower_data;
170 }
171
172 static u64 amd_sdw_send_cmd_get_resp(struct amd_sdw_manager *amd_manager, u32 lower_data,
173                                      u32 upper_data)
174 {
175         u64 resp;
176         u32 lower_resp, upper_resp;
177         u32 sts;
178         int ret;
179
180         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts,
181                                  !(sts & AMD_SDW_IMM_CMD_BUSY), ACP_DELAY_US, AMD_SDW_TIMEOUT);
182         if (ret) {
183                 dev_err(amd_manager->dev, "SDW%x previous cmd status clear failed\n",
184                         amd_manager->instance);
185                 return ret;
186         }
187
188         if (sts & AMD_SDW_IMM_RES_VALID) {
189                 dev_err(amd_manager->dev, "SDW%x manager is in bad state\n", amd_manager->instance);
190                 writel(0x00, amd_manager->mmio + ACP_SW_IMM_CMD_STS);
191         }
192         writel(upper_data, amd_manager->mmio + ACP_SW_IMM_CMD_UPPER_WORD);
193         writel(lower_data, amd_manager->mmio + ACP_SW_IMM_CMD_LOWER_QWORD);
194
195         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts,
196                                  (sts & AMD_SDW_IMM_RES_VALID), ACP_DELAY_US, AMD_SDW_TIMEOUT);
197         if (ret) {
198                 dev_err(amd_manager->dev, "SDW%x cmd response timeout occurred\n",
199                         amd_manager->instance);
200                 return ret;
201         }
202         upper_resp = readl(amd_manager->mmio + ACP_SW_IMM_RESP_UPPER_WORD);
203         lower_resp = readl(amd_manager->mmio + ACP_SW_IMM_RESP_LOWER_QWORD);
204
205         writel(AMD_SDW_IMM_RES_VALID, amd_manager->mmio + ACP_SW_IMM_CMD_STS);
206         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts,
207                                  !(sts & AMD_SDW_IMM_RES_VALID), ACP_DELAY_US, AMD_SDW_TIMEOUT);
208         if (ret) {
209                 dev_err(amd_manager->dev, "SDW%x cmd status retry failed\n",
210                         amd_manager->instance);
211                 return ret;
212         }
213         resp = upper_resp;
214         resp = (resp << 32) | lower_resp;
215         return resp;
216 }
217
218 static enum sdw_command_response
219 amd_program_scp_addr(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg)
220 {
221         struct sdw_msg scp_msg = {0};
222         u64 response_buf[2] = {0};
223         u32 upper_data = 0, lower_data = 0;
224         int index;
225
226         scp_msg.dev_num = msg->dev_num;
227         scp_msg.addr = SDW_SCP_ADDRPAGE1;
228         scp_msg.buf = &msg->addr_page1;
229         scp_msg.flags = SDW_MSG_FLAG_WRITE;
230         amd_sdw_ctl_word_prep(&lower_data, &upper_data, &scp_msg, 0);
231         response_buf[0] = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data);
232         scp_msg.addr = SDW_SCP_ADDRPAGE2;
233         scp_msg.buf = &msg->addr_page2;
234         amd_sdw_ctl_word_prep(&lower_data, &upper_data, &scp_msg, 0);
235         response_buf[1] = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data);
236
237         for (index = 0; index < 2; index++) {
238                 if (response_buf[index] == -ETIMEDOUT) {
239                         dev_err_ratelimited(amd_manager->dev,
240                                             "SCP_addrpage command timeout for Slave %d\n",
241                                             msg->dev_num);
242                         return SDW_CMD_TIMEOUT;
243                 } else if (!(response_buf[index] & AMD_SDW_MCP_RESP_ACK)) {
244                         if (response_buf[index] & AMD_SDW_MCP_RESP_NACK) {
245                                 dev_err_ratelimited(amd_manager->dev,
246                                                     "SCP_addrpage NACKed for Slave %d\n",
247                                                     msg->dev_num);
248                                 return SDW_CMD_FAIL;
249                         }
250                         dev_dbg_ratelimited(amd_manager->dev, "SCP_addrpage ignored for Slave %d\n",
251                                             msg->dev_num);
252                         return SDW_CMD_IGNORED;
253                 }
254         }
255         return SDW_CMD_OK;
256 }
257
258 static int amd_prep_msg(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg)
259 {
260         int ret;
261
262         if (msg->page) {
263                 ret = amd_program_scp_addr(amd_manager, msg);
264                 if (ret) {
265                         msg->len = 0;
266                         return ret;
267                 }
268         }
269         switch (msg->flags) {
270         case SDW_MSG_FLAG_READ:
271         case SDW_MSG_FLAG_WRITE:
272                 break;
273         default:
274                 dev_err(amd_manager->dev, "Invalid msg cmd: %d\n", msg->flags);
275                 return -EINVAL;
276         }
277         return 0;
278 }
279
280 static enum sdw_command_response amd_sdw_fill_msg_resp(struct amd_sdw_manager *amd_manager,
281                                                        struct sdw_msg *msg, u64 response,
282                                                        int offset)
283 {
284         if (response & AMD_SDW_MCP_RESP_ACK) {
285                 if (msg->flags == SDW_MSG_FLAG_READ)
286                         msg->buf[offset] = FIELD_GET(AMD_SDW_MCP_RESP_RDATA, response);
287         } else {
288                 if (response == -ETIMEDOUT) {
289                         dev_err_ratelimited(amd_manager->dev, "command timeout for Slave %d\n",
290                                             msg->dev_num);
291                         return SDW_CMD_TIMEOUT;
292                 } else if (response & AMD_SDW_MCP_RESP_NACK) {
293                         dev_err_ratelimited(amd_manager->dev,
294                                             "command response NACK received for Slave %d\n",
295                                             msg->dev_num);
296                         return SDW_CMD_FAIL;
297                 }
298                 dev_err_ratelimited(amd_manager->dev, "command is ignored for Slave %d\n",
299                                     msg->dev_num);
300                 return SDW_CMD_IGNORED;
301         }
302         return SDW_CMD_OK;
303 }
304
305 static unsigned int _amd_sdw_xfer_msg(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg,
306                                       int cmd_offset)
307 {
308         u64 response;
309         u32 upper_data = 0, lower_data = 0;
310
311         amd_sdw_ctl_word_prep(&lower_data, &upper_data, msg, cmd_offset);
312         response = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data);
313         return amd_sdw_fill_msg_resp(amd_manager, msg, response, cmd_offset);
314 }
315
316 static enum sdw_command_response amd_sdw_xfer_msg(struct sdw_bus *bus, struct sdw_msg *msg)
317 {
318         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
319         int ret, i;
320
321         ret = amd_prep_msg(amd_manager, msg);
322         if (ret)
323                 return SDW_CMD_FAIL_OTHER;
324         for (i = 0; i < msg->len; i++) {
325                 ret = _amd_sdw_xfer_msg(amd_manager, msg, i);
326                 if (ret)
327                         return ret;
328         }
329         return SDW_CMD_OK;
330 }
331
332 static void amd_sdw_fill_slave_status(struct amd_sdw_manager *amd_manager, u16 index, u32 status)
333 {
334         switch (status) {
335         case SDW_SLAVE_ATTACHED:
336         case SDW_SLAVE_UNATTACHED:
337         case SDW_SLAVE_ALERT:
338                 amd_manager->status[index] = status;
339                 break;
340         default:
341                 amd_manager->status[index] = SDW_SLAVE_RESERVED;
342                 break;
343         }
344 }
345
346 static void amd_sdw_process_ping_status(u64 response, struct amd_sdw_manager *amd_manager)
347 {
348         u64 slave_stat;
349         u32 val;
350         u16 dev_index;
351
352         /* slave status response */
353         slave_stat = FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_0_3, response);
354         slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_4_11, response) << 8;
355         dev_dbg(amd_manager->dev, "slave_stat:0x%llx\n", slave_stat);
356         for (dev_index = 0; dev_index <= SDW_MAX_DEVICES; ++dev_index) {
357                 val = (slave_stat >> (dev_index * 2)) & AMD_SDW_MCP_SLAVE_STATUS_MASK;
358                 dev_dbg(amd_manager->dev, "val:0x%x\n", val);
359                 amd_sdw_fill_slave_status(amd_manager, dev_index, val);
360         }
361 }
362
363 static void amd_sdw_read_and_process_ping_status(struct amd_sdw_manager *amd_manager)
364 {
365         u64 response;
366
367         mutex_lock(&amd_manager->bus.msg_lock);
368         response = amd_sdw_send_cmd_get_resp(amd_manager, 0, 0);
369         mutex_unlock(&amd_manager->bus.msg_lock);
370         amd_sdw_process_ping_status(response, amd_manager);
371 }
372
373 static u32 amd_sdw_read_ping_status(struct sdw_bus *bus)
374 {
375         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
376         u64 response;
377         u32 slave_stat;
378
379         response = amd_sdw_send_cmd_get_resp(amd_manager, 0, 0);
380         /* slave status from ping response */
381         slave_stat = FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_0_3, response);
382         slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_4_11, response) << 8;
383         dev_dbg(amd_manager->dev, "slave_stat:0x%x\n", slave_stat);
384         return slave_stat;
385 }
386
387 static int amd_sdw_compute_params(struct sdw_bus *bus)
388 {
389         struct sdw_transport_data t_data = {0};
390         struct sdw_master_runtime *m_rt;
391         struct sdw_port_runtime *p_rt;
392         struct sdw_bus_params *b_params = &bus->params;
393         int port_bo, hstart, hstop, sample_int;
394         unsigned int rate, bps;
395
396         port_bo = 0;
397         hstart = 1;
398         hstop = bus->params.col - 1;
399         t_data.hstop = hstop;
400         t_data.hstart = hstart;
401
402         list_for_each_entry(m_rt, &bus->m_rt_list, bus_node) {
403                 rate = m_rt->stream->params.rate;
404                 bps = m_rt->stream->params.bps;
405                 sample_int = (bus->params.curr_dr_freq / rate);
406                 list_for_each_entry(p_rt, &m_rt->port_list, port_node) {
407                         port_bo = (p_rt->num * 64) + 1;
408                         dev_dbg(bus->dev, "p_rt->num=%d hstart=%d hstop=%d port_bo=%d\n",
409                                 p_rt->num, hstart, hstop, port_bo);
410                         sdw_fill_xport_params(&p_rt->transport_params, p_rt->num,
411                                               false, SDW_BLK_GRP_CNT_1, sample_int,
412                                               port_bo, port_bo >> 8, hstart, hstop,
413                                               SDW_BLK_PKG_PER_PORT, 0x0);
414
415                         sdw_fill_port_params(&p_rt->port_params,
416                                              p_rt->num, bps,
417                                              SDW_PORT_FLOW_MODE_ISOCH,
418                                              b_params->m_data_mode);
419                         t_data.hstart = hstart;
420                         t_data.hstop = hstop;
421                         t_data.block_offset = port_bo;
422                         t_data.sub_block_offset = 0;
423                 }
424                 sdw_compute_slave_ports(m_rt, &t_data);
425         }
426         return 0;
427 }
428
429 static int amd_sdw_port_params(struct sdw_bus *bus, struct sdw_port_params *p_params,
430                                unsigned int bank)
431 {
432         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
433         u32 frame_fmt_reg, dpn_frame_fmt;
434
435         dev_dbg(amd_manager->dev, "p_params->num:0x%x\n", p_params->num);
436         switch (amd_manager->instance) {
437         case ACP_SDW0:
438                 frame_fmt_reg = sdw0_manager_dp_reg[p_params->num].frame_fmt_reg;
439                 break;
440         case ACP_SDW1:
441                 frame_fmt_reg = sdw1_manager_dp_reg[p_params->num].frame_fmt_reg;
442                 break;
443         default:
444                 return -EINVAL;
445         }
446
447         dpn_frame_fmt = readl(amd_manager->mmio + frame_fmt_reg);
448         u32p_replace_bits(&dpn_frame_fmt, p_params->flow_mode, AMD_DPN_FRAME_FMT_PFM);
449         u32p_replace_bits(&dpn_frame_fmt, p_params->data_mode, AMD_DPN_FRAME_FMT_PDM);
450         u32p_replace_bits(&dpn_frame_fmt, p_params->bps - 1, AMD_DPN_FRAME_FMT_WORD_LEN);
451         writel(dpn_frame_fmt, amd_manager->mmio + frame_fmt_reg);
452         return 0;
453 }
454
455 static int amd_sdw_transport_params(struct sdw_bus *bus,
456                                     struct sdw_transport_params *params,
457                                     enum sdw_reg_bank bank)
458 {
459         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
460         u32 dpn_frame_fmt;
461         u32 dpn_sampleinterval;
462         u32 dpn_hctrl;
463         u32 dpn_offsetctrl;
464         u32 dpn_lanectrl;
465         u32 frame_fmt_reg, sample_int_reg, hctrl_dp0_reg;
466         u32 offset_reg, lane_ctrl_ch_en_reg;
467
468         switch (amd_manager->instance) {
469         case ACP_SDW0:
470                 frame_fmt_reg = sdw0_manager_dp_reg[params->port_num].frame_fmt_reg;
471                 sample_int_reg = sdw0_manager_dp_reg[params->port_num].sample_int_reg;
472                 hctrl_dp0_reg = sdw0_manager_dp_reg[params->port_num].hctrl_dp0_reg;
473                 offset_reg = sdw0_manager_dp_reg[params->port_num].offset_reg;
474                 lane_ctrl_ch_en_reg = sdw0_manager_dp_reg[params->port_num].lane_ctrl_ch_en_reg;
475                 break;
476         case ACP_SDW1:
477                 frame_fmt_reg = sdw1_manager_dp_reg[params->port_num].frame_fmt_reg;
478                 sample_int_reg = sdw1_manager_dp_reg[params->port_num].sample_int_reg;
479                 hctrl_dp0_reg = sdw1_manager_dp_reg[params->port_num].hctrl_dp0_reg;
480                 offset_reg = sdw1_manager_dp_reg[params->port_num].offset_reg;
481                 lane_ctrl_ch_en_reg = sdw1_manager_dp_reg[params->port_num].lane_ctrl_ch_en_reg;
482                 break;
483         default:
484                 return -EINVAL;
485         }
486         writel(AMD_SDW_SSP_COUNTER_VAL, amd_manager->mmio + ACP_SW_SSP_COUNTER);
487
488         dpn_frame_fmt = readl(amd_manager->mmio + frame_fmt_reg);
489         u32p_replace_bits(&dpn_frame_fmt, params->blk_pkg_mode, AMD_DPN_FRAME_FMT_BLK_PKG_MODE);
490         u32p_replace_bits(&dpn_frame_fmt, params->blk_grp_ctrl, AMD_DPN_FRAME_FMT_BLK_GRP_CTRL);
491         u32p_replace_bits(&dpn_frame_fmt, SDW_STREAM_PCM, AMD_DPN_FRAME_FMT_PCM_OR_PDM);
492         writel(dpn_frame_fmt, amd_manager->mmio + frame_fmt_reg);
493
494         dpn_sampleinterval = params->sample_interval - 1;
495         writel(dpn_sampleinterval, amd_manager->mmio + sample_int_reg);
496
497         dpn_hctrl = FIELD_PREP(AMD_DPN_HCTRL_HSTOP, params->hstop);
498         dpn_hctrl |= FIELD_PREP(AMD_DPN_HCTRL_HSTART, params->hstart);
499         writel(dpn_hctrl, amd_manager->mmio + hctrl_dp0_reg);
500
501         dpn_offsetctrl = FIELD_PREP(AMD_DPN_OFFSET_CTRL_1, params->offset1);
502         dpn_offsetctrl |= FIELD_PREP(AMD_DPN_OFFSET_CTRL_2, params->offset2);
503         writel(dpn_offsetctrl, amd_manager->mmio + offset_reg);
504
505         /*
506          * lane_ctrl_ch_en_reg will be used to program lane_ctrl and ch_mask
507          * parameters.
508          */
509         dpn_lanectrl = readl(amd_manager->mmio + lane_ctrl_ch_en_reg);
510         u32p_replace_bits(&dpn_lanectrl, params->lane_ctrl, AMD_DPN_CH_EN_LCTRL);
511         writel(dpn_lanectrl, amd_manager->mmio + lane_ctrl_ch_en_reg);
512         return 0;
513 }
514
515 static int amd_sdw_port_enable(struct sdw_bus *bus,
516                                struct sdw_enable_ch *enable_ch,
517                                unsigned int bank)
518 {
519         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
520         u32 dpn_ch_enable;
521         u32 lane_ctrl_ch_en_reg;
522
523         switch (amd_manager->instance) {
524         case ACP_SDW0:
525                 lane_ctrl_ch_en_reg = sdw0_manager_dp_reg[enable_ch->port_num].lane_ctrl_ch_en_reg;
526                 break;
527         case ACP_SDW1:
528                 lane_ctrl_ch_en_reg = sdw1_manager_dp_reg[enable_ch->port_num].lane_ctrl_ch_en_reg;
529                 break;
530         default:
531                 return -EINVAL;
532         }
533
534         /*
535          * lane_ctrl_ch_en_reg will be used to program lane_ctrl and ch_mask
536          * parameters.
537          */
538         dpn_ch_enable = readl(amd_manager->mmio + lane_ctrl_ch_en_reg);
539         u32p_replace_bits(&dpn_ch_enable, enable_ch->ch_mask, AMD_DPN_CH_EN_CHMASK);
540         if (enable_ch->enable)
541                 writel(dpn_ch_enable, amd_manager->mmio + lane_ctrl_ch_en_reg);
542         else
543                 writel(0, amd_manager->mmio + lane_ctrl_ch_en_reg);
544         return 0;
545 }
546
547 static int sdw_master_read_amd_prop(struct sdw_bus *bus)
548 {
549         struct amd_sdw_manager *amd_manager = to_amd_sdw(bus);
550         struct fwnode_handle *link;
551         struct sdw_master_prop *prop;
552         u32 quirk_mask = 0;
553         u32 wake_en_mask = 0;
554         u32 power_mode_mask = 0;
555         char name[32];
556
557         prop = &bus->prop;
558         /* Find manager handle */
559         snprintf(name, sizeof(name), "mipi-sdw-link-%d-subproperties", bus->link_id);
560         link = device_get_named_child_node(bus->dev, name);
561         if (!link) {
562                 dev_err(bus->dev, "Manager node %s not found\n", name);
563                 return -EIO;
564         }
565         fwnode_property_read_u32(link, "amd-sdw-enable", &quirk_mask);
566         if (!(quirk_mask & AMD_SDW_QUIRK_MASK_BUS_ENABLE))
567                 prop->hw_disabled = true;
568         prop->quirks = SDW_MASTER_QUIRKS_CLEAR_INITIAL_CLASH |
569                        SDW_MASTER_QUIRKS_CLEAR_INITIAL_PARITY;
570
571         fwnode_property_read_u32(link, "amd-sdw-wakeup-enable", &wake_en_mask);
572         amd_manager->wake_en_mask = wake_en_mask;
573         fwnode_property_read_u32(link, "amd-sdw-power-mode", &power_mode_mask);
574         amd_manager->power_mode_mask = power_mode_mask;
575         return 0;
576 }
577
578 static int amd_prop_read(struct sdw_bus *bus)
579 {
580         sdw_master_read_prop(bus);
581         sdw_master_read_amd_prop(bus);
582         return 0;
583 }
584
585 static const struct sdw_master_port_ops amd_sdw_port_ops = {
586         .dpn_set_port_params = amd_sdw_port_params,
587         .dpn_set_port_transport_params = amd_sdw_transport_params,
588         .dpn_port_enable_ch = amd_sdw_port_enable,
589 };
590
591 static const struct sdw_master_ops amd_sdw_ops = {
592         .read_prop = amd_prop_read,
593         .xfer_msg = amd_sdw_xfer_msg,
594         .read_ping_status = amd_sdw_read_ping_status,
595 };
596
597 static int amd_sdw_hw_params(struct snd_pcm_substream *substream,
598                              struct snd_pcm_hw_params *params,
599                              struct snd_soc_dai *dai)
600 {
601         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
602         struct sdw_amd_dai_runtime *dai_runtime;
603         struct sdw_stream_config sconfig;
604         struct sdw_port_config *pconfig;
605         int ch, dir;
606         int ret;
607
608         dai_runtime = amd_manager->dai_runtime_array[dai->id];
609         if (!dai_runtime)
610                 return -EIO;
611
612         ch = params_channels(params);
613         if (substream->stream == SNDRV_PCM_STREAM_CAPTURE)
614                 dir = SDW_DATA_DIR_RX;
615         else
616                 dir = SDW_DATA_DIR_TX;
617         dev_dbg(amd_manager->dev, "dir:%d dai->id:0x%x\n", dir, dai->id);
618
619         sconfig.direction = dir;
620         sconfig.ch_count = ch;
621         sconfig.frame_rate = params_rate(params);
622         sconfig.type = dai_runtime->stream_type;
623
624         sconfig.bps = snd_pcm_format_width(params_format(params));
625
626         /* Port configuration */
627         pconfig = kzalloc(sizeof(*pconfig), GFP_KERNEL);
628         if (!pconfig) {
629                 ret =  -ENOMEM;
630                 goto error;
631         }
632
633         pconfig->num = dai->id;
634         pconfig->ch_mask = (1 << ch) - 1;
635         ret = sdw_stream_add_master(&amd_manager->bus, &sconfig,
636                                     pconfig, 1, dai_runtime->stream);
637         if (ret)
638                 dev_err(amd_manager->dev, "add manager to stream failed:%d\n", ret);
639
640         kfree(pconfig);
641 error:
642         return ret;
643 }
644
645 static int amd_sdw_hw_free(struct snd_pcm_substream *substream, struct snd_soc_dai *dai)
646 {
647         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
648         struct sdw_amd_dai_runtime *dai_runtime;
649         int ret;
650
651         dai_runtime = amd_manager->dai_runtime_array[dai->id];
652         if (!dai_runtime)
653                 return -EIO;
654
655         ret = sdw_stream_remove_master(&amd_manager->bus, dai_runtime->stream);
656         if (ret < 0)
657                 dev_err(dai->dev, "remove manager from stream %s failed: %d\n",
658                         dai_runtime->stream->name, ret);
659         return ret;
660 }
661
662 static int amd_set_sdw_stream(struct snd_soc_dai *dai, void *stream, int direction)
663 {
664         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
665         struct sdw_amd_dai_runtime *dai_runtime;
666
667         dai_runtime = amd_manager->dai_runtime_array[dai->id];
668         if (stream) {
669                 /* first paranoia check */
670                 if (dai_runtime) {
671                         dev_err(dai->dev, "dai_runtime already allocated for dai %s\n", dai->name);
672                         return -EINVAL;
673                 }
674
675                 /* allocate and set dai_runtime info */
676                 dai_runtime = kzalloc(sizeof(*dai_runtime), GFP_KERNEL);
677                 if (!dai_runtime)
678                         return -ENOMEM;
679
680                 dai_runtime->stream_type = SDW_STREAM_PCM;
681                 dai_runtime->bus = &amd_manager->bus;
682                 dai_runtime->stream = stream;
683                 amd_manager->dai_runtime_array[dai->id] = dai_runtime;
684         } else {
685                 /* second paranoia check */
686                 if (!dai_runtime) {
687                         dev_err(dai->dev, "dai_runtime not allocated for dai %s\n", dai->name);
688                         return -EINVAL;
689                 }
690
691                 /* for NULL stream we release allocated dai_runtime */
692                 kfree(dai_runtime);
693                 amd_manager->dai_runtime_array[dai->id] = NULL;
694         }
695         return 0;
696 }
697
698 static int amd_pcm_set_sdw_stream(struct snd_soc_dai *dai, void *stream, int direction)
699 {
700         return amd_set_sdw_stream(dai, stream, direction);
701 }
702
703 static void *amd_get_sdw_stream(struct snd_soc_dai *dai, int direction)
704 {
705         struct amd_sdw_manager *amd_manager = snd_soc_dai_get_drvdata(dai);
706         struct sdw_amd_dai_runtime *dai_runtime;
707
708         dai_runtime = amd_manager->dai_runtime_array[dai->id];
709         if (!dai_runtime)
710                 return ERR_PTR(-EINVAL);
711
712         return dai_runtime->stream;
713 }
714
715 static const struct snd_soc_dai_ops amd_sdw_dai_ops = {
716         .hw_params = amd_sdw_hw_params,
717         .hw_free = amd_sdw_hw_free,
718         .set_stream = amd_pcm_set_sdw_stream,
719         .get_stream = amd_get_sdw_stream,
720 };
721
722 static const struct snd_soc_component_driver amd_sdw_dai_component = {
723         .name = "soundwire",
724 };
725
726 static int amd_sdw_register_dais(struct amd_sdw_manager *amd_manager)
727 {
728         struct sdw_amd_dai_runtime **dai_runtime_array;
729         struct snd_soc_dai_driver *dais;
730         struct snd_soc_pcm_stream *stream;
731         struct device *dev;
732         int i, num_dais;
733
734         dev = amd_manager->dev;
735         num_dais = amd_manager->num_dout_ports + amd_manager->num_din_ports;
736         dais = devm_kcalloc(dev, num_dais, sizeof(*dais), GFP_KERNEL);
737         if (!dais)
738                 return -ENOMEM;
739
740         dai_runtime_array = devm_kcalloc(dev, num_dais,
741                                          sizeof(struct sdw_amd_dai_runtime *),
742                                          GFP_KERNEL);
743         if (!dai_runtime_array)
744                 return -ENOMEM;
745         amd_manager->dai_runtime_array = dai_runtime_array;
746         for (i = 0; i < num_dais; i++) {
747                 dais[i].name = devm_kasprintf(dev, GFP_KERNEL, "SDW%d Pin%d", amd_manager->instance,
748                                               i);
749                 if (!dais[i].name)
750                         return -ENOMEM;
751                 if (i < amd_manager->num_dout_ports)
752                         stream = &dais[i].playback;
753                 else
754                         stream = &dais[i].capture;
755
756                 stream->channels_min = 2;
757                 stream->channels_max = 2;
758                 stream->rates = SNDRV_PCM_RATE_48000;
759                 stream->formats = SNDRV_PCM_FMTBIT_S16_LE;
760
761                 dais[i].ops = &amd_sdw_dai_ops;
762                 dais[i].id = i;
763         }
764
765         return devm_snd_soc_register_component(dev, &amd_sdw_dai_component,
766                                                dais, num_dais);
767 }
768
769 static void amd_sdw_update_slave_status_work(struct work_struct *work)
770 {
771         struct amd_sdw_manager *amd_manager =
772                 container_of(work, struct amd_sdw_manager, amd_sdw_work);
773         int retry_count = 0;
774
775         if (amd_manager->status[0] == SDW_SLAVE_ATTACHED) {
776                 writel(0, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
777                 writel(0, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
778         }
779
780 update_status:
781         sdw_handle_slave_status(&amd_manager->bus, amd_manager->status);
782         /*
783          * During the peripheral enumeration sequence, the SoundWire manager interrupts
784          * are masked. Once the device number programming is done for all peripherals,
785          * interrupts will be unmasked. Read the peripheral device status from ping command
786          * and process the response. This sequence will ensure all peripheral devices enumerated
787          * and initialized properly.
788          */
789         if (amd_manager->status[0] == SDW_SLAVE_ATTACHED) {
790                 if (retry_count++ < SDW_MAX_DEVICES) {
791                         writel(AMD_SDW_IRQ_MASK_0TO7, amd_manager->mmio +
792                                ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7);
793                         writel(AMD_SDW_IRQ_MASK_8TO11, amd_manager->mmio +
794                                ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11);
795                         amd_sdw_read_and_process_ping_status(amd_manager);
796                         goto update_status;
797                 } else {
798                         dev_err_ratelimited(amd_manager->dev,
799                                             "Device0 detected after %d iterations\n",
800                                             retry_count);
801                 }
802         }
803 }
804
805 static void amd_sdw_update_slave_status(u32 status_change_0to7, u32 status_change_8to11,
806                                         struct amd_sdw_manager *amd_manager)
807 {
808         u64 slave_stat;
809         u32 val;
810         int dev_index;
811
812         if (status_change_0to7 == AMD_SDW_SLAVE_0_ATTACHED)
813                 memset(amd_manager->status, 0, sizeof(amd_manager->status));
814         slave_stat = status_change_0to7;
815         slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STATUS_8TO_11, status_change_8to11) << 32;
816         dev_dbg(amd_manager->dev, "status_change_0to7:0x%x status_change_8to11:0x%x\n",
817                 status_change_0to7, status_change_8to11);
818         if (slave_stat) {
819                 for (dev_index = 0; dev_index <= SDW_MAX_DEVICES; ++dev_index) {
820                         if (slave_stat & AMD_SDW_MCP_SLAVE_STATUS_VALID_MASK(dev_index)) {
821                                 val = (slave_stat >> AMD_SDW_MCP_SLAVE_STAT_SHIFT_MASK(dev_index)) &
822                                       AMD_SDW_MCP_SLAVE_STATUS_MASK;
823                                 amd_sdw_fill_slave_status(amd_manager, dev_index, val);
824                         }
825                 }
826         }
827 }
828
829 static void amd_sdw_process_wake_event(struct amd_sdw_manager *amd_manager)
830 {
831         pm_request_resume(amd_manager->dev);
832         writel(0x00, amd_manager->acp_mmio + ACP_SW_WAKE_EN(amd_manager->instance));
833         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_8TO11);
834 }
835
836 static void amd_sdw_irq_thread(struct work_struct *work)
837 {
838         struct amd_sdw_manager *amd_manager =
839                         container_of(work, struct amd_sdw_manager, amd_sdw_irq_thread);
840         u32 status_change_8to11;
841         u32 status_change_0to7;
842
843         status_change_8to11 = readl(amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_8TO11);
844         status_change_0to7 = readl(amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_0TO7);
845         dev_dbg(amd_manager->dev, "[SDW%d] SDW INT: 0to7=0x%x, 8to11=0x%x\n",
846                 amd_manager->instance, status_change_0to7, status_change_8to11);
847         if (status_change_8to11 & AMD_SDW_WAKE_STAT_MASK)
848                 return amd_sdw_process_wake_event(amd_manager);
849
850         if (status_change_8to11 & AMD_SDW_PREQ_INTR_STAT) {
851                 amd_sdw_read_and_process_ping_status(amd_manager);
852         } else {
853                 /* Check for the updated status on peripheral device */
854                 amd_sdw_update_slave_status(status_change_0to7, status_change_8to11, amd_manager);
855         }
856         if (status_change_8to11 || status_change_0to7)
857                 schedule_work(&amd_manager->amd_sdw_work);
858         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_8TO11);
859         writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_0TO7);
860 }
861
862 int amd_sdw_manager_start(struct amd_sdw_manager *amd_manager)
863 {
864         struct sdw_master_prop *prop;
865         int ret;
866
867         prop = &amd_manager->bus.prop;
868         if (!prop->hw_disabled) {
869                 ret = amd_init_sdw_manager(amd_manager);
870                 if (ret)
871                         return ret;
872                 amd_enable_sdw_interrupts(amd_manager);
873                 ret = amd_enable_sdw_manager(amd_manager);
874                 if (ret)
875                         return ret;
876                 amd_sdw_set_frameshape(amd_manager);
877         }
878         /* Enable runtime PM */
879         pm_runtime_set_autosuspend_delay(amd_manager->dev, AMD_SDW_MASTER_SUSPEND_DELAY_MS);
880         pm_runtime_use_autosuspend(amd_manager->dev);
881         pm_runtime_mark_last_busy(amd_manager->dev);
882         pm_runtime_set_active(amd_manager->dev);
883         pm_runtime_enable(amd_manager->dev);
884         return 0;
885 }
886
887 static int amd_sdw_manager_probe(struct platform_device *pdev)
888 {
889         const struct acp_sdw_pdata *pdata = pdev->dev.platform_data;
890         struct resource *res;
891         struct device *dev = &pdev->dev;
892         struct sdw_master_prop *prop;
893         struct sdw_bus_params *params;
894         struct amd_sdw_manager *amd_manager;
895         int ret;
896
897         amd_manager = devm_kzalloc(dev, sizeof(struct amd_sdw_manager), GFP_KERNEL);
898         if (!amd_manager)
899                 return -ENOMEM;
900
901         res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
902         if (!res)
903                 return -ENOMEM;
904
905         amd_manager->acp_mmio = devm_ioremap(dev, res->start, resource_size(res));
906         if (!amd_manager->acp_mmio) {
907                 dev_err(dev, "mmio not found\n");
908                 return -ENOMEM;
909         }
910         amd_manager->instance = pdata->instance;
911         amd_manager->mmio = amd_manager->acp_mmio +
912                             (amd_manager->instance * SDW_MANAGER_REG_OFFSET);
913         amd_manager->acp_sdw_lock = pdata->acp_sdw_lock;
914         amd_manager->cols_index = sdw_find_col_index(AMD_SDW_DEFAULT_COLUMNS);
915         amd_manager->rows_index = sdw_find_row_index(AMD_SDW_DEFAULT_ROWS);
916         amd_manager->dev = dev;
917         amd_manager->bus.ops = &amd_sdw_ops;
918         amd_manager->bus.port_ops = &amd_sdw_port_ops;
919         amd_manager->bus.compute_params = &amd_sdw_compute_params;
920         amd_manager->bus.clk_stop_timeout = 200;
921         amd_manager->bus.link_id = amd_manager->instance;
922
923         /*
924          * Due to BIOS compatibility, the two links are exposed within
925          * the scope of a single controller. If this changes, the
926          * controller_id will have to be updated with drv_data
927          * information.
928          */
929         amd_manager->bus.controller_id = 0;
930
931         switch (amd_manager->instance) {
932         case ACP_SDW0:
933                 amd_manager->num_dout_ports = AMD_SDW0_MAX_TX_PORTS;
934                 amd_manager->num_din_ports = AMD_SDW0_MAX_RX_PORTS;
935                 break;
936         case ACP_SDW1:
937                 amd_manager->num_dout_ports = AMD_SDW1_MAX_TX_PORTS;
938                 amd_manager->num_din_ports = AMD_SDW1_MAX_RX_PORTS;
939                 break;
940         default:
941                 return -EINVAL;
942         }
943
944         params = &amd_manager->bus.params;
945
946         params->col = AMD_SDW_DEFAULT_COLUMNS;
947         params->row = AMD_SDW_DEFAULT_ROWS;
948         prop = &amd_manager->bus.prop;
949         prop->clk_freq = &amd_sdw_freq_tbl[0];
950         prop->mclk_freq = AMD_SDW_BUS_BASE_FREQ;
951         prop->max_clk_freq = AMD_SDW_DEFAULT_CLK_FREQ;
952
953         ret = sdw_bus_master_add(&amd_manager->bus, dev, dev->fwnode);
954         if (ret) {
955                 dev_err(dev, "Failed to register SoundWire manager(%d)\n", ret);
956                 return ret;
957         }
958         ret = amd_sdw_register_dais(amd_manager);
959         if (ret) {
960                 dev_err(dev, "CPU DAI registration failed\n");
961                 sdw_bus_master_delete(&amd_manager->bus);
962                 return ret;
963         }
964         dev_set_drvdata(dev, amd_manager);
965         INIT_WORK(&amd_manager->amd_sdw_irq_thread, amd_sdw_irq_thread);
966         INIT_WORK(&amd_manager->amd_sdw_work, amd_sdw_update_slave_status_work);
967         return 0;
968 }
969
970 static void amd_sdw_manager_remove(struct platform_device *pdev)
971 {
972         struct amd_sdw_manager *amd_manager = dev_get_drvdata(&pdev->dev);
973         int ret;
974
975         pm_runtime_disable(&pdev->dev);
976         amd_disable_sdw_interrupts(amd_manager);
977         sdw_bus_master_delete(&amd_manager->bus);
978         ret = amd_disable_sdw_manager(amd_manager);
979         if (ret)
980                 dev_err(&pdev->dev, "Failed to disable device (%pe)\n", ERR_PTR(ret));
981 }
982
983 static int amd_sdw_clock_stop(struct amd_sdw_manager *amd_manager)
984 {
985         u32 val;
986         int ret;
987
988         ret = sdw_bus_prep_clk_stop(&amd_manager->bus);
989         if (ret < 0 && ret != -ENODATA) {
990                 dev_err(amd_manager->dev, "prepare clock stop failed %d", ret);
991                 return 0;
992         }
993         ret = sdw_bus_clk_stop(&amd_manager->bus);
994         if (ret < 0 && ret != -ENODATA) {
995                 dev_err(amd_manager->dev, "bus clock stop failed %d", ret);
996                 return 0;
997         }
998
999         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL, val,
1000                                  (val & AMD_SDW_CLK_STOP_DONE), ACP_DELAY_US, AMD_SDW_TIMEOUT);
1001         if (ret) {
1002                 dev_err(amd_manager->dev, "SDW%x clock stop failed\n", amd_manager->instance);
1003                 return 0;
1004         }
1005
1006         amd_manager->clk_stopped = true;
1007         if (amd_manager->wake_en_mask)
1008                 writel(0x01, amd_manager->acp_mmio + ACP_SW_WAKE_EN(amd_manager->instance));
1009
1010         dev_dbg(amd_manager->dev, "SDW%x clock stop successful\n", amd_manager->instance);
1011         return 0;
1012 }
1013
1014 static int amd_sdw_clock_stop_exit(struct amd_sdw_manager *amd_manager)
1015 {
1016         int ret;
1017         u32 val;
1018
1019         if (amd_manager->clk_stopped) {
1020                 val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1021                 val |= AMD_SDW_CLK_RESUME_REQ;
1022                 writel(val, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1023                 ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL, val,
1024                                          (val & AMD_SDW_CLK_RESUME_DONE), ACP_DELAY_US,
1025                                          AMD_SDW_TIMEOUT);
1026                 if (val & AMD_SDW_CLK_RESUME_DONE) {
1027                         writel(0, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1028                         ret = sdw_bus_exit_clk_stop(&amd_manager->bus);
1029                         if (ret < 0)
1030                                 dev_err(amd_manager->dev, "bus failed to exit clock stop %d\n",
1031                                         ret);
1032                         amd_manager->clk_stopped = false;
1033                 }
1034         }
1035         if (amd_manager->clk_stopped) {
1036                 dev_err(amd_manager->dev, "SDW%x clock stop exit failed\n", amd_manager->instance);
1037                 return 0;
1038         }
1039         dev_dbg(amd_manager->dev, "SDW%x clock stop exit successful\n", amd_manager->instance);
1040         return 0;
1041 }
1042
1043 static int amd_resume_child_device(struct device *dev, void *data)
1044 {
1045         struct sdw_slave *slave = dev_to_sdw_dev(dev);
1046         int ret;
1047
1048         if (!slave->probed) {
1049                 dev_dbg(dev, "skipping device, no probed driver\n");
1050                 return 0;
1051         }
1052         if (!slave->dev_num_sticky) {
1053                 dev_dbg(dev, "skipping device, never detected on bus\n");
1054                 return 0;
1055         }
1056         ret = pm_request_resume(dev);
1057         if (ret < 0) {
1058                 dev_err(dev, "pm_request_resume failed: %d\n", ret);
1059                 return ret;
1060         }
1061         return 0;
1062 }
1063
1064 static int __maybe_unused amd_pm_prepare(struct device *dev)
1065 {
1066         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1067         struct sdw_bus *bus = &amd_manager->bus;
1068         int ret;
1069
1070         if (bus->prop.hw_disabled) {
1071                 dev_dbg(bus->dev, "SoundWire manager %d is disabled, ignoring\n",
1072                         bus->link_id);
1073                 return 0;
1074         }
1075         /*
1076          * When multiple peripheral devices connected over the same link, if SoundWire manager
1077          * device is not in runtime suspend state, observed that device alerts are missing
1078          * without pm_prepare on AMD platforms in clockstop mode0.
1079          */
1080         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1081                 ret = pm_request_resume(dev);
1082                 if (ret < 0) {
1083                         dev_err(bus->dev, "pm_request_resume failed: %d\n", ret);
1084                         return 0;
1085                 }
1086         }
1087         /* To force peripheral devices to system level suspend state, resume the devices
1088          * from runtime suspend state first. Without that unable to dispatch the alert
1089          * status to peripheral driver during system level resume as they are in runtime
1090          * suspend state.
1091          */
1092         ret = device_for_each_child(bus->dev, NULL, amd_resume_child_device);
1093         if (ret < 0)
1094                 dev_err(dev, "amd_resume_child_device failed: %d\n", ret);
1095         return 0;
1096 }
1097
1098 static int __maybe_unused amd_suspend(struct device *dev)
1099 {
1100         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1101         struct sdw_bus *bus = &amd_manager->bus;
1102         int ret;
1103
1104         if (bus->prop.hw_disabled) {
1105                 dev_dbg(bus->dev, "SoundWire manager %d is disabled, ignoring\n",
1106                         bus->link_id);
1107                 return 0;
1108         }
1109
1110         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1111                 amd_sdw_wake_enable(amd_manager, false);
1112                 return amd_sdw_clock_stop(amd_manager);
1113         } else if (amd_manager->power_mode_mask & AMD_SDW_POWER_OFF_MODE) {
1114                 /*
1115                  * As per hardware programming sequence on AMD platforms,
1116                  * clock stop should be invoked first before powering-off
1117                  */
1118                 ret = amd_sdw_clock_stop(amd_manager);
1119                 if (ret)
1120                         return ret;
1121                 return amd_deinit_sdw_manager(amd_manager);
1122         }
1123         return 0;
1124 }
1125
1126 static int __maybe_unused amd_suspend_runtime(struct device *dev)
1127 {
1128         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1129         struct sdw_bus *bus = &amd_manager->bus;
1130         int ret;
1131
1132         if (bus->prop.hw_disabled) {
1133                 dev_dbg(bus->dev, "SoundWire manager %d is disabled,\n",
1134                         bus->link_id);
1135                 return 0;
1136         }
1137         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1138                 amd_sdw_wake_enable(amd_manager, true);
1139                 return amd_sdw_clock_stop(amd_manager);
1140         } else if (amd_manager->power_mode_mask & AMD_SDW_POWER_OFF_MODE) {
1141                 ret = amd_sdw_clock_stop(amd_manager);
1142                 if (ret)
1143                         return ret;
1144                 return amd_deinit_sdw_manager(amd_manager);
1145         }
1146         return 0;
1147 }
1148
1149 static int __maybe_unused amd_resume_runtime(struct device *dev)
1150 {
1151         struct amd_sdw_manager *amd_manager = dev_get_drvdata(dev);
1152         struct sdw_bus *bus = &amd_manager->bus;
1153         int ret;
1154         u32 val;
1155
1156         if (bus->prop.hw_disabled) {
1157                 dev_dbg(bus->dev, "SoundWire manager %d is disabled, ignoring\n",
1158                         bus->link_id);
1159                 return 0;
1160         }
1161
1162         if (amd_manager->power_mode_mask & AMD_SDW_CLK_STOP_MODE) {
1163                 return amd_sdw_clock_stop_exit(amd_manager);
1164         } else if (amd_manager->power_mode_mask & AMD_SDW_POWER_OFF_MODE) {
1165                 val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1166                 if (val) {
1167                         val |= AMD_SDW_CLK_RESUME_REQ;
1168                         writel(val, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1169                         ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL, val,
1170                                                  (val & AMD_SDW_CLK_RESUME_DONE), ACP_DELAY_US,
1171                                                  AMD_SDW_TIMEOUT);
1172                         if (val & AMD_SDW_CLK_RESUME_DONE) {
1173                                 writel(0, amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL);
1174                                 amd_manager->clk_stopped = false;
1175                         }
1176                 }
1177                 sdw_clear_slave_status(bus, SDW_UNATTACH_REQUEST_MASTER_RESET);
1178                 amd_init_sdw_manager(amd_manager);
1179                 amd_enable_sdw_interrupts(amd_manager);
1180                 ret = amd_enable_sdw_manager(amd_manager);
1181                 if (ret)
1182                         return ret;
1183                 amd_sdw_set_frameshape(amd_manager);
1184         }
1185         return 0;
1186 }
1187
1188 static const struct dev_pm_ops amd_pm = {
1189         .prepare = amd_pm_prepare,
1190         SET_SYSTEM_SLEEP_PM_OPS(amd_suspend, amd_resume_runtime)
1191         SET_RUNTIME_PM_OPS(amd_suspend_runtime, amd_resume_runtime, NULL)
1192 };
1193
1194 static struct platform_driver amd_sdw_driver = {
1195         .probe  = &amd_sdw_manager_probe,
1196         .remove_new = &amd_sdw_manager_remove,
1197         .driver = {
1198                 .name   = "amd_sdw_manager",
1199                 .pm = &amd_pm,
1200         }
1201 };
1202 module_platform_driver(amd_sdw_driver);
1203
1204 MODULE_AUTHOR("Vijendar.Mukunda@amd.com");
1205 MODULE_DESCRIPTION("AMD SoundWire driver");
1206 MODULE_LICENSE("Dual BSD/GPL");
1207 MODULE_ALIAS("platform:" DRV_NAME);