1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (C) 2022 MediaTek Inc.
4 * Author: Lorenzo Bianconi <lorenzo@kernel.org>
5 * Sujuan Chen <sujuan.chen@mediatek.com>
8 #include <linux/firmware.h>
9 #include <linux/of_address.h>
10 #include <linux/of_reserved_mem.h>
11 #include <linux/mfd/syscon.h>
12 #include <linux/soc/mediatek/mtk_wed.h>
13 #include <asm/unaligned.h>
15 #include "mtk_wed_regs.h"
16 #include "mtk_wed_wo.h"
19 static struct mtk_wed_wo_memory_region mem_region[] = {
20 [MTK_WED_WO_REGION_EMI] = {
23 [MTK_WED_WO_REGION_ILM] = {
26 [MTK_WED_WO_REGION_DATA] = {
30 [MTK_WED_WO_REGION_BOOT] = {
35 static u32 wo_r32(u32 reg)
37 return readl(mem_region[MTK_WED_WO_REGION_BOOT].addr + reg);
40 static void wo_w32(u32 reg, u32 val)
42 writel(val, mem_region[MTK_WED_WO_REGION_BOOT].addr + reg);
45 static struct sk_buff *
46 mtk_wed_mcu_msg_alloc(const void *data, int data_len)
48 int length = sizeof(struct mtk_wed_mcu_hdr) + data_len;
51 skb = alloc_skb(length, GFP_KERNEL);
55 memset(skb->head, 0, length);
56 skb_reserve(skb, sizeof(struct mtk_wed_mcu_hdr));
58 skb_put_data(skb, data, data_len);
63 static struct sk_buff *
64 mtk_wed_mcu_get_response(struct mtk_wed_wo *wo, unsigned long expires)
66 if (!time_is_after_jiffies(expires))
69 wait_event_timeout(wo->mcu.wait, !skb_queue_empty(&wo->mcu.res_q),
71 return skb_dequeue(&wo->mcu.res_q);
74 void mtk_wed_mcu_rx_event(struct mtk_wed_wo *wo, struct sk_buff *skb)
76 skb_queue_tail(&wo->mcu.res_q, skb);
77 wake_up(&wo->mcu.wait);
81 mtk_wed_update_rx_stats(struct mtk_wed_device *wed, struct sk_buff *skb)
83 u32 count = get_unaligned_le32(skb->data);
84 struct mtk_wed_wo_rx_stats *stats;
87 if (!wed->wlan.update_wo_rx_stats)
90 if (count * sizeof(*stats) > skb->len - sizeof(u32))
93 stats = (struct mtk_wed_wo_rx_stats *)(skb->data + sizeof(u32));
94 for (i = 0 ; i < count ; i++)
95 wed->wlan.update_wo_rx_stats(wed, &stats[i]);
98 void mtk_wed_mcu_rx_unsolicited_event(struct mtk_wed_wo *wo,
101 struct mtk_wed_mcu_hdr *hdr = (struct mtk_wed_mcu_hdr *)skb->data;
103 skb_pull(skb, sizeof(*hdr));
106 case MTK_WED_WO_EVT_LOG_DUMP:
107 dev_notice(wo->hw->dev, "%s\n", skb->data);
109 case MTK_WED_WO_EVT_PROFILING: {
110 struct mtk_wed_wo_log_info *info = (void *)skb->data;
111 u32 count = skb->len / sizeof(*info);
114 for (i = 0 ; i < count ; i++)
115 dev_notice(wo->hw->dev,
116 "SN:%u latency: total=%u, rro:%u, mod:%u\n",
117 le32_to_cpu(info[i].sn),
118 le32_to_cpu(info[i].total),
119 le32_to_cpu(info[i].rro),
120 le32_to_cpu(info[i].mod));
123 case MTK_WED_WO_EVT_RXCNT_INFO:
124 mtk_wed_update_rx_stats(wo->hw->wed_dev, skb);
134 mtk_wed_mcu_skb_send_msg(struct mtk_wed_wo *wo, struct sk_buff *skb,
135 int id, int cmd, u16 *wait_seq, bool wait_resp)
137 struct mtk_wed_mcu_hdr *hdr;
139 /* TODO: make it dynamic based on cmd */
140 wo->mcu.timeout = 20 * HZ;
142 hdr = (struct mtk_wed_mcu_hdr *)skb_push(skb, sizeof(*hdr));
144 hdr->length = cpu_to_le16(skb->len);
146 if (wait_resp && wait_seq) {
147 u16 seq = ++wo->mcu.seq;
153 hdr->flag |= cpu_to_le16(MTK_WED_WARP_CMD_FLAG_NEED_RSP);
154 hdr->seq = cpu_to_le16(seq);
156 if (id == MTK_WED_MODULE_ID_WO)
157 hdr->flag |= cpu_to_le16(MTK_WED_WARP_CMD_FLAG_FROM_TO_WO);
159 return mtk_wed_wo_queue_tx_skb(wo, &wo->q_tx, skb);
163 mtk_wed_mcu_parse_response(struct mtk_wed_wo *wo, struct sk_buff *skb,
166 struct mtk_wed_mcu_hdr *hdr;
169 dev_err(wo->hw->dev, "Message %08x (seq %d) timeout\n",
174 hdr = (struct mtk_wed_mcu_hdr *)skb->data;
175 if (le16_to_cpu(hdr->seq) != seq)
178 skb_pull(skb, sizeof(*hdr));
180 case MTK_WED_WO_CMD_RXCNT_INFO:
181 mtk_wed_update_rx_stats(wo->hw->wed_dev, skb);
190 int mtk_wed_mcu_send_msg(struct mtk_wed_wo *wo, int id, int cmd,
191 const void *data, int len, bool wait_resp)
193 unsigned long expires;
198 skb = mtk_wed_mcu_msg_alloc(data, len);
202 mutex_lock(&wo->mcu.mutex);
204 ret = mtk_wed_mcu_skb_send_msg(wo, skb, id, cmd, &seq, wait_resp);
205 if (ret || !wait_resp)
208 expires = jiffies + wo->mcu.timeout;
210 skb = mtk_wed_mcu_get_response(wo, expires);
211 ret = mtk_wed_mcu_parse_response(wo, skb, cmd, seq);
213 } while (ret == -EAGAIN);
216 mutex_unlock(&wo->mcu.mutex);
221 int mtk_wed_mcu_msg_update(struct mtk_wed_device *dev, int id, void *data,
224 struct mtk_wed_wo *wo = dev->hw->wed_wo;
226 if (!mtk_wed_get_rx_capa(dev))
232 return mtk_wed_mcu_send_msg(wo, MTK_WED_MODULE_ID_WO, id, data, len,
237 mtk_wed_get_memory_region(struct mtk_wed_hw *hw, int index,
238 struct mtk_wed_wo_memory_region *region)
240 struct reserved_mem *rmem;
241 struct device_node *np;
243 np = of_parse_phandle(hw->node, "memory-region", index);
247 rmem = of_reserved_mem_lookup(np);
253 region->phy_addr = rmem->base;
254 region->size = rmem->size;
255 region->addr = devm_ioremap(hw->dev, region->phy_addr, region->size);
257 return !region->addr ? -EINVAL : 0;
261 mtk_wed_mcu_run_firmware(struct mtk_wed_wo *wo, const struct firmware *fw)
263 const u8 *first_region_ptr, *region_ptr, *trailer_ptr, *ptr = fw->data;
264 const struct mtk_wed_fw_trailer *trailer;
265 const struct mtk_wed_fw_region *fw_region;
267 trailer_ptr = fw->data + fw->size - sizeof(*trailer);
268 trailer = (const struct mtk_wed_fw_trailer *)trailer_ptr;
269 region_ptr = trailer_ptr - trailer->num_region * sizeof(*fw_region);
270 first_region_ptr = region_ptr;
272 while (region_ptr < trailer_ptr) {
276 fw_region = (const struct mtk_wed_fw_region *)region_ptr;
277 length = le32_to_cpu(fw_region->len);
278 if (first_region_ptr < ptr + length)
281 for (i = 0; i < ARRAY_SIZE(mem_region); i++) {
282 struct mtk_wed_wo_memory_region *region;
284 region = &mem_region[i];
285 if (region->phy_addr != le32_to_cpu(fw_region->addr))
288 if (region->size < length)
291 if (region->shared && region->consumed)
294 if (!region->shared || !region->consumed) {
295 memcpy_toio(region->addr, ptr, length);
296 region->consumed = true;
301 if (i == ARRAY_SIZE(mem_region))
304 region_ptr += sizeof(*fw_region);
312 mtk_wed_mcu_load_firmware(struct mtk_wed_wo *wo)
314 const struct mtk_wed_fw_trailer *trailer;
315 const struct firmware *fw;
320 /* load firmware region metadata */
321 for (i = 0; i < ARRAY_SIZE(mem_region); i++) {
322 int index = of_property_match_string(wo->hw->node,
323 "memory-region-names",
328 ret = mtk_wed_get_memory_region(wo->hw, index, &mem_region[i]);
334 wed_w32(wo->hw->wed_dev, MTK_WED_SCR0 + 4 * MTK_WED_DUMMY_CR_FWDL,
338 switch (wo->hw->version) {
340 if (of_device_is_compatible(wo->hw->node,
341 "mediatek,mt7981-wed"))
342 fw_name = MT7981_FIRMWARE_WO;
344 fw_name = wo->hw->index ? MT7986_FIRMWARE_WO1
345 : MT7986_FIRMWARE_WO0;
348 fw_name = wo->hw->index ? MT7988_FIRMWARE_WO1
349 : MT7988_FIRMWARE_WO0;
355 ret = reject_firmware(&fw, fw_name, wo->hw->dev);
359 trailer = (void *)(fw->data + fw->size -
360 sizeof(struct mtk_wed_fw_trailer));
361 dev_info(wo->hw->dev,
362 "MTK WED WO Firmware Version: %.10s, Build Time: %.15s\n",
363 trailer->fw_ver, trailer->build_date);
364 dev_info(wo->hw->dev, "MTK WED WO Chip ID %02x Region %d\n",
365 trailer->chip_id, trailer->num_region);
367 ret = mtk_wed_mcu_run_firmware(wo, fw);
371 /* set the start address */
372 if (!mtk_wed_is_v3_or_greater(wo->hw) && wo->hw->index)
373 boot_cr = MTK_WO_MCU_CFG_LS_WA_BOOT_ADDR_ADDR;
375 boot_cr = MTK_WO_MCU_CFG_LS_WM_BOOT_ADDR_ADDR;
376 wo_w32(boot_cr, mem_region[MTK_WED_WO_REGION_EMI].phy_addr >> 16);
377 /* wo firmware reset */
378 wo_w32(MTK_WO_MCU_CFG_LS_WF_MCCR_CLR_ADDR, 0xc00);
380 val = wo_r32(MTK_WO_MCU_CFG_LS_WF_MCU_CFG_WM_WA_ADDR) |
381 MTK_WO_MCU_CFG_LS_WF_WM_WA_WM_CPU_RSTB_MASK;
382 wo_w32(MTK_WO_MCU_CFG_LS_WF_MCU_CFG_WM_WA_ADDR, val);
384 release_firmware(fw);
390 mtk_wed_mcu_read_fw_dl(struct mtk_wed_wo *wo)
392 return wed_r32(wo->hw->wed_dev,
393 MTK_WED_SCR0 + 4 * MTK_WED_DUMMY_CR_FWDL);
396 int mtk_wed_mcu_init(struct mtk_wed_wo *wo)
401 skb_queue_head_init(&wo->mcu.res_q);
402 init_waitqueue_head(&wo->mcu.wait);
403 mutex_init(&wo->mcu.mutex);
405 ret = mtk_wed_mcu_load_firmware(wo);
409 return readx_poll_timeout(mtk_wed_mcu_read_fw_dl, wo, val, !val,
410 100, MTK_FW_DL_TIMEOUT);