1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
2 /* Copyright (C) 2021 Martin Blumenstingl <martin.blumenstingl@googlemail.com>
3 * Copyright (C) 2021 Jernej Skrabec <jernej.skrabec@gmail.com>
4 *
5 * Based on rtw88/pci.c:
6 * Copyright(c) 2018-2019 Realtek Corporation
7 */
8
9 #include <linux/module.h>
10 #include <linux/mmc/host.h>
11 #include <linux/mmc/sdio_func.h>
12 #include "main.h"
13 #include "mac.h"
14 #include "debug.h"
15 #include "fw.h"
16 #include "ps.h"
17 #include "reg.h"
18 #include "rx.h"
19 #include "sdio.h"
20 #include "tx.h"
21
22 #define RTW_SDIO_INDIRECT_RW_RETRIES 50
23
rtw_sdio_is_bus_addr(u32 addr)24 static bool rtw_sdio_is_bus_addr(u32 addr)
25 {
26 return !!(addr & RTW_SDIO_BUS_MSK);
27 }
28
rtw_sdio_bus_claim_needed(struct rtw_sdio * rtwsdio)29 static bool rtw_sdio_bus_claim_needed(struct rtw_sdio *rtwsdio)
30 {
31 return !rtwsdio->irq_thread ||
32 rtwsdio->irq_thread != current;
33 }
34
rtw_sdio_to_bus_offset(struct rtw_dev * rtwdev,u32 addr)35 static u32 rtw_sdio_to_bus_offset(struct rtw_dev *rtwdev, u32 addr)
36 {
37 switch (addr & RTW_SDIO_BUS_MSK) {
38 case WLAN_IOREG_OFFSET:
39 addr &= WLAN_IOREG_REG_MSK;
40 addr |= FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
41 REG_SDIO_CMD_ADDR_MAC_REG);
42 break;
43 case SDIO_LOCAL_OFFSET:
44 addr &= SDIO_LOCAL_REG_MSK;
45 addr |= FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
46 REG_SDIO_CMD_ADDR_SDIO_REG);
47 break;
48 default:
49 rtw_warn(rtwdev, "Cannot convert addr 0x%08x to bus offset",
50 addr);
51 }
52
53 return addr;
54 }
55
rtw_sdio_use_memcpy_io(struct rtw_dev * rtwdev,u32 addr,u8 alignment)56 static bool rtw_sdio_use_memcpy_io(struct rtw_dev *rtwdev, u32 addr,
57 u8 alignment)
58 {
59 return IS_ALIGNED(addr, alignment) &&
60 test_bit(RTW_FLAG_POWERON, rtwdev->flags);
61 }
62
rtw_sdio_writel(struct rtw_dev * rtwdev,u32 val,u32 addr,int * err_ret)63 static void rtw_sdio_writel(struct rtw_dev *rtwdev, u32 val, u32 addr,
64 int *err_ret)
65 {
66 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
67 u8 buf[4];
68 int i;
69
70 if (rtw_sdio_use_memcpy_io(rtwdev, addr, 4)) {
71 sdio_writel(rtwsdio->sdio_func, val, addr, err_ret);
72 return;
73 }
74
75 *(__le32 *)buf = cpu_to_le32(val);
76
77 for (i = 0; i < 4; i++) {
78 sdio_writeb(rtwsdio->sdio_func, buf[i], addr + i, err_ret);
79 if (*err_ret)
80 return;
81 }
82 }
83
rtw_sdio_writew(struct rtw_dev * rtwdev,u16 val,u32 addr,int * err_ret)84 static void rtw_sdio_writew(struct rtw_dev *rtwdev, u16 val, u32 addr,
85 int *err_ret)
86 {
87 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
88 u8 buf[2];
89 int i;
90
91 *(__le16 *)buf = cpu_to_le16(val);
92
93 for (i = 0; i < 2; i++) {
94 sdio_writeb(rtwsdio->sdio_func, buf[i], addr + i, err_ret);
95 if (*err_ret)
96 return;
97 }
98 }
99
rtw_sdio_readl(struct rtw_dev * rtwdev,u32 addr,int * err_ret)100 static u32 rtw_sdio_readl(struct rtw_dev *rtwdev, u32 addr, int *err_ret)
101 {
102 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
103 u8 buf[4];
104 int i;
105
106 if (rtw_sdio_use_memcpy_io(rtwdev, addr, 4))
107 return sdio_readl(rtwsdio->sdio_func, addr, err_ret);
108
109 for (i = 0; i < 4; i++) {
110 buf[i] = sdio_readb(rtwsdio->sdio_func, addr + i, err_ret);
111 if (*err_ret)
112 return 0;
113 }
114
115 return le32_to_cpu(*(__le32 *)buf);
116 }
117
rtw_sdio_readw(struct rtw_dev * rtwdev,u32 addr,int * err_ret)118 static u16 rtw_sdio_readw(struct rtw_dev *rtwdev, u32 addr, int *err_ret)
119 {
120 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
121 u8 buf[2];
122 int i;
123
124 for (i = 0; i < 2; i++) {
125 buf[i] = sdio_readb(rtwsdio->sdio_func, addr + i, err_ret);
126 if (*err_ret)
127 return 0;
128 }
129
130 return le16_to_cpu(*(__le16 *)buf);
131 }
132
rtw_sdio_to_io_address(struct rtw_dev * rtwdev,u32 addr,bool direct)133 static u32 rtw_sdio_to_io_address(struct rtw_dev *rtwdev, u32 addr,
134 bool direct)
135 {
136 if (!direct)
137 return addr;
138
139 if (!rtw_sdio_is_bus_addr(addr))
140 addr |= WLAN_IOREG_OFFSET;
141
142 return rtw_sdio_to_bus_offset(rtwdev, addr);
143 }
144
rtw_sdio_use_direct_io(struct rtw_dev * rtwdev,u32 addr)145 static bool rtw_sdio_use_direct_io(struct rtw_dev *rtwdev, u32 addr)
146 {
147 return !rtw_sdio_is_sdio30_supported(rtwdev) ||
148 rtw_sdio_is_bus_addr(addr);
149 }
150
rtw_sdio_indirect_reg_cfg(struct rtw_dev * rtwdev,u32 addr,u32 cfg)151 static int rtw_sdio_indirect_reg_cfg(struct rtw_dev *rtwdev, u32 addr, u32 cfg)
152 {
153 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
154 unsigned int retry;
155 u32 reg_cfg;
156 int ret;
157 u8 tmp;
158
159 reg_cfg = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_CFG);
160
161 rtw_sdio_writel(rtwdev, addr | cfg | BIT_SDIO_INDIRECT_REG_CFG_UNK20,
162 reg_cfg, &ret);
163 if (ret)
164 return ret;
165
166 for (retry = 0; retry < RTW_SDIO_INDIRECT_RW_RETRIES; retry++) {
167 tmp = sdio_readb(rtwsdio->sdio_func, reg_cfg + 2, &ret);
168 if (!ret && (tmp & BIT(4)))
169 return 0;
170 }
171
172 return -ETIMEDOUT;
173 }
174
rtw_sdio_indirect_read8(struct rtw_dev * rtwdev,u32 addr,int * err_ret)175 static u8 rtw_sdio_indirect_read8(struct rtw_dev *rtwdev, u32 addr,
176 int *err_ret)
177 {
178 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
179 u32 reg_data;
180
181 *err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
182 BIT_SDIO_INDIRECT_REG_CFG_READ);
183 if (*err_ret)
184 return 0;
185
186 reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
187 return sdio_readb(rtwsdio->sdio_func, reg_data, err_ret);
188 }
189
rtw_sdio_indirect_read_bytes(struct rtw_dev * rtwdev,u32 addr,u8 * buf,int count)190 static int rtw_sdio_indirect_read_bytes(struct rtw_dev *rtwdev, u32 addr,
191 u8 *buf, int count)
192 {
193 int i, ret = 0;
194
195 for (i = 0; i < count; i++) {
196 buf[i] = rtw_sdio_indirect_read8(rtwdev, addr + i, &ret);
197 if (ret)
198 break;
199 }
200
201 return ret;
202 }
203
rtw_sdio_indirect_read16(struct rtw_dev * rtwdev,u32 addr,int * err_ret)204 static u16 rtw_sdio_indirect_read16(struct rtw_dev *rtwdev, u32 addr,
205 int *err_ret)
206 {
207 u32 reg_data;
208 u8 buf[2];
209
210 if (!IS_ALIGNED(addr, 2)) {
211 *err_ret = rtw_sdio_indirect_read_bytes(rtwdev, addr, buf, 2);
212 if (*err_ret)
213 return 0;
214
215 return le16_to_cpu(*(__le16 *)buf);
216 }
217
218 *err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
219 BIT_SDIO_INDIRECT_REG_CFG_READ);
220 if (*err_ret)
221 return 0;
222
223 reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
224 return rtw_sdio_readw(rtwdev, reg_data, err_ret);
225 }
226
rtw_sdio_indirect_read32(struct rtw_dev * rtwdev,u32 addr,int * err_ret)227 static u32 rtw_sdio_indirect_read32(struct rtw_dev *rtwdev, u32 addr,
228 int *err_ret)
229 {
230 u32 reg_data;
231 u8 buf[4];
232
233 if (!IS_ALIGNED(addr, 4)) {
234 *err_ret = rtw_sdio_indirect_read_bytes(rtwdev, addr, buf, 4);
235 if (*err_ret)
236 return 0;
237
238 return le32_to_cpu(*(__le32 *)buf);
239 }
240
241 *err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
242 BIT_SDIO_INDIRECT_REG_CFG_READ);
243 if (*err_ret)
244 return 0;
245
246 reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
247 return rtw_sdio_readl(rtwdev, reg_data, err_ret);
248 }
249
rtw_sdio_read8(struct rtw_dev * rtwdev,u32 addr)250 static u8 rtw_sdio_read8(struct rtw_dev *rtwdev, u32 addr)
251 {
252 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
253 bool direct, bus_claim;
254 int ret;
255 u8 val;
256
257 direct = rtw_sdio_use_direct_io(rtwdev, addr);
258 addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
259 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
260
261 if (bus_claim)
262 sdio_claim_host(rtwsdio->sdio_func);
263
264 if (direct)
265 val = sdio_readb(rtwsdio->sdio_func, addr, &ret);
266 else
267 val = rtw_sdio_indirect_read8(rtwdev, addr, &ret);
268
269 if (bus_claim)
270 sdio_release_host(rtwsdio->sdio_func);
271
272 if (ret)
273 rtw_warn(rtwdev, "sdio read8 failed (0x%x): %d", addr, ret);
274
275 return val;
276 }
277
rtw_sdio_read16(struct rtw_dev * rtwdev,u32 addr)278 static u16 rtw_sdio_read16(struct rtw_dev *rtwdev, u32 addr)
279 {
280 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
281 bool direct, bus_claim;
282 int ret;
283 u16 val;
284
285 direct = rtw_sdio_use_direct_io(rtwdev, addr);
286 addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
287 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
288
289 if (bus_claim)
290 sdio_claim_host(rtwsdio->sdio_func);
291
292 if (direct)
293 val = rtw_sdio_readw(rtwdev, addr, &ret);
294 else
295 val = rtw_sdio_indirect_read16(rtwdev, addr, &ret);
296
297 if (bus_claim)
298 sdio_release_host(rtwsdio->sdio_func);
299
300 if (ret)
301 rtw_warn(rtwdev, "sdio read16 failed (0x%x): %d", addr, ret);
302
303 return val;
304 }
305
rtw_sdio_read32(struct rtw_dev * rtwdev,u32 addr)306 static u32 rtw_sdio_read32(struct rtw_dev *rtwdev, u32 addr)
307 {
308 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
309 bool direct, bus_claim;
310 u32 val;
311 int ret;
312
313 direct = rtw_sdio_use_direct_io(rtwdev, addr);
314 addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
315 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
316
317 if (bus_claim)
318 sdio_claim_host(rtwsdio->sdio_func);
319
320 if (direct)
321 val = rtw_sdio_readl(rtwdev, addr, &ret);
322 else
323 val = rtw_sdio_indirect_read32(rtwdev, addr, &ret);
324
325 if (bus_claim)
326 sdio_release_host(rtwsdio->sdio_func);
327
328 if (ret)
329 rtw_warn(rtwdev, "sdio read32 failed (0x%x): %d", addr, ret);
330
331 return val;
332 }
333
rtw_sdio_indirect_write8(struct rtw_dev * rtwdev,u8 val,u32 addr,int * err_ret)334 static void rtw_sdio_indirect_write8(struct rtw_dev *rtwdev, u8 val, u32 addr,
335 int *err_ret)
336 {
337 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
338 u32 reg_data;
339
340 reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
341 sdio_writeb(rtwsdio->sdio_func, val, reg_data, err_ret);
342 if (*err_ret)
343 return;
344
345 *err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
346 BIT_SDIO_INDIRECT_REG_CFG_WRITE);
347 }
348
rtw_sdio_indirect_write16(struct rtw_dev * rtwdev,u16 val,u32 addr,int * err_ret)349 static void rtw_sdio_indirect_write16(struct rtw_dev *rtwdev, u16 val, u32 addr,
350 int *err_ret)
351 {
352 u32 reg_data;
353
354 if (!IS_ALIGNED(addr, 2)) {
355 addr = rtw_sdio_to_io_address(rtwdev, addr, true);
356 rtw_sdio_writew(rtwdev, val, addr, err_ret);
357 return;
358 }
359
360 reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
361 rtw_sdio_writew(rtwdev, val, reg_data, err_ret);
362 if (*err_ret)
363 return;
364
365 *err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
366 BIT_SDIO_INDIRECT_REG_CFG_WRITE |
367 BIT_SDIO_INDIRECT_REG_CFG_WORD);
368 }
369
rtw_sdio_indirect_write32(struct rtw_dev * rtwdev,u32 val,u32 addr,int * err_ret)370 static void rtw_sdio_indirect_write32(struct rtw_dev *rtwdev, u32 val,
371 u32 addr, int *err_ret)
372 {
373 u32 reg_data;
374
375 if (!IS_ALIGNED(addr, 4)) {
376 addr = rtw_sdio_to_io_address(rtwdev, addr, true);
377 rtw_sdio_writel(rtwdev, val, addr, err_ret);
378 return;
379 }
380
381 reg_data = rtw_sdio_to_bus_offset(rtwdev, REG_SDIO_INDIRECT_REG_DATA);
382 rtw_sdio_writel(rtwdev, val, reg_data, err_ret);
383
384 *err_ret = rtw_sdio_indirect_reg_cfg(rtwdev, addr,
385 BIT_SDIO_INDIRECT_REG_CFG_WRITE |
386 BIT_SDIO_INDIRECT_REG_CFG_DWORD);
387 }
388
rtw_sdio_write8(struct rtw_dev * rtwdev,u32 addr,u8 val)389 static void rtw_sdio_write8(struct rtw_dev *rtwdev, u32 addr, u8 val)
390 {
391 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
392 bool direct, bus_claim;
393 int ret;
394
395 direct = rtw_sdio_use_direct_io(rtwdev, addr);
396 addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
397 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
398
399 if (bus_claim)
400 sdio_claim_host(rtwsdio->sdio_func);
401
402 if (direct)
403 sdio_writeb(rtwsdio->sdio_func, val, addr, &ret);
404 else
405 rtw_sdio_indirect_write8(rtwdev, val, addr, &ret);
406
407 if (bus_claim)
408 sdio_release_host(rtwsdio->sdio_func);
409
410 if (ret)
411 rtw_warn(rtwdev, "sdio write8 failed (0x%x): %d", addr, ret);
412 }
413
rtw_sdio_write16(struct rtw_dev * rtwdev,u32 addr,u16 val)414 static void rtw_sdio_write16(struct rtw_dev *rtwdev, u32 addr, u16 val)
415 {
416 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
417 bool direct, bus_claim;
418 int ret;
419
420 direct = rtw_sdio_use_direct_io(rtwdev, addr);
421 addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
422 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
423
424 if (bus_claim)
425 sdio_claim_host(rtwsdio->sdio_func);
426
427 if (direct)
428 rtw_sdio_writew(rtwdev, val, addr, &ret);
429 else
430 rtw_sdio_indirect_write16(rtwdev, val, addr, &ret);
431
432 if (bus_claim)
433 sdio_release_host(rtwsdio->sdio_func);
434
435 if (ret)
436 rtw_warn(rtwdev, "sdio write16 failed (0x%x): %d", addr, ret);
437 }
438
rtw_sdio_write32(struct rtw_dev * rtwdev,u32 addr,u32 val)439 static void rtw_sdio_write32(struct rtw_dev *rtwdev, u32 addr, u32 val)
440 {
441 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
442 bool direct, bus_claim;
443 int ret;
444
445 direct = rtw_sdio_use_direct_io(rtwdev, addr);
446 addr = rtw_sdio_to_io_address(rtwdev, addr, direct);
447 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
448
449 if (bus_claim)
450 sdio_claim_host(rtwsdio->sdio_func);
451
452 if (direct)
453 rtw_sdio_writel(rtwdev, val, addr, &ret);
454 else
455 rtw_sdio_indirect_write32(rtwdev, val, addr, &ret);
456
457 if (bus_claim)
458 sdio_release_host(rtwsdio->sdio_func);
459
460 if (ret)
461 rtw_warn(rtwdev, "sdio write32 failed (0x%x): %d", addr, ret);
462 }
463
rtw_sdio_get_tx_addr(struct rtw_dev * rtwdev,size_t size,enum rtw_tx_queue_type queue)464 static u32 rtw_sdio_get_tx_addr(struct rtw_dev *rtwdev, size_t size,
465 enum rtw_tx_queue_type queue)
466 {
467 u32 txaddr;
468
469 switch (queue) {
470 case RTW_TX_QUEUE_BCN:
471 case RTW_TX_QUEUE_H2C:
472 case RTW_TX_QUEUE_HI0:
473 txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
474 REG_SDIO_CMD_ADDR_TXFF_HIGH);
475 break;
476 case RTW_TX_QUEUE_VI:
477 case RTW_TX_QUEUE_VO:
478 txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
479 REG_SDIO_CMD_ADDR_TXFF_NORMAL);
480 break;
481 case RTW_TX_QUEUE_BE:
482 case RTW_TX_QUEUE_BK:
483 txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
484 REG_SDIO_CMD_ADDR_TXFF_LOW);
485 break;
486 case RTW_TX_QUEUE_MGMT:
487 txaddr = FIELD_PREP(REG_SDIO_CMD_ADDR_MSK,
488 REG_SDIO_CMD_ADDR_TXFF_EXTRA);
489 break;
490 default:
491 rtw_warn(rtwdev, "Unsupported queue for TX addr: 0x%02x\n",
492 queue);
493 return 0;
494 }
495
496 txaddr += DIV_ROUND_UP(size, 4);
497
498 return txaddr;
499 };
500
rtw_sdio_read_port(struct rtw_dev * rtwdev,u8 * buf,size_t count)501 static int rtw_sdio_read_port(struct rtw_dev *rtwdev, u8 *buf, size_t count)
502 {
503 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
504 struct mmc_host *host = rtwsdio->sdio_func->card->host;
505 bool bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
506 u32 rxaddr = rtwsdio->rx_addr++;
507 int ret = 0, err;
508 size_t bytes;
509
510 if (bus_claim)
511 sdio_claim_host(rtwsdio->sdio_func);
512
513 while (count > 0) {
514 bytes = min_t(size_t, host->max_req_size, count);
515
516 err = sdio_memcpy_fromio(rtwsdio->sdio_func, buf,
517 RTW_SDIO_ADDR_RX_RX0FF_GEN(rxaddr),
518 bytes);
519 if (err) {
520 rtw_warn(rtwdev,
521 "Failed to read %zu byte(s) from SDIO port 0x%08x: %d",
522 bytes, rxaddr, err);
523
524 /* Signal to the caller that reading did not work and
525 * that the data in the buffer is short/corrupted.
526 */
527 ret = err;
528
529 /* Don't stop here - instead drain the remaining data
530 * from the card's buffer, else the card will return
531 * corrupt data for the next rtw_sdio_read_port() call.
532 */
533 }
534
535 count -= bytes;
536 buf += bytes;
537 }
538
539 if (bus_claim)
540 sdio_release_host(rtwsdio->sdio_func);
541
542 return ret;
543 }
544
rtw_sdio_check_free_txpg(struct rtw_dev * rtwdev,u8 queue,size_t count)545 static int rtw_sdio_check_free_txpg(struct rtw_dev *rtwdev, u8 queue,
546 size_t count)
547 {
548 unsigned int pages_free, pages_needed;
549
550 if (rtw_chip_wcpu_11n(rtwdev)) {
551 u32 free_txpg;
552
553 free_txpg = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG);
554
555 switch (queue) {
556 case RTW_TX_QUEUE_BCN:
557 case RTW_TX_QUEUE_H2C:
558 case RTW_TX_QUEUE_HI0:
559 case RTW_TX_QUEUE_MGMT:
560 /* high */
561 pages_free = free_txpg & 0xff;
562 break;
563 case RTW_TX_QUEUE_VI:
564 case RTW_TX_QUEUE_VO:
565 /* normal */
566 pages_free = (free_txpg >> 8) & 0xff;
567 break;
568 case RTW_TX_QUEUE_BE:
569 case RTW_TX_QUEUE_BK:
570 /* low */
571 pages_free = (free_txpg >> 16) & 0xff;
572 break;
573 default:
574 rtw_warn(rtwdev, "Unknown mapping for queue %u\n", queue);
575 return -EINVAL;
576 }
577
578 /* add the pages from the public queue */
579 pages_free += (free_txpg >> 24) & 0xff;
580 } else {
581 u32 free_txpg[3];
582
583 free_txpg[0] = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG);
584 free_txpg[1] = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG + 4);
585 free_txpg[2] = rtw_sdio_read32(rtwdev, REG_SDIO_FREE_TXPG + 8);
586
587 switch (queue) {
588 case RTW_TX_QUEUE_BCN:
589 case RTW_TX_QUEUE_H2C:
590 case RTW_TX_QUEUE_HI0:
591 /* high */
592 pages_free = free_txpg[0] & 0xfff;
593 break;
594 case RTW_TX_QUEUE_VI:
595 case RTW_TX_QUEUE_VO:
596 /* normal */
597 pages_free = (free_txpg[0] >> 16) & 0xfff;
598 break;
599 case RTW_TX_QUEUE_BE:
600 case RTW_TX_QUEUE_BK:
601 /* low */
602 pages_free = free_txpg[1] & 0xfff;
603 break;
604 case RTW_TX_QUEUE_MGMT:
605 /* extra */
606 pages_free = free_txpg[2] & 0xfff;
607 break;
608 default:
609 rtw_warn(rtwdev, "Unknown mapping for queue %u\n", queue);
610 return -EINVAL;
611 }
612
613 /* add the pages from the public queue */
614 pages_free += (free_txpg[1] >> 16) & 0xfff;
615 }
616
617 pages_needed = DIV_ROUND_UP(count, rtwdev->chip->page_size);
618
619 if (pages_needed > pages_free) {
620 rtw_dbg(rtwdev, RTW_DBG_SDIO,
621 "Not enough free pages (%u needed, %u free) in queue %u for %zu bytes\n",
622 pages_needed, pages_free, queue, count);
623 return -EBUSY;
624 }
625
626 return 0;
627 }
628
rtw_sdio_write_port(struct rtw_dev * rtwdev,struct sk_buff * skb,enum rtw_tx_queue_type queue)629 static int rtw_sdio_write_port(struct rtw_dev *rtwdev, struct sk_buff *skb,
630 enum rtw_tx_queue_type queue)
631 {
632 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
633 bool bus_claim;
634 size_t txsize;
635 u32 txaddr;
636 int ret;
637
638 txaddr = rtw_sdio_get_tx_addr(rtwdev, skb->len, queue);
639 if (!txaddr)
640 return -EINVAL;
641
642 txsize = sdio_align_size(rtwsdio->sdio_func, skb->len);
643
644 ret = rtw_sdio_check_free_txpg(rtwdev, queue, txsize);
645 if (ret)
646 return ret;
647
648 if (!IS_ALIGNED((unsigned long)skb->data, RTW_SDIO_DATA_PTR_ALIGN))
649 rtw_warn(rtwdev, "Got unaligned SKB in %s() for queue %u\n",
650 __func__, queue);
651
652 bus_claim = rtw_sdio_bus_claim_needed(rtwsdio);
653
654 if (bus_claim)
655 sdio_claim_host(rtwsdio->sdio_func);
656
657 ret = sdio_memcpy_toio(rtwsdio->sdio_func, txaddr, skb->data, txsize);
658
659 if (bus_claim)
660 sdio_release_host(rtwsdio->sdio_func);
661
662 if (ret)
663 rtw_warn(rtwdev,
664 "Failed to write %zu byte(s) to SDIO port 0x%08x",
665 txsize, txaddr);
666
667 return ret;
668 }
669
rtw_sdio_init(struct rtw_dev * rtwdev)670 static void rtw_sdio_init(struct rtw_dev *rtwdev)
671 {
672 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
673
674 rtwsdio->irq_mask = REG_SDIO_HIMR_RX_REQUEST | REG_SDIO_HIMR_CPWM1;
675 }
676
rtw_sdio_enable_rx_aggregation(struct rtw_dev * rtwdev)677 static void rtw_sdio_enable_rx_aggregation(struct rtw_dev *rtwdev)
678 {
679 u8 size, timeout;
680
681 if (rtw_chip_wcpu_11n(rtwdev)) {
682 size = 0x6;
683 timeout = 0x6;
684 } else {
685 size = 0xff;
686 timeout = 0x1;
687 }
688
689 /* Make the firmware honor the size limit configured below */
690 rtw_write32_set(rtwdev, REG_RXDMA_AGG_PG_TH, BIT_EN_PRE_CALC);
691
692 rtw_write8_set(rtwdev, REG_TXDMA_PQ_MAP, BIT_RXDMA_AGG_EN);
693
694 rtw_write16(rtwdev, REG_RXDMA_AGG_PG_TH,
695 FIELD_PREP(BIT_RXDMA_AGG_PG_TH, size) |
696 FIELD_PREP(BIT_DMA_AGG_TO_V1, timeout));
697
698 rtw_write8_set(rtwdev, REG_RXDMA_MODE, BIT_DMA_MODE);
699 }
700
rtw_sdio_enable_interrupt(struct rtw_dev * rtwdev)701 static void rtw_sdio_enable_interrupt(struct rtw_dev *rtwdev)
702 {
703 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
704
705 rtw_write32(rtwdev, REG_SDIO_HIMR, rtwsdio->irq_mask);
706 }
707
rtw_sdio_disable_interrupt(struct rtw_dev * rtwdev)708 static void rtw_sdio_disable_interrupt(struct rtw_dev *rtwdev)
709 {
710 rtw_write32(rtwdev, REG_SDIO_HIMR, 0x0);
711 }
712
rtw_sdio_get_tx_qsel(struct rtw_dev * rtwdev,struct sk_buff * skb,u8 queue)713 static u8 rtw_sdio_get_tx_qsel(struct rtw_dev *rtwdev, struct sk_buff *skb,
714 u8 queue)
715 {
716 switch (queue) {
717 case RTW_TX_QUEUE_BCN:
718 return TX_DESC_QSEL_BEACON;
719 case RTW_TX_QUEUE_H2C:
720 return TX_DESC_QSEL_H2C;
721 case RTW_TX_QUEUE_MGMT:
722 return TX_DESC_QSEL_MGMT;
723 case RTW_TX_QUEUE_HI0:
724 return TX_DESC_QSEL_HIGH;
725 default:
726 return skb->priority;
727 }
728 }
729
rtw_sdio_setup(struct rtw_dev * rtwdev)730 static int rtw_sdio_setup(struct rtw_dev *rtwdev)
731 {
732 /* nothing to do */
733 return 0;
734 }
735
rtw_sdio_start(struct rtw_dev * rtwdev)736 static int rtw_sdio_start(struct rtw_dev *rtwdev)
737 {
738 rtw_sdio_enable_rx_aggregation(rtwdev);
739 rtw_sdio_enable_interrupt(rtwdev);
740
741 return 0;
742 }
743
rtw_sdio_stop(struct rtw_dev * rtwdev)744 static void rtw_sdio_stop(struct rtw_dev *rtwdev)
745 {
746 rtw_sdio_disable_interrupt(rtwdev);
747 }
748
rtw_sdio_deep_ps_enter(struct rtw_dev * rtwdev)749 static void rtw_sdio_deep_ps_enter(struct rtw_dev *rtwdev)
750 {
751 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
752 bool tx_empty = true;
753 u8 queue;
754
755 if (!rtw_fw_feature_check(&rtwdev->fw, FW_FEATURE_TX_WAKE)) {
756 /* Deep PS state is not allowed to TX-DMA */
757 for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) {
758 /* BCN queue is rsvd page, does not have DMA interrupt
759 * H2C queue is managed by firmware
760 */
761 if (queue == RTW_TX_QUEUE_BCN ||
762 queue == RTW_TX_QUEUE_H2C)
763 continue;
764
765 /* check if there is any skb DMAing */
766 if (skb_queue_len(&rtwsdio->tx_queue[queue])) {
767 tx_empty = false;
768 break;
769 }
770 }
771 }
772
773 if (!tx_empty) {
774 rtw_dbg(rtwdev, RTW_DBG_PS,
775 "TX path not empty, cannot enter deep power save state\n");
776 return;
777 }
778
779 set_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags);
780 rtw_power_mode_change(rtwdev, true);
781 }
782
rtw_sdio_deep_ps_leave(struct rtw_dev * rtwdev)783 static void rtw_sdio_deep_ps_leave(struct rtw_dev *rtwdev)
784 {
785 if (test_and_clear_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags))
786 rtw_power_mode_change(rtwdev, false);
787 }
788
rtw_sdio_deep_ps(struct rtw_dev * rtwdev,bool enter)789 static void rtw_sdio_deep_ps(struct rtw_dev *rtwdev, bool enter)
790 {
791 if (enter && !test_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags))
792 rtw_sdio_deep_ps_enter(rtwdev);
793
794 if (!enter && test_bit(RTW_FLAG_LEISURE_PS_DEEP, rtwdev->flags))
795 rtw_sdio_deep_ps_leave(rtwdev);
796 }
797
rtw_sdio_tx_kick_off(struct rtw_dev * rtwdev)798 static void rtw_sdio_tx_kick_off(struct rtw_dev *rtwdev)
799 {
800 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
801
802 queue_work(rtwsdio->txwq, &rtwsdio->tx_handler_data->work);
803 }
804
rtw_sdio_link_ps(struct rtw_dev * rtwdev,bool enter)805 static void rtw_sdio_link_ps(struct rtw_dev *rtwdev, bool enter)
806 {
807 /* nothing to do */
808 }
809
rtw_sdio_interface_cfg(struct rtw_dev * rtwdev)810 static void rtw_sdio_interface_cfg(struct rtw_dev *rtwdev)
811 {
812 u32 val;
813
814 rtw_read32(rtwdev, REG_SDIO_FREE_TXPG);
815
816 val = rtw_read32(rtwdev, REG_SDIO_TX_CTRL);
817 val &= 0xfff8;
818 rtw_write32(rtwdev, REG_SDIO_TX_CTRL, val);
819 }
820
rtw_sdio_get_tx_data(struct sk_buff * skb)821 static struct rtw_sdio_tx_data *rtw_sdio_get_tx_data(struct sk_buff *skb)
822 {
823 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
824
825 BUILD_BUG_ON(sizeof(struct rtw_sdio_tx_data) >
826 sizeof(info->status.status_driver_data));
827
828 return (struct rtw_sdio_tx_data *)info->status.status_driver_data;
829 }
830
rtw_sdio_tx_skb_prepare(struct rtw_dev * rtwdev,struct rtw_tx_pkt_info * pkt_info,struct sk_buff * skb,enum rtw_tx_queue_type queue)831 static void rtw_sdio_tx_skb_prepare(struct rtw_dev *rtwdev,
832 struct rtw_tx_pkt_info *pkt_info,
833 struct sk_buff *skb,
834 enum rtw_tx_queue_type queue)
835 {
836 const struct rtw_chip_info *chip = rtwdev->chip;
837 unsigned long data_addr, aligned_addr;
838 size_t offset;
839 u8 *pkt_desc;
840
841 pkt_desc = skb_push(skb, chip->tx_pkt_desc_sz);
842
843 data_addr = (unsigned long)pkt_desc;
844 aligned_addr = ALIGN(data_addr, RTW_SDIO_DATA_PTR_ALIGN);
845
846 if (data_addr != aligned_addr) {
847 /* Ensure that the start of the pkt_desc is always aligned at
848 * RTW_SDIO_DATA_PTR_ALIGN.
849 */
850 offset = RTW_SDIO_DATA_PTR_ALIGN - (aligned_addr - data_addr);
851
852 pkt_desc = skb_push(skb, offset);
853
854 /* By inserting padding to align the start of the pkt_desc we
855 * need to inform the firmware that the actual data starts at
856 * a different offset than normal.
857 */
858 pkt_info->offset += offset;
859 }
860
861 memset(pkt_desc, 0, chip->tx_pkt_desc_sz);
862
863 pkt_info->qsel = rtw_sdio_get_tx_qsel(rtwdev, skb, queue);
864
865 rtw_tx_fill_tx_desc(pkt_info, skb);
866 rtw_tx_fill_txdesc_checksum(rtwdev, pkt_info, pkt_desc);
867 }
868
rtw_sdio_write_data(struct rtw_dev * rtwdev,struct rtw_tx_pkt_info * pkt_info,struct sk_buff * skb,enum rtw_tx_queue_type queue)869 static int rtw_sdio_write_data(struct rtw_dev *rtwdev,
870 struct rtw_tx_pkt_info *pkt_info,
871 struct sk_buff *skb,
872 enum rtw_tx_queue_type queue)
873 {
874 int ret;
875
876 rtw_sdio_tx_skb_prepare(rtwdev, pkt_info, skb, queue);
877
878 ret = rtw_sdio_write_port(rtwdev, skb, queue);
879 dev_kfree_skb_any(skb);
880
881 return ret;
882 }
883
rtw_sdio_write_data_rsvd_page(struct rtw_dev * rtwdev,u8 * buf,u32 size)884 static int rtw_sdio_write_data_rsvd_page(struct rtw_dev *rtwdev, u8 *buf,
885 u32 size)
886 {
887 struct rtw_tx_pkt_info pkt_info = {};
888 struct sk_buff *skb;
889
890 skb = rtw_tx_write_data_rsvd_page_get(rtwdev, &pkt_info, buf, size);
891 if (!skb)
892 return -ENOMEM;
893
894 return rtw_sdio_write_data(rtwdev, &pkt_info, skb, RTW_TX_QUEUE_BCN);
895 }
896
rtw_sdio_write_data_h2c(struct rtw_dev * rtwdev,u8 * buf,u32 size)897 static int rtw_sdio_write_data_h2c(struct rtw_dev *rtwdev, u8 *buf, u32 size)
898 {
899 struct rtw_tx_pkt_info pkt_info = {};
900 struct sk_buff *skb;
901
902 skb = rtw_tx_write_data_h2c_get(rtwdev, &pkt_info, buf, size);
903 if (!skb)
904 return -ENOMEM;
905
906 return rtw_sdio_write_data(rtwdev, &pkt_info, skb, RTW_TX_QUEUE_H2C);
907 }
908
rtw_sdio_tx_write(struct rtw_dev * rtwdev,struct rtw_tx_pkt_info * pkt_info,struct sk_buff * skb)909 static int rtw_sdio_tx_write(struct rtw_dev *rtwdev,
910 struct rtw_tx_pkt_info *pkt_info,
911 struct sk_buff *skb)
912 {
913 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
914 enum rtw_tx_queue_type queue = rtw_tx_queue_mapping(skb);
915 struct rtw_sdio_tx_data *tx_data;
916
917 rtw_sdio_tx_skb_prepare(rtwdev, pkt_info, skb, queue);
918
919 tx_data = rtw_sdio_get_tx_data(skb);
920 tx_data->sn = pkt_info->sn;
921
922 skb_queue_tail(&rtwsdio->tx_queue[queue], skb);
923
924 return 0;
925 }
926
rtw_sdio_tx_err_isr(struct rtw_dev * rtwdev)927 static void rtw_sdio_tx_err_isr(struct rtw_dev *rtwdev)
928 {
929 u32 val = rtw_read32(rtwdev, REG_TXDMA_STATUS);
930
931 rtw_write32(rtwdev, REG_TXDMA_STATUS, val);
932 }
933
rtw_sdio_rx_skb(struct rtw_dev * rtwdev,struct sk_buff * skb,u32 pkt_offset,struct rtw_rx_pkt_stat * pkt_stat,struct ieee80211_rx_status * rx_status)934 static void rtw_sdio_rx_skb(struct rtw_dev *rtwdev, struct sk_buff *skb,
935 u32 pkt_offset, struct rtw_rx_pkt_stat *pkt_stat,
936 struct ieee80211_rx_status *rx_status)
937 {
938 *IEEE80211_SKB_RXCB(skb) = *rx_status;
939
940 if (pkt_stat->is_c2h) {
941 skb_put(skb, pkt_stat->pkt_len + pkt_offset);
942 rtw_fw_c2h_cmd_rx_irqsafe(rtwdev, pkt_offset, skb);
943 return;
944 }
945
946 skb_put(skb, pkt_stat->pkt_len);
947 skb_reserve(skb, pkt_offset);
948
949 rtw_update_rx_freq_for_invalid(rtwdev, skb, rx_status, pkt_stat);
950 rtw_rx_stats(rtwdev, pkt_stat->vif, skb);
951
952 ieee80211_rx_irqsafe(rtwdev->hw, skb);
953 }
954
rtw_sdio_rxfifo_recv(struct rtw_dev * rtwdev,u32 rx_len)955 static void rtw_sdio_rxfifo_recv(struct rtw_dev *rtwdev, u32 rx_len)
956 {
957 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
958 const struct rtw_chip_info *chip = rtwdev->chip;
959 u32 pkt_desc_sz = chip->rx_pkt_desc_sz;
960 struct ieee80211_rx_status rx_status;
961 struct rtw_rx_pkt_stat pkt_stat;
962 struct sk_buff *skb, *split_skb;
963 u32 pkt_offset, curr_pkt_len;
964 size_t bufsz;
965 u8 *rx_desc;
966 int ret;
967
968 bufsz = sdio_align_size(rtwsdio->sdio_func, rx_len);
969
970 skb = dev_alloc_skb(bufsz);
971 if (!skb)
972 return;
973
974 ret = rtw_sdio_read_port(rtwdev, skb->data, bufsz);
975 if (ret) {
976 dev_kfree_skb_any(skb);
977 return;
978 }
979
980 while (true) {
981 rx_desc = skb->data;
982 chip->ops->query_rx_desc(rtwdev, rx_desc, &pkt_stat,
983 &rx_status);
984 pkt_offset = pkt_desc_sz + pkt_stat.drv_info_sz +
985 pkt_stat.shift;
986
987 curr_pkt_len = ALIGN(pkt_offset + pkt_stat.pkt_len,
988 RTW_SDIO_DATA_PTR_ALIGN);
989
990 if ((curr_pkt_len + pkt_desc_sz) >= rx_len) {
991 /* Use the original skb (with it's adjusted offset)
992 * when processing the last (or even the only) entry to
993 * have it's memory freed automatically.
994 */
995 rtw_sdio_rx_skb(rtwdev, skb, pkt_offset, &pkt_stat,
996 &rx_status);
997 break;
998 }
999
1000 split_skb = dev_alloc_skb(curr_pkt_len);
1001 if (!split_skb) {
1002 rtw_sdio_rx_skb(rtwdev, skb, pkt_offset, &pkt_stat,
1003 &rx_status);
1004 break;
1005 }
1006
1007 skb_copy_header(split_skb, skb);
1008 memcpy(split_skb->data, skb->data, curr_pkt_len);
1009
1010 rtw_sdio_rx_skb(rtwdev, split_skb, pkt_offset, &pkt_stat,
1011 &rx_status);
1012
1013 /* Move to the start of the next RX descriptor */
1014 skb_reserve(skb, curr_pkt_len);
1015 rx_len -= curr_pkt_len;
1016 }
1017 }
1018
rtw_sdio_rx_isr(struct rtw_dev * rtwdev)1019 static void rtw_sdio_rx_isr(struct rtw_dev *rtwdev)
1020 {
1021 u32 rx_len, hisr, total_rx_bytes = 0;
1022
1023 do {
1024 if (rtw_chip_wcpu_11n(rtwdev))
1025 rx_len = rtw_read16(rtwdev, REG_SDIO_RX0_REQ_LEN);
1026 else
1027 rx_len = rtw_read32(rtwdev, REG_SDIO_RX0_REQ_LEN);
1028
1029 if (!rx_len)
1030 break;
1031
1032 rtw_sdio_rxfifo_recv(rtwdev, rx_len);
1033
1034 total_rx_bytes += rx_len;
1035
1036 if (rtw_chip_wcpu_11n(rtwdev)) {
1037 /* Stop if no more RX requests are pending, even if
1038 * rx_len could be greater than zero in the next
1039 * iteration. This is needed because the RX buffer may
1040 * already contain data while either HW or FW are not
1041 * done filling that buffer yet. Still reading the
1042 * buffer can result in packets where
1043 * rtw_rx_pkt_stat.pkt_len is zero or points beyond the
1044 * end of the buffer.
1045 */
1046 hisr = rtw_read32(rtwdev, REG_SDIO_HISR);
1047 } else {
1048 /* RTW_WCPU_11AC chips have improved hardware or
1049 * firmware and can use rx_len unconditionally.
1050 */
1051 hisr = REG_SDIO_HISR_RX_REQUEST;
1052 }
1053 } while (total_rx_bytes < SZ_64K && hisr & REG_SDIO_HISR_RX_REQUEST);
1054 }
1055
rtw_sdio_handle_interrupt(struct sdio_func * sdio_func)1056 static void rtw_sdio_handle_interrupt(struct sdio_func *sdio_func)
1057 {
1058 struct ieee80211_hw *hw = sdio_get_drvdata(sdio_func);
1059 struct rtw_sdio *rtwsdio;
1060 struct rtw_dev *rtwdev;
1061 u32 hisr;
1062
1063 rtwdev = hw->priv;
1064 rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1065
1066 rtwsdio->irq_thread = current;
1067
1068 hisr = rtw_read32(rtwdev, REG_SDIO_HISR);
1069
1070 if (hisr & REG_SDIO_HISR_TXERR)
1071 rtw_sdio_tx_err_isr(rtwdev);
1072 if (hisr & REG_SDIO_HISR_RX_REQUEST) {
1073 hisr &= ~REG_SDIO_HISR_RX_REQUEST;
1074 rtw_sdio_rx_isr(rtwdev);
1075 }
1076
1077 rtw_write32(rtwdev, REG_SDIO_HISR, hisr);
1078
1079 rtwsdio->irq_thread = NULL;
1080 }
1081
rtw_sdio_suspend(struct device * dev)1082 static int __maybe_unused rtw_sdio_suspend(struct device *dev)
1083 {
1084 struct sdio_func *func = dev_to_sdio_func(dev);
1085 struct ieee80211_hw *hw = dev_get_drvdata(dev);
1086 struct rtw_dev *rtwdev = hw->priv;
1087 int ret;
1088
1089 ret = sdio_set_host_pm_flags(func, MMC_PM_KEEP_POWER);
1090 if (ret)
1091 rtw_err(rtwdev, "Failed to host PM flag MMC_PM_KEEP_POWER");
1092
1093 return ret;
1094 }
1095
rtw_sdio_resume(struct device * dev)1096 static int __maybe_unused rtw_sdio_resume(struct device *dev)
1097 {
1098 return 0;
1099 }
1100
1101 SIMPLE_DEV_PM_OPS(rtw_sdio_pm_ops, rtw_sdio_suspend, rtw_sdio_resume);
1102 EXPORT_SYMBOL(rtw_sdio_pm_ops);
1103
rtw_sdio_claim(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1104 static int rtw_sdio_claim(struct rtw_dev *rtwdev, struct sdio_func *sdio_func)
1105 {
1106 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1107 int ret;
1108
1109 sdio_claim_host(sdio_func);
1110
1111 ret = sdio_enable_func(sdio_func);
1112 if (ret) {
1113 rtw_err(rtwdev, "Failed to enable SDIO func");
1114 goto err_release_host;
1115 }
1116
1117 ret = sdio_set_block_size(sdio_func, RTW_SDIO_BLOCK_SIZE);
1118 if (ret) {
1119 rtw_err(rtwdev, "Failed to set SDIO block size to 512");
1120 goto err_disable_func;
1121 }
1122
1123 rtwsdio->sdio_func = sdio_func;
1124
1125 rtwsdio->sdio3_bus_mode = mmc_card_uhs(sdio_func->card);
1126
1127 sdio_set_drvdata(sdio_func, rtwdev->hw);
1128 SET_IEEE80211_DEV(rtwdev->hw, &sdio_func->dev);
1129
1130 sdio_release_host(sdio_func);
1131
1132 return 0;
1133
1134 err_disable_func:
1135 sdio_disable_func(sdio_func);
1136 err_release_host:
1137 sdio_release_host(sdio_func);
1138 return ret;
1139 }
1140
rtw_sdio_declaim(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1141 static void rtw_sdio_declaim(struct rtw_dev *rtwdev,
1142 struct sdio_func *sdio_func)
1143 {
1144 sdio_claim_host(sdio_func);
1145 sdio_disable_func(sdio_func);
1146 sdio_release_host(sdio_func);
1147 }
1148
1149 static struct rtw_hci_ops rtw_sdio_ops = {
1150 .tx_write = rtw_sdio_tx_write,
1151 .tx_kick_off = rtw_sdio_tx_kick_off,
1152 .setup = rtw_sdio_setup,
1153 .start = rtw_sdio_start,
1154 .stop = rtw_sdio_stop,
1155 .deep_ps = rtw_sdio_deep_ps,
1156 .link_ps = rtw_sdio_link_ps,
1157 .interface_cfg = rtw_sdio_interface_cfg,
1158 .dynamic_rx_agg = NULL,
1159 .write_firmware_page = rtw_write_firmware_page,
1160
1161 .read8 = rtw_sdio_read8,
1162 .read16 = rtw_sdio_read16,
1163 .read32 = rtw_sdio_read32,
1164 .write8 = rtw_sdio_write8,
1165 .write16 = rtw_sdio_write16,
1166 .write32 = rtw_sdio_write32,
1167 .write_data_rsvd_page = rtw_sdio_write_data_rsvd_page,
1168 .write_data_h2c = rtw_sdio_write_data_h2c,
1169 };
1170
rtw_sdio_request_irq(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1171 static int rtw_sdio_request_irq(struct rtw_dev *rtwdev,
1172 struct sdio_func *sdio_func)
1173 {
1174 int ret;
1175
1176 sdio_claim_host(sdio_func);
1177 ret = sdio_claim_irq(sdio_func, &rtw_sdio_handle_interrupt);
1178 sdio_release_host(sdio_func);
1179
1180 if (ret) {
1181 rtw_err(rtwdev, "failed to claim SDIO IRQ");
1182 return ret;
1183 }
1184
1185 return 0;
1186 }
1187
rtw_sdio_indicate_tx_status(struct rtw_dev * rtwdev,struct sk_buff * skb)1188 static void rtw_sdio_indicate_tx_status(struct rtw_dev *rtwdev,
1189 struct sk_buff *skb)
1190 {
1191 struct rtw_sdio_tx_data *tx_data = rtw_sdio_get_tx_data(skb);
1192 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
1193 struct ieee80211_hw *hw = rtwdev->hw;
1194
1195 skb_pull(skb, rtwdev->chip->tx_pkt_desc_sz);
1196
1197 /* enqueue to wait for tx report */
1198 if (info->flags & IEEE80211_TX_CTL_REQ_TX_STATUS) {
1199 rtw_tx_report_enqueue(rtwdev, skb, tx_data->sn);
1200 return;
1201 }
1202
1203 /* always ACK for others, then they won't be marked as drop */
1204 ieee80211_tx_info_clear_status(info);
1205 if (info->flags & IEEE80211_TX_CTL_NO_ACK)
1206 info->flags |= IEEE80211_TX_STAT_NOACK_TRANSMITTED;
1207 else
1208 info->flags |= IEEE80211_TX_STAT_ACK;
1209
1210 ieee80211_tx_status_irqsafe(hw, skb);
1211 }
1212
rtw_sdio_process_tx_queue(struct rtw_dev * rtwdev,enum rtw_tx_queue_type queue)1213 static void rtw_sdio_process_tx_queue(struct rtw_dev *rtwdev,
1214 enum rtw_tx_queue_type queue)
1215 {
1216 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1217 struct sk_buff *skb;
1218 int ret;
1219
1220 skb = skb_dequeue(&rtwsdio->tx_queue[queue]);
1221 if (!skb)
1222 return;
1223
1224 ret = rtw_sdio_write_port(rtwdev, skb, queue);
1225 if (ret) {
1226 skb_queue_head(&rtwsdio->tx_queue[queue], skb);
1227 return;
1228 }
1229
1230 rtw_sdio_indicate_tx_status(rtwdev, skb);
1231 }
1232
rtw_sdio_tx_handler(struct work_struct * work)1233 static void rtw_sdio_tx_handler(struct work_struct *work)
1234 {
1235 struct rtw_sdio_work_data *work_data =
1236 container_of(work, struct rtw_sdio_work_data, work);
1237 struct rtw_sdio *rtwsdio;
1238 struct rtw_dev *rtwdev;
1239 int limit, queue;
1240
1241 rtwdev = work_data->rtwdev;
1242 rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1243
1244 if (!rtw_fw_feature_check(&rtwdev->fw, FW_FEATURE_TX_WAKE))
1245 rtw_sdio_deep_ps_leave(rtwdev);
1246
1247 for (queue = RTK_MAX_TX_QUEUE_NUM - 1; queue >= 0; queue--) {
1248 for (limit = 0; limit < 1000; limit++) {
1249 rtw_sdio_process_tx_queue(rtwdev, queue);
1250
1251 if (skb_queue_empty(&rtwsdio->tx_queue[queue]))
1252 break;
1253 }
1254 }
1255 }
1256
rtw_sdio_free_irq(struct rtw_dev * rtwdev,struct sdio_func * sdio_func)1257 static void rtw_sdio_free_irq(struct rtw_dev *rtwdev,
1258 struct sdio_func *sdio_func)
1259 {
1260 sdio_claim_host(sdio_func);
1261 sdio_release_irq(sdio_func);
1262 sdio_release_host(sdio_func);
1263 }
1264
rtw_sdio_init_tx(struct rtw_dev * rtwdev)1265 static int rtw_sdio_init_tx(struct rtw_dev *rtwdev)
1266 {
1267 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1268 int i;
1269
1270 rtwsdio->txwq = create_singlethread_workqueue("rtw88_sdio: tx wq");
1271 if (!rtwsdio->txwq) {
1272 rtw_err(rtwdev, "failed to create TX work queue\n");
1273 return -ENOMEM;
1274 }
1275
1276 for (i = 0; i < RTK_MAX_TX_QUEUE_NUM; i++)
1277 skb_queue_head_init(&rtwsdio->tx_queue[i]);
1278 rtwsdio->tx_handler_data = kmalloc(sizeof(*rtwsdio->tx_handler_data),
1279 GFP_KERNEL);
1280 if (!rtwsdio->tx_handler_data)
1281 goto err_destroy_wq;
1282
1283 rtwsdio->tx_handler_data->rtwdev = rtwdev;
1284 INIT_WORK(&rtwsdio->tx_handler_data->work, rtw_sdio_tx_handler);
1285
1286 return 0;
1287
1288 err_destroy_wq:
1289 destroy_workqueue(rtwsdio->txwq);
1290 return -ENOMEM;
1291 }
1292
rtw_sdio_deinit_tx(struct rtw_dev * rtwdev)1293 static void rtw_sdio_deinit_tx(struct rtw_dev *rtwdev)
1294 {
1295 struct rtw_sdio *rtwsdio = (struct rtw_sdio *)rtwdev->priv;
1296 int i;
1297
1298 flush_workqueue(rtwsdio->txwq);
1299 destroy_workqueue(rtwsdio->txwq);
1300 kfree(rtwsdio->tx_handler_data);
1301
1302 for (i = 0; i < RTK_MAX_TX_QUEUE_NUM; i++)
1303 ieee80211_purge_tx_queue(rtwdev->hw, &rtwsdio->tx_queue[i]);
1304 }
1305
rtw_sdio_probe(struct sdio_func * sdio_func,const struct sdio_device_id * id)1306 int rtw_sdio_probe(struct sdio_func *sdio_func,
1307 const struct sdio_device_id *id)
1308 {
1309 struct ieee80211_hw *hw;
1310 struct rtw_dev *rtwdev;
1311 int drv_data_size;
1312 int ret;
1313
1314 drv_data_size = sizeof(struct rtw_dev) + sizeof(struct rtw_sdio);
1315 hw = ieee80211_alloc_hw(drv_data_size, &rtw_ops);
1316 if (!hw) {
1317 dev_err(&sdio_func->dev, "failed to allocate hw");
1318 return -ENOMEM;
1319 }
1320
1321 rtwdev = hw->priv;
1322 rtwdev->hw = hw;
1323 rtwdev->dev = &sdio_func->dev;
1324 rtwdev->chip = (struct rtw_chip_info *)id->driver_data;
1325 rtwdev->hci.ops = &rtw_sdio_ops;
1326 rtwdev->hci.type = RTW_HCI_TYPE_SDIO;
1327
1328 ret = rtw_core_init(rtwdev);
1329 if (ret)
1330 goto err_release_hw;
1331
1332 rtw_dbg(rtwdev, RTW_DBG_SDIO,
1333 "rtw88 SDIO probe: vendor=0x%04x device=%04x class=%02x",
1334 id->vendor, id->device, id->class);
1335
1336 ret = rtw_sdio_claim(rtwdev, sdio_func);
1337 if (ret) {
1338 rtw_err(rtwdev, "failed to claim SDIO device");
1339 goto err_deinit_core;
1340 }
1341
1342 rtw_sdio_init(rtwdev);
1343
1344 ret = rtw_sdio_init_tx(rtwdev);
1345 if (ret) {
1346 rtw_err(rtwdev, "failed to init SDIO TX queue\n");
1347 goto err_sdio_declaim;
1348 }
1349
1350 ret = rtw_chip_info_setup(rtwdev);
1351 if (ret) {
1352 rtw_err(rtwdev, "failed to setup chip information");
1353 goto err_destroy_txwq;
1354 }
1355
1356 ret = rtw_sdio_request_irq(rtwdev, sdio_func);
1357 if (ret)
1358 goto err_destroy_txwq;
1359
1360 ret = rtw_register_hw(rtwdev, hw);
1361 if (ret) {
1362 rtw_err(rtwdev, "failed to register hw");
1363 goto err_free_irq;
1364 }
1365
1366 return 0;
1367
1368 err_free_irq:
1369 rtw_sdio_free_irq(rtwdev, sdio_func);
1370 err_destroy_txwq:
1371 rtw_sdio_deinit_tx(rtwdev);
1372 err_sdio_declaim:
1373 rtw_sdio_declaim(rtwdev, sdio_func);
1374 err_deinit_core:
1375 rtw_core_deinit(rtwdev);
1376 err_release_hw:
1377 ieee80211_free_hw(hw);
1378
1379 return ret;
1380 }
1381 EXPORT_SYMBOL(rtw_sdio_probe);
1382
rtw_sdio_remove(struct sdio_func * sdio_func)1383 void rtw_sdio_remove(struct sdio_func *sdio_func)
1384 {
1385 struct ieee80211_hw *hw = sdio_get_drvdata(sdio_func);
1386 struct rtw_dev *rtwdev;
1387
1388 if (!hw)
1389 return;
1390
1391 rtwdev = hw->priv;
1392
1393 rtw_unregister_hw(rtwdev, hw);
1394 rtw_sdio_disable_interrupt(rtwdev);
1395 rtw_sdio_free_irq(rtwdev, sdio_func);
1396 rtw_sdio_declaim(rtwdev, sdio_func);
1397 rtw_sdio_deinit_tx(rtwdev);
1398 rtw_core_deinit(rtwdev);
1399 ieee80211_free_hw(hw);
1400 }
1401 EXPORT_SYMBOL(rtw_sdio_remove);
1402
rtw_sdio_shutdown(struct device * dev)1403 void rtw_sdio_shutdown(struct device *dev)
1404 {
1405 struct sdio_func *sdio_func = dev_to_sdio_func(dev);
1406 const struct rtw_chip_info *chip;
1407 struct ieee80211_hw *hw;
1408 struct rtw_dev *rtwdev;
1409
1410 hw = sdio_get_drvdata(sdio_func);
1411 if (!hw)
1412 return;
1413
1414 rtwdev = hw->priv;
1415 chip = rtwdev->chip;
1416
1417 if (chip->ops->shutdown)
1418 chip->ops->shutdown(rtwdev);
1419 }
1420 EXPORT_SYMBOL(rtw_sdio_shutdown);
1421
1422 MODULE_AUTHOR("Martin Blumenstingl");
1423 MODULE_AUTHOR("Jernej Skrabec");
1424 MODULE_DESCRIPTION("Realtek 802.11ac wireless SDIO driver");
1425 MODULE_LICENSE("Dual BSD/GPL");
1426