1 /*
2 * linux/drivers/mmc/core/mmc_ops.h
3 *
4 * Copyright 2006-2007 Pierre Ossman
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or (at
9 * your option) any later version.
10 */
11
12 #include <linux/slab.h>
13 #include <linux/export.h>
14 #include <linux/types.h>
15 #include <linux/scatterlist.h>
16
17 #include <linux/mmc/host.h>
18 #include <linux/mmc/card.h>
19 #include <linux/mmc/mmc.h>
20
21 #include "core.h"
22 #include "host.h"
23 #include "mmc_ops.h"
24
25 #define MMC_OPS_TIMEOUT_MS (10 * 60 * 1000) /* 10 minute timeout */
26
27 static const u8 tuning_blk_pattern_4bit[] = {
28 0xff, 0x0f, 0xff, 0x00, 0xff, 0xcc, 0xc3, 0xcc,
29 0xc3, 0x3c, 0xcc, 0xff, 0xfe, 0xff, 0xfe, 0xef,
30 0xff, 0xdf, 0xff, 0xdd, 0xff, 0xfb, 0xff, 0xfb,
31 0xbf, 0xff, 0x7f, 0xff, 0x77, 0xf7, 0xbd, 0xef,
32 0xff, 0xf0, 0xff, 0xf0, 0x0f, 0xfc, 0xcc, 0x3c,
33 0xcc, 0x33, 0xcc, 0xcf, 0xff, 0xef, 0xff, 0xee,
34 0xff, 0xfd, 0xff, 0xfd, 0xdf, 0xff, 0xbf, 0xff,
35 0xbb, 0xff, 0xf7, 0xff, 0xf7, 0x7f, 0x7b, 0xde,
36 };
37
38 static const u8 tuning_blk_pattern_8bit[] = {
39 0xff, 0xff, 0x00, 0xff, 0xff, 0xff, 0x00, 0x00,
40 0xff, 0xff, 0xcc, 0xcc, 0xcc, 0x33, 0xcc, 0xcc,
41 0xcc, 0x33, 0x33, 0xcc, 0xcc, 0xcc, 0xff, 0xff,
42 0xff, 0xee, 0xff, 0xff, 0xff, 0xee, 0xee, 0xff,
43 0xff, 0xff, 0xdd, 0xff, 0xff, 0xff, 0xdd, 0xdd,
44 0xff, 0xff, 0xff, 0xbb, 0xff, 0xff, 0xff, 0xbb,
45 0xbb, 0xff, 0xff, 0xff, 0x77, 0xff, 0xff, 0xff,
46 0x77, 0x77, 0xff, 0x77, 0xbb, 0xdd, 0xee, 0xff,
47 0xff, 0xff, 0xff, 0x00, 0xff, 0xff, 0xff, 0x00,
48 0x00, 0xff, 0xff, 0xcc, 0xcc, 0xcc, 0x33, 0xcc,
49 0xcc, 0xcc, 0x33, 0x33, 0xcc, 0xcc, 0xcc, 0xff,
50 0xff, 0xff, 0xee, 0xff, 0xff, 0xff, 0xee, 0xee,
51 0xff, 0xff, 0xff, 0xdd, 0xff, 0xff, 0xff, 0xdd,
52 0xdd, 0xff, 0xff, 0xff, 0xbb, 0xff, 0xff, 0xff,
53 0xbb, 0xbb, 0xff, 0xff, 0xff, 0x77, 0xff, 0xff,
54 0xff, 0x77, 0x77, 0xff, 0x77, 0xbb, 0xdd, 0xee,
55 };
56
__mmc_send_status(struct mmc_card * card,u32 * status,bool ignore_crc)57 static inline int __mmc_send_status(struct mmc_card *card, u32 *status,
58 bool ignore_crc)
59 {
60 int err;
61 struct mmc_command cmd = {0};
62
63 BUG_ON(!card);
64 BUG_ON(!card->host);
65
66 cmd.opcode = MMC_SEND_STATUS;
67 if (!mmc_host_is_spi(card->host))
68 cmd.arg = card->rca << 16;
69 cmd.flags = MMC_RSP_SPI_R2 | MMC_RSP_R1 | MMC_CMD_AC;
70 if (ignore_crc)
71 cmd.flags &= ~MMC_RSP_CRC;
72
73 err = mmc_wait_for_cmd(card->host, &cmd, MMC_CMD_RETRIES);
74 if (err)
75 return err;
76
77 /* NOTE: callers are required to understand the difference
78 * between "native" and SPI format status words!
79 */
80 if (status)
81 *status = cmd.resp[0];
82
83 return 0;
84 }
85
mmc_send_status(struct mmc_card * card,u32 * status)86 int mmc_send_status(struct mmc_card *card, u32 *status)
87 {
88 return __mmc_send_status(card, status, false);
89 }
90
_mmc_select_card(struct mmc_host * host,struct mmc_card * card)91 static int _mmc_select_card(struct mmc_host *host, struct mmc_card *card)
92 {
93 struct mmc_command cmd = {0};
94
95 BUG_ON(!host);
96
97 cmd.opcode = MMC_SELECT_CARD;
98
99 if (card) {
100 cmd.arg = card->rca << 16;
101 cmd.flags = MMC_RSP_R1 | MMC_CMD_AC;
102 } else {
103 cmd.arg = 0;
104 cmd.flags = MMC_RSP_NONE | MMC_CMD_AC;
105 }
106
107 return mmc_wait_for_cmd(host, &cmd, MMC_CMD_RETRIES);
108 }
109
mmc_select_card(struct mmc_card * card)110 int mmc_select_card(struct mmc_card *card)
111 {
112 BUG_ON(!card);
113
114 return _mmc_select_card(card->host, card);
115 }
116
mmc_deselect_cards(struct mmc_host * host)117 int mmc_deselect_cards(struct mmc_host *host)
118 {
119 return _mmc_select_card(host, NULL);
120 }
121
122 /*
123 * Write the value specified in the device tree or board code into the optional
124 * 16 bit Driver Stage Register. This can be used to tune raise/fall times and
125 * drive strength of the DAT and CMD outputs. The actual meaning of a given
126 * value is hardware dependant.
127 * The presence of the DSR register can be determined from the CSD register,
128 * bit 76.
129 */
mmc_set_dsr(struct mmc_host * host)130 int mmc_set_dsr(struct mmc_host *host)
131 {
132 struct mmc_command cmd = {0};
133
134 cmd.opcode = MMC_SET_DSR;
135
136 cmd.arg = (host->dsr << 16) | 0xffff;
137 cmd.flags = MMC_RSP_NONE | MMC_CMD_AC;
138
139 return mmc_wait_for_cmd(host, &cmd, MMC_CMD_RETRIES);
140 }
141
mmc_go_idle(struct mmc_host * host)142 int mmc_go_idle(struct mmc_host *host)
143 {
144 int err;
145 struct mmc_command cmd = {0};
146
147 /*
148 * Non-SPI hosts need to prevent chipselect going active during
149 * GO_IDLE; that would put chips into SPI mode. Remind them of
150 * that in case of hardware that won't pull up DAT3/nCS otherwise.
151 *
152 * SPI hosts ignore ios.chip_select; it's managed according to
153 * rules that must accommodate non-MMC slaves which this layer
154 * won't even know about.
155 */
156 if (!mmc_host_is_spi(host)) {
157 mmc_set_chip_select(host, MMC_CS_HIGH);
158 mmc_delay(1);
159 }
160
161 cmd.opcode = MMC_GO_IDLE_STATE;
162 cmd.arg = 0;
163 cmd.flags = MMC_RSP_SPI_R1 | MMC_RSP_NONE | MMC_CMD_BC;
164
165 err = mmc_wait_for_cmd(host, &cmd, 0);
166
167 mmc_delay(1);
168
169 if (!mmc_host_is_spi(host)) {
170 mmc_set_chip_select(host, MMC_CS_DONTCARE);
171 mmc_delay(1);
172 }
173
174 host->use_spi_crc = 0;
175
176 return err;
177 }
178
mmc_send_op_cond(struct mmc_host * host,u32 ocr,u32 * rocr)179 int mmc_send_op_cond(struct mmc_host *host, u32 ocr, u32 *rocr)
180 {
181 struct mmc_command cmd = {0};
182 int i, err = 0;
183
184 BUG_ON(!host);
185
186 cmd.opcode = MMC_SEND_OP_COND;
187 cmd.arg = mmc_host_is_spi(host) ? 0 : ocr;
188 cmd.flags = MMC_RSP_SPI_R1 | MMC_RSP_R3 | MMC_CMD_BCR;
189
190 for (i = 100; i; i--) {
191 err = mmc_wait_for_cmd(host, &cmd, 0);
192 if (err)
193 break;
194
195 /* if we're just probing, do a single pass */
196 if (ocr == 0)
197 break;
198
199 /* otherwise wait until reset completes */
200 if (mmc_host_is_spi(host)) {
201 if (!(cmd.resp[0] & R1_SPI_IDLE))
202 break;
203 } else {
204 if (cmd.resp[0] & MMC_CARD_BUSY)
205 break;
206 }
207
208 err = -ETIMEDOUT;
209
210 mmc_delay(10);
211 }
212
213 if (rocr && !mmc_host_is_spi(host))
214 *rocr = cmd.resp[0];
215
216 return err;
217 }
218
mmc_all_send_cid(struct mmc_host * host,u32 * cid)219 int mmc_all_send_cid(struct mmc_host *host, u32 *cid)
220 {
221 int err;
222 struct mmc_command cmd = {0};
223
224 BUG_ON(!host);
225 BUG_ON(!cid);
226
227 cmd.opcode = MMC_ALL_SEND_CID;
228 cmd.arg = 0;
229 cmd.flags = MMC_RSP_R2 | MMC_CMD_BCR;
230
231 err = mmc_wait_for_cmd(host, &cmd, MMC_CMD_RETRIES);
232 if (err)
233 return err;
234
235 memcpy(cid, cmd.resp, sizeof(u32) * 4);
236
237 return 0;
238 }
239
mmc_set_relative_addr(struct mmc_card * card)240 int mmc_set_relative_addr(struct mmc_card *card)
241 {
242 struct mmc_command cmd = {0};
243
244 BUG_ON(!card);
245 BUG_ON(!card->host);
246
247 cmd.opcode = MMC_SET_RELATIVE_ADDR;
248 cmd.arg = card->rca << 16;
249 cmd.flags = MMC_RSP_R1 | MMC_CMD_AC;
250
251 return mmc_wait_for_cmd(card->host, &cmd, MMC_CMD_RETRIES);
252 }
253
254 static int
mmc_send_cxd_native(struct mmc_host * host,u32 arg,u32 * cxd,int opcode)255 mmc_send_cxd_native(struct mmc_host *host, u32 arg, u32 *cxd, int opcode)
256 {
257 int err;
258 struct mmc_command cmd = {0};
259
260 BUG_ON(!host);
261 BUG_ON(!cxd);
262
263 cmd.opcode = opcode;
264 cmd.arg = arg;
265 cmd.flags = MMC_RSP_R2 | MMC_CMD_AC;
266
267 err = mmc_wait_for_cmd(host, &cmd, MMC_CMD_RETRIES);
268 if (err)
269 return err;
270
271 memcpy(cxd, cmd.resp, sizeof(u32) * 4);
272
273 return 0;
274 }
275
276 /*
277 * NOTE: void *buf, caller for the buf is required to use DMA-capable
278 * buffer or on-stack buffer (with some overhead in callee).
279 */
280 static int
mmc_send_cxd_data(struct mmc_card * card,struct mmc_host * host,u32 opcode,void * buf,unsigned len)281 mmc_send_cxd_data(struct mmc_card *card, struct mmc_host *host,
282 u32 opcode, void *buf, unsigned len)
283 {
284 struct mmc_request mrq = {NULL};
285 struct mmc_command cmd = {0};
286 struct mmc_data data = {0};
287 struct scatterlist sg;
288
289 mrq.cmd = &cmd;
290 mrq.data = &data;
291
292 cmd.opcode = opcode;
293 cmd.arg = 0;
294
295 /* NOTE HACK: the MMC_RSP_SPI_R1 is always correct here, but we
296 * rely on callers to never use this with "native" calls for reading
297 * CSD or CID. Native versions of those commands use the R2 type,
298 * not R1 plus a data block.
299 */
300 cmd.flags = MMC_RSP_SPI_R1 | MMC_RSP_R1 | MMC_CMD_ADTC;
301
302 data.blksz = len;
303 data.blocks = 1;
304 data.flags = MMC_DATA_READ;
305 data.sg = &sg;
306 data.sg_len = 1;
307
308 sg_init_one(&sg, buf, len);
309
310 if (opcode == MMC_SEND_CSD || opcode == MMC_SEND_CID) {
311 /*
312 * The spec states that CSR and CID accesses have a timeout
313 * of 64 clock cycles.
314 */
315 data.timeout_ns = 0;
316 data.timeout_clks = 64;
317 } else
318 mmc_set_data_timeout(&data, card);
319
320 mmc_wait_for_req(host, &mrq);
321
322 if (cmd.error)
323 return cmd.error;
324 if (data.error)
325 return data.error;
326
327 return 0;
328 }
329
mmc_send_csd(struct mmc_card * card,u32 * csd)330 int mmc_send_csd(struct mmc_card *card, u32 *csd)
331 {
332 int ret, i;
333 u32 *csd_tmp;
334
335 if (!mmc_host_is_spi(card->host))
336 return mmc_send_cxd_native(card->host, card->rca << 16,
337 csd, MMC_SEND_CSD);
338
339 csd_tmp = kzalloc(16, GFP_KERNEL);
340 if (!csd_tmp)
341 return -ENOMEM;
342
343 ret = mmc_send_cxd_data(card, card->host, MMC_SEND_CSD, csd_tmp, 16);
344 if (ret)
345 goto err;
346
347 for (i = 0;i < 4;i++)
348 csd[i] = be32_to_cpu(csd_tmp[i]);
349
350 err:
351 kfree(csd_tmp);
352 return ret;
353 }
354
mmc_send_cid(struct mmc_host * host,u32 * cid)355 int mmc_send_cid(struct mmc_host *host, u32 *cid)
356 {
357 int ret, i;
358 u32 *cid_tmp;
359
360 if (!mmc_host_is_spi(host)) {
361 if (!host->card)
362 return -EINVAL;
363 return mmc_send_cxd_native(host, host->card->rca << 16,
364 cid, MMC_SEND_CID);
365 }
366
367 cid_tmp = kzalloc(16, GFP_KERNEL);
368 if (!cid_tmp)
369 return -ENOMEM;
370
371 ret = mmc_send_cxd_data(NULL, host, MMC_SEND_CID, cid_tmp, 16);
372 if (ret)
373 goto err;
374
375 for (i = 0;i < 4;i++)
376 cid[i] = be32_to_cpu(cid_tmp[i]);
377
378 err:
379 kfree(cid_tmp);
380 return ret;
381 }
382
mmc_get_ext_csd(struct mmc_card * card,u8 ** new_ext_csd)383 int mmc_get_ext_csd(struct mmc_card *card, u8 **new_ext_csd)
384 {
385 int err;
386 u8 *ext_csd;
387
388 if (!card || !new_ext_csd)
389 return -EINVAL;
390
391 if (!mmc_can_ext_csd(card))
392 return -EOPNOTSUPP;
393
394 /*
395 * As the ext_csd is so large and mostly unused, we don't store the
396 * raw block in mmc_card.
397 */
398 ext_csd = kzalloc(512, GFP_KERNEL);
399 if (!ext_csd)
400 return -ENOMEM;
401
402 err = mmc_send_cxd_data(card, card->host, MMC_SEND_EXT_CSD, ext_csd,
403 512);
404 if (err)
405 kfree(ext_csd);
406 else
407 *new_ext_csd = ext_csd;
408
409 return err;
410 }
411 EXPORT_SYMBOL_GPL(mmc_get_ext_csd);
412
mmc_spi_read_ocr(struct mmc_host * host,int highcap,u32 * ocrp)413 int mmc_spi_read_ocr(struct mmc_host *host, int highcap, u32 *ocrp)
414 {
415 struct mmc_command cmd = {0};
416 int err;
417
418 cmd.opcode = MMC_SPI_READ_OCR;
419 cmd.arg = highcap ? (1 << 30) : 0;
420 cmd.flags = MMC_RSP_SPI_R3;
421
422 err = mmc_wait_for_cmd(host, &cmd, 0);
423
424 *ocrp = cmd.resp[1];
425 return err;
426 }
427
mmc_spi_set_crc(struct mmc_host * host,int use_crc)428 int mmc_spi_set_crc(struct mmc_host *host, int use_crc)
429 {
430 struct mmc_command cmd = {0};
431 int err;
432
433 cmd.opcode = MMC_SPI_CRC_ON_OFF;
434 cmd.flags = MMC_RSP_SPI_R1;
435 cmd.arg = use_crc;
436
437 err = mmc_wait_for_cmd(host, &cmd, 0);
438 if (!err)
439 host->use_spi_crc = use_crc;
440 return err;
441 }
442
mmc_switch_status_error(struct mmc_host * host,u32 status)443 int mmc_switch_status_error(struct mmc_host *host, u32 status)
444 {
445 if (mmc_host_is_spi(host)) {
446 if (status & R1_SPI_ILLEGAL_COMMAND)
447 return -EBADMSG;
448 } else {
449 if (status & 0xFDFFA000)
450 pr_warn("%s: unexpected status %#x after switch\n",
451 mmc_hostname(host), status);
452 if (status & R1_SWITCH_ERROR)
453 return -EBADMSG;
454 }
455 return 0;
456 }
457
458 /**
459 * __mmc_switch - modify EXT_CSD register
460 * @card: the MMC card associated with the data transfer
461 * @set: cmd set values
462 * @index: EXT_CSD register index
463 * @value: value to program into EXT_CSD register
464 * @timeout_ms: timeout (ms) for operation performed by register write,
465 * timeout of zero implies maximum possible timeout
466 * @use_busy_signal: use the busy signal as response type
467 * @send_status: send status cmd to poll for busy
468 * @ignore_crc: ignore CRC errors when sending status cmd to poll for busy
469 *
470 * Modifies the EXT_CSD register for selected card.
471 */
__mmc_switch(struct mmc_card * card,u8 set,u8 index,u8 value,unsigned int timeout_ms,bool use_busy_signal,bool send_status,bool ignore_crc)472 int __mmc_switch(struct mmc_card *card, u8 set, u8 index, u8 value,
473 unsigned int timeout_ms, bool use_busy_signal, bool send_status,
474 bool ignore_crc)
475 {
476 struct mmc_host *host = card->host;
477 int err;
478 struct mmc_command cmd = {0};
479 unsigned long timeout;
480 u32 status = 0;
481 bool use_r1b_resp = use_busy_signal;
482 bool expired = false;
483 bool busy = false;
484
485 mmc_retune_hold(host);
486
487 /*
488 * If the cmd timeout and the max_busy_timeout of the host are both
489 * specified, let's validate them. A failure means we need to prevent
490 * the host from doing hw busy detection, which is done by converting
491 * to a R1 response instead of a R1B.
492 */
493 if (timeout_ms && host->max_busy_timeout &&
494 (timeout_ms > host->max_busy_timeout))
495 use_r1b_resp = false;
496
497 cmd.opcode = MMC_SWITCH;
498 cmd.arg = (MMC_SWITCH_MODE_WRITE_BYTE << 24) |
499 (index << 16) |
500 (value << 8) |
501 set;
502 cmd.flags = MMC_CMD_AC;
503 if (use_r1b_resp) {
504 cmd.flags |= MMC_RSP_SPI_R1B | MMC_RSP_R1B;
505 /*
506 * A busy_timeout of zero means the host can decide to use
507 * whatever value it finds suitable.
508 */
509 cmd.busy_timeout = timeout_ms;
510 } else {
511 cmd.flags |= MMC_RSP_SPI_R1 | MMC_RSP_R1;
512 }
513
514 if (index == EXT_CSD_SANITIZE_START)
515 cmd.sanitize_busy = true;
516
517 err = mmc_wait_for_cmd(host, &cmd, MMC_CMD_RETRIES);
518 if (err)
519 goto out;
520
521 /* No need to check card status in case of unblocking command */
522 if (!use_busy_signal)
523 goto out;
524
525 /*
526 * CRC errors shall only be ignored in cases were CMD13 is used to poll
527 * to detect busy completion.
528 */
529 if ((host->caps & MMC_CAP_WAIT_WHILE_BUSY) && use_r1b_resp)
530 ignore_crc = false;
531
532 /* We have an unspecified cmd timeout, use the fallback value. */
533 if (!timeout_ms)
534 timeout_ms = MMC_OPS_TIMEOUT_MS;
535
536 /* Must check status to be sure of no errors. */
537 timeout = jiffies + msecs_to_jiffies(timeout_ms) + 1;
538 do {
539 /*
540 * Due to the possibility of being preempted after
541 * sending the status command, check the expiration
542 * time first.
543 */
544 expired = time_after(jiffies, timeout);
545 if (send_status) {
546 err = __mmc_send_status(card, &status, ignore_crc);
547 if (err)
548 goto out;
549 }
550 if ((host->caps & MMC_CAP_WAIT_WHILE_BUSY) && use_r1b_resp)
551 break;
552 if (host->ops->card_busy) {
553 if (!host->ops->card_busy(host))
554 break;
555 busy = true;
556 }
557 if (mmc_host_is_spi(host))
558 break;
559
560 /*
561 * We are not allowed to issue a status command and the host
562 * does'nt support MMC_CAP_WAIT_WHILE_BUSY, then we can only
563 * rely on waiting for the stated timeout to be sufficient.
564 */
565 if (!send_status && !host->ops->card_busy) {
566 mmc_delay(timeout_ms);
567 goto out;
568 }
569
570 /* Timeout if the device never leaves the program state. */
571 if (expired &&
572 (R1_CURRENT_STATE(status) == R1_STATE_PRG || busy)) {
573 pr_err("%s: Card stuck in programming state! %s\n",
574 mmc_hostname(host), __func__);
575 err = -ETIMEDOUT;
576 goto out;
577 }
578 } while (R1_CURRENT_STATE(status) == R1_STATE_PRG || busy);
579
580 err = mmc_switch_status_error(host, status);
581 out:
582 mmc_retune_release(host);
583
584 return err;
585 }
586
mmc_switch(struct mmc_card * card,u8 set,u8 index,u8 value,unsigned int timeout_ms)587 int mmc_switch(struct mmc_card *card, u8 set, u8 index, u8 value,
588 unsigned int timeout_ms)
589 {
590 return __mmc_switch(card, set, index, value, timeout_ms, true, true,
591 false);
592 }
593 EXPORT_SYMBOL_GPL(mmc_switch);
594
mmc_send_tuning(struct mmc_host * host,u32 opcode,int * cmd_error)595 int mmc_send_tuning(struct mmc_host *host, u32 opcode, int *cmd_error)
596 {
597 struct mmc_request mrq = {NULL};
598 struct mmc_command cmd = {0};
599 struct mmc_data data = {0};
600 struct scatterlist sg;
601 struct mmc_ios *ios = &host->ios;
602 const u8 *tuning_block_pattern;
603 int size, err = 0;
604 u8 *data_buf;
605
606 if (ios->bus_width == MMC_BUS_WIDTH_8) {
607 tuning_block_pattern = tuning_blk_pattern_8bit;
608 size = sizeof(tuning_blk_pattern_8bit);
609 } else if (ios->bus_width == MMC_BUS_WIDTH_4) {
610 tuning_block_pattern = tuning_blk_pattern_4bit;
611 size = sizeof(tuning_blk_pattern_4bit);
612 } else
613 return -EINVAL;
614
615 data_buf = kzalloc(size, GFP_KERNEL);
616 if (!data_buf)
617 return -ENOMEM;
618
619 mrq.cmd = &cmd;
620 mrq.data = &data;
621
622 cmd.opcode = opcode;
623 cmd.flags = MMC_RSP_R1 | MMC_CMD_ADTC;
624
625 data.blksz = size;
626 data.blocks = 1;
627 data.flags = MMC_DATA_READ;
628
629 /*
630 * According to the tuning specs, Tuning process
631 * is normally shorter 40 executions of CMD19,
632 * and timeout value should be shorter than 150 ms
633 */
634 data.timeout_ns = 150 * NSEC_PER_MSEC;
635
636 data.sg = &sg;
637 data.sg_len = 1;
638 sg_init_one(&sg, data_buf, size);
639
640 mmc_wait_for_req(host, &mrq);
641
642 if (cmd_error)
643 *cmd_error = cmd.error;
644
645 if (cmd.error) {
646 err = cmd.error;
647 goto out;
648 }
649
650 if (data.error) {
651 err = data.error;
652 goto out;
653 }
654
655 if (memcmp(data_buf, tuning_block_pattern, size))
656 err = -EIO;
657
658 out:
659 kfree(data_buf);
660 return err;
661 }
662 EXPORT_SYMBOL_GPL(mmc_send_tuning);
663
664 static int
mmc_send_bus_test(struct mmc_card * card,struct mmc_host * host,u8 opcode,u8 len)665 mmc_send_bus_test(struct mmc_card *card, struct mmc_host *host, u8 opcode,
666 u8 len)
667 {
668 struct mmc_request mrq = {NULL};
669 struct mmc_command cmd = {0};
670 struct mmc_data data = {0};
671 struct scatterlist sg;
672 u8 *data_buf;
673 u8 *test_buf;
674 int i, err;
675 static u8 testdata_8bit[8] = { 0x55, 0xaa, 0, 0, 0, 0, 0, 0 };
676 static u8 testdata_4bit[4] = { 0x5a, 0, 0, 0 };
677
678 /* dma onto stack is unsafe/nonportable, but callers to this
679 * routine normally provide temporary on-stack buffers ...
680 */
681 data_buf = kmalloc(len, GFP_KERNEL);
682 if (!data_buf)
683 return -ENOMEM;
684
685 if (len == 8)
686 test_buf = testdata_8bit;
687 else if (len == 4)
688 test_buf = testdata_4bit;
689 else {
690 pr_err("%s: Invalid bus_width %d\n",
691 mmc_hostname(host), len);
692 kfree(data_buf);
693 return -EINVAL;
694 }
695
696 if (opcode == MMC_BUS_TEST_W)
697 memcpy(data_buf, test_buf, len);
698
699 mrq.cmd = &cmd;
700 mrq.data = &data;
701 cmd.opcode = opcode;
702 cmd.arg = 0;
703
704 /* NOTE HACK: the MMC_RSP_SPI_R1 is always correct here, but we
705 * rely on callers to never use this with "native" calls for reading
706 * CSD or CID. Native versions of those commands use the R2 type,
707 * not R1 plus a data block.
708 */
709 cmd.flags = MMC_RSP_SPI_R1 | MMC_RSP_R1 | MMC_CMD_ADTC;
710
711 data.blksz = len;
712 data.blocks = 1;
713 if (opcode == MMC_BUS_TEST_R)
714 data.flags = MMC_DATA_READ;
715 else
716 data.flags = MMC_DATA_WRITE;
717
718 data.sg = &sg;
719 data.sg_len = 1;
720 mmc_set_data_timeout(&data, card);
721 sg_init_one(&sg, data_buf, len);
722 mmc_wait_for_req(host, &mrq);
723 err = 0;
724 if (opcode == MMC_BUS_TEST_R) {
725 for (i = 0; i < len / 4; i++)
726 if ((test_buf[i] ^ data_buf[i]) != 0xff) {
727 err = -EIO;
728 break;
729 }
730 }
731 kfree(data_buf);
732
733 if (cmd.error)
734 return cmd.error;
735 if (data.error)
736 return data.error;
737
738 return err;
739 }
740
mmc_bus_test(struct mmc_card * card,u8 bus_width)741 int mmc_bus_test(struct mmc_card *card, u8 bus_width)
742 {
743 int width;
744
745 if (bus_width == MMC_BUS_WIDTH_8)
746 width = 8;
747 else if (bus_width == MMC_BUS_WIDTH_4)
748 width = 4;
749 else if (bus_width == MMC_BUS_WIDTH_1)
750 return 0; /* no need for test */
751 else
752 return -EINVAL;
753
754 /*
755 * Ignore errors from BUS_TEST_W. BUS_TEST_R will fail if there
756 * is a problem. This improves chances that the test will work.
757 */
758 mmc_send_bus_test(card, card->host, MMC_BUS_TEST_W, width);
759 return mmc_send_bus_test(card, card->host, MMC_BUS_TEST_R, width);
760 }
761
mmc_send_hpi_cmd(struct mmc_card * card,u32 * status)762 int mmc_send_hpi_cmd(struct mmc_card *card, u32 *status)
763 {
764 struct mmc_command cmd = {0};
765 unsigned int opcode;
766 int err;
767
768 if (!card->ext_csd.hpi) {
769 pr_warn("%s: Card didn't support HPI command\n",
770 mmc_hostname(card->host));
771 return -EINVAL;
772 }
773
774 opcode = card->ext_csd.hpi_cmd;
775 if (opcode == MMC_STOP_TRANSMISSION)
776 cmd.flags = MMC_RSP_R1B | MMC_CMD_AC;
777 else if (opcode == MMC_SEND_STATUS)
778 cmd.flags = MMC_RSP_R1 | MMC_CMD_AC;
779
780 cmd.opcode = opcode;
781 cmd.arg = card->rca << 16 | 1;
782
783 err = mmc_wait_for_cmd(card->host, &cmd, 0);
784 if (err) {
785 pr_warn("%s: error %d interrupting operation. "
786 "HPI command response %#x\n", mmc_hostname(card->host),
787 err, cmd.resp[0]);
788 return err;
789 }
790 if (status)
791 *status = cmd.resp[0];
792
793 return 0;
794 }
795
mmc_can_ext_csd(struct mmc_card * card)796 int mmc_can_ext_csd(struct mmc_card *card)
797 {
798 return (card && card->csd.mmca_vsn > CSD_SPEC_VER_3);
799 }
800