1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Portions copyright (C) 2003 Russell King, PXA MMCI Driver
4 * Portions copyright (C) 2004-2005 Pierre Ossman, W83L51xD SD/MMC driver
5 *
6 * Copyright 2008 Embedded Alley Solutions, Inc.
7 * Copyright 2009-2011 Freescale Semiconductor, Inc.
8 */
9
10 #include <linux/kernel.h>
11 #include <linux/init.h>
12 #include <linux/ioport.h>
13 #include <linux/of.h>
14 #include <linux/of_device.h>
15 #include <linux/platform_device.h>
16 #include <linux/delay.h>
17 #include <linux/interrupt.h>
18 #include <linux/dma-mapping.h>
19 #include <linux/dmaengine.h>
20 #include <linux/dma/mxs-dma.h>
21 #include <linux/highmem.h>
22 #include <linux/clk.h>
23 #include <linux/err.h>
24 #include <linux/completion.h>
25 #include <linux/mmc/host.h>
26 #include <linux/mmc/mmc.h>
27 #include <linux/mmc/sdio.h>
28 #include <linux/mmc/slot-gpio.h>
29 #include <linux/regulator/consumer.h>
30 #include <linux/module.h>
31 #include <linux/stmp_device.h>
32 #include <linux/spi/mxs-spi.h>
33
34 #define DRIVER_NAME "mxs-mmc"
35
36 #define MXS_MMC_IRQ_BITS (BM_SSP_CTRL1_SDIO_IRQ | \
37 BM_SSP_CTRL1_RESP_ERR_IRQ | \
38 BM_SSP_CTRL1_RESP_TIMEOUT_IRQ | \
39 BM_SSP_CTRL1_DATA_TIMEOUT_IRQ | \
40 BM_SSP_CTRL1_DATA_CRC_IRQ | \
41 BM_SSP_CTRL1_FIFO_UNDERRUN_IRQ | \
42 BM_SSP_CTRL1_RECV_TIMEOUT_IRQ | \
43 BM_SSP_CTRL1_FIFO_OVERRUN_IRQ)
44
45 /* card detect polling timeout */
46 #define MXS_MMC_DETECT_TIMEOUT (HZ/2)
47
48 struct mxs_mmc_host {
49 struct mxs_ssp ssp;
50
51 struct mmc_host *mmc;
52 struct mmc_request *mrq;
53 struct mmc_command *cmd;
54 struct mmc_data *data;
55
56 unsigned char bus_width;
57 spinlock_t lock;
58 int sdio_irq_en;
59 bool broken_cd;
60 };
61
mxs_mmc_get_cd(struct mmc_host * mmc)62 static int mxs_mmc_get_cd(struct mmc_host *mmc)
63 {
64 struct mxs_mmc_host *host = mmc_priv(mmc);
65 struct mxs_ssp *ssp = &host->ssp;
66 int present, ret;
67
68 if (host->broken_cd)
69 return -ENOSYS;
70
71 ret = mmc_gpio_get_cd(mmc);
72 if (ret >= 0)
73 return ret;
74
75 present = mmc->caps & MMC_CAP_NEEDS_POLL ||
76 !(readl(ssp->base + HW_SSP_STATUS(ssp)) &
77 BM_SSP_STATUS_CARD_DETECT);
78
79 if (mmc->caps2 & MMC_CAP2_CD_ACTIVE_HIGH)
80 present = !present;
81
82 return present;
83 }
84
mxs_mmc_reset(struct mxs_mmc_host * host)85 static int mxs_mmc_reset(struct mxs_mmc_host *host)
86 {
87 struct mxs_ssp *ssp = &host->ssp;
88 u32 ctrl0, ctrl1;
89 int ret;
90
91 ret = stmp_reset_block(ssp->base);
92 if (ret)
93 return ret;
94
95 ctrl0 = BM_SSP_CTRL0_IGNORE_CRC;
96 ctrl1 = BF_SSP(0x3, CTRL1_SSP_MODE) |
97 BF_SSP(0x7, CTRL1_WORD_LENGTH) |
98 BM_SSP_CTRL1_DMA_ENABLE |
99 BM_SSP_CTRL1_POLARITY |
100 BM_SSP_CTRL1_RECV_TIMEOUT_IRQ_EN |
101 BM_SSP_CTRL1_DATA_CRC_IRQ_EN |
102 BM_SSP_CTRL1_DATA_TIMEOUT_IRQ_EN |
103 BM_SSP_CTRL1_RESP_TIMEOUT_IRQ_EN |
104 BM_SSP_CTRL1_RESP_ERR_IRQ_EN;
105
106 writel(BF_SSP(0xffff, TIMING_TIMEOUT) |
107 BF_SSP(2, TIMING_CLOCK_DIVIDE) |
108 BF_SSP(0, TIMING_CLOCK_RATE),
109 ssp->base + HW_SSP_TIMING(ssp));
110
111 if (host->sdio_irq_en) {
112 ctrl0 |= BM_SSP_CTRL0_SDIO_IRQ_CHECK;
113 ctrl1 |= BM_SSP_CTRL1_SDIO_IRQ_EN;
114 }
115
116 writel(ctrl0, ssp->base + HW_SSP_CTRL0);
117 writel(ctrl1, ssp->base + HW_SSP_CTRL1(ssp));
118 return 0;
119 }
120
121 static void mxs_mmc_start_cmd(struct mxs_mmc_host *host,
122 struct mmc_command *cmd);
123
mxs_mmc_request_done(struct mxs_mmc_host * host)124 static void mxs_mmc_request_done(struct mxs_mmc_host *host)
125 {
126 struct mmc_command *cmd = host->cmd;
127 struct mmc_data *data = host->data;
128 struct mmc_request *mrq = host->mrq;
129 struct mxs_ssp *ssp = &host->ssp;
130
131 if (mmc_resp_type(cmd) & MMC_RSP_PRESENT) {
132 if (mmc_resp_type(cmd) & MMC_RSP_136) {
133 cmd->resp[3] = readl(ssp->base + HW_SSP_SDRESP0(ssp));
134 cmd->resp[2] = readl(ssp->base + HW_SSP_SDRESP1(ssp));
135 cmd->resp[1] = readl(ssp->base + HW_SSP_SDRESP2(ssp));
136 cmd->resp[0] = readl(ssp->base + HW_SSP_SDRESP3(ssp));
137 } else {
138 cmd->resp[0] = readl(ssp->base + HW_SSP_SDRESP0(ssp));
139 }
140 }
141
142 if (cmd == mrq->sbc) {
143 /* Finished CMD23, now send actual command. */
144 mxs_mmc_start_cmd(host, mrq->cmd);
145 return;
146 } else if (data) {
147 dma_unmap_sg(mmc_dev(host->mmc), data->sg,
148 data->sg_len, ssp->dma_dir);
149 /*
150 * If there was an error on any block, we mark all
151 * data blocks as being in error.
152 */
153 if (!data->error)
154 data->bytes_xfered = data->blocks * data->blksz;
155 else
156 data->bytes_xfered = 0;
157
158 host->data = NULL;
159 if (data->stop && (data->error || !mrq->sbc)) {
160 mxs_mmc_start_cmd(host, mrq->stop);
161 return;
162 }
163 }
164
165 host->mrq = NULL;
166 mmc_request_done(host->mmc, mrq);
167 }
168
mxs_mmc_dma_irq_callback(void * param)169 static void mxs_mmc_dma_irq_callback(void *param)
170 {
171 struct mxs_mmc_host *host = param;
172
173 mxs_mmc_request_done(host);
174 }
175
mxs_mmc_irq_handler(int irq,void * dev_id)176 static irqreturn_t mxs_mmc_irq_handler(int irq, void *dev_id)
177 {
178 struct mxs_mmc_host *host = dev_id;
179 struct mmc_command *cmd = host->cmd;
180 struct mmc_data *data = host->data;
181 struct mxs_ssp *ssp = &host->ssp;
182 u32 stat;
183
184 spin_lock(&host->lock);
185
186 stat = readl(ssp->base + HW_SSP_CTRL1(ssp));
187 writel(stat & MXS_MMC_IRQ_BITS,
188 ssp->base + HW_SSP_CTRL1(ssp) + STMP_OFFSET_REG_CLR);
189
190 spin_unlock(&host->lock);
191
192 if ((stat & BM_SSP_CTRL1_SDIO_IRQ) && (stat & BM_SSP_CTRL1_SDIO_IRQ_EN))
193 mmc_signal_sdio_irq(host->mmc);
194
195 if (stat & BM_SSP_CTRL1_RESP_TIMEOUT_IRQ)
196 cmd->error = -ETIMEDOUT;
197 else if (stat & BM_SSP_CTRL1_RESP_ERR_IRQ)
198 cmd->error = -EIO;
199
200 if (data) {
201 if (stat & (BM_SSP_CTRL1_DATA_TIMEOUT_IRQ |
202 BM_SSP_CTRL1_RECV_TIMEOUT_IRQ))
203 data->error = -ETIMEDOUT;
204 else if (stat & BM_SSP_CTRL1_DATA_CRC_IRQ)
205 data->error = -EILSEQ;
206 else if (stat & (BM_SSP_CTRL1_FIFO_UNDERRUN_IRQ |
207 BM_SSP_CTRL1_FIFO_OVERRUN_IRQ))
208 data->error = -EIO;
209 }
210
211 return IRQ_HANDLED;
212 }
213
mxs_mmc_prep_dma(struct mxs_mmc_host * host,unsigned long flags)214 static struct dma_async_tx_descriptor *mxs_mmc_prep_dma(
215 struct mxs_mmc_host *host, unsigned long flags)
216 {
217 struct mxs_ssp *ssp = &host->ssp;
218 struct dma_async_tx_descriptor *desc;
219 struct mmc_data *data = host->data;
220 struct scatterlist * sgl;
221 unsigned int sg_len;
222
223 if (data) {
224 /* data */
225 dma_map_sg(mmc_dev(host->mmc), data->sg,
226 data->sg_len, ssp->dma_dir);
227 sgl = data->sg;
228 sg_len = data->sg_len;
229 } else {
230 /* pio */
231 sgl = (struct scatterlist *) ssp->ssp_pio_words;
232 sg_len = SSP_PIO_NUM;
233 }
234
235 desc = dmaengine_prep_slave_sg(ssp->dmach,
236 sgl, sg_len, ssp->slave_dirn, flags);
237 if (desc) {
238 desc->callback = mxs_mmc_dma_irq_callback;
239 desc->callback_param = host;
240 } else {
241 if (data)
242 dma_unmap_sg(mmc_dev(host->mmc), data->sg,
243 data->sg_len, ssp->dma_dir);
244 }
245
246 return desc;
247 }
248
mxs_mmc_bc(struct mxs_mmc_host * host)249 static void mxs_mmc_bc(struct mxs_mmc_host *host)
250 {
251 struct mxs_ssp *ssp = &host->ssp;
252 struct mmc_command *cmd = host->cmd;
253 struct dma_async_tx_descriptor *desc;
254 u32 ctrl0, cmd0, cmd1;
255
256 ctrl0 = BM_SSP_CTRL0_ENABLE | BM_SSP_CTRL0_IGNORE_CRC;
257 cmd0 = BF_SSP(cmd->opcode, CMD0_CMD) | BM_SSP_CMD0_APPEND_8CYC;
258 cmd1 = cmd->arg;
259
260 if (host->sdio_irq_en) {
261 ctrl0 |= BM_SSP_CTRL0_SDIO_IRQ_CHECK;
262 cmd0 |= BM_SSP_CMD0_CONT_CLKING_EN | BM_SSP_CMD0_SLOW_CLKING_EN;
263 }
264
265 ssp->ssp_pio_words[0] = ctrl0;
266 ssp->ssp_pio_words[1] = cmd0;
267 ssp->ssp_pio_words[2] = cmd1;
268 ssp->dma_dir = DMA_NONE;
269 ssp->slave_dirn = DMA_TRANS_NONE;
270 desc = mxs_mmc_prep_dma(host, MXS_DMA_CTRL_WAIT4END);
271 if (!desc)
272 goto out;
273
274 dmaengine_submit(desc);
275 dma_async_issue_pending(ssp->dmach);
276 return;
277
278 out:
279 dev_warn(mmc_dev(host->mmc),
280 "%s: failed to prep dma\n", __func__);
281 }
282
mxs_mmc_ac(struct mxs_mmc_host * host)283 static void mxs_mmc_ac(struct mxs_mmc_host *host)
284 {
285 struct mxs_ssp *ssp = &host->ssp;
286 struct mmc_command *cmd = host->cmd;
287 struct dma_async_tx_descriptor *desc;
288 u32 ignore_crc, get_resp, long_resp;
289 u32 ctrl0, cmd0, cmd1;
290
291 ignore_crc = (mmc_resp_type(cmd) & MMC_RSP_CRC) ?
292 0 : BM_SSP_CTRL0_IGNORE_CRC;
293 get_resp = (mmc_resp_type(cmd) & MMC_RSP_PRESENT) ?
294 BM_SSP_CTRL0_GET_RESP : 0;
295 long_resp = (mmc_resp_type(cmd) & MMC_RSP_136) ?
296 BM_SSP_CTRL0_LONG_RESP : 0;
297
298 ctrl0 = BM_SSP_CTRL0_ENABLE | ignore_crc | get_resp | long_resp;
299 cmd0 = BF_SSP(cmd->opcode, CMD0_CMD);
300 cmd1 = cmd->arg;
301
302 if (cmd->opcode == MMC_STOP_TRANSMISSION)
303 cmd0 |= BM_SSP_CMD0_APPEND_8CYC;
304
305 if (host->sdio_irq_en) {
306 ctrl0 |= BM_SSP_CTRL0_SDIO_IRQ_CHECK;
307 cmd0 |= BM_SSP_CMD0_CONT_CLKING_EN | BM_SSP_CMD0_SLOW_CLKING_EN;
308 }
309
310 ssp->ssp_pio_words[0] = ctrl0;
311 ssp->ssp_pio_words[1] = cmd0;
312 ssp->ssp_pio_words[2] = cmd1;
313 ssp->dma_dir = DMA_NONE;
314 ssp->slave_dirn = DMA_TRANS_NONE;
315 desc = mxs_mmc_prep_dma(host, MXS_DMA_CTRL_WAIT4END);
316 if (!desc)
317 goto out;
318
319 dmaengine_submit(desc);
320 dma_async_issue_pending(ssp->dmach);
321 return;
322
323 out:
324 dev_warn(mmc_dev(host->mmc),
325 "%s: failed to prep dma\n", __func__);
326 }
327
mxs_ns_to_ssp_ticks(unsigned clock_rate,unsigned ns)328 static unsigned short mxs_ns_to_ssp_ticks(unsigned clock_rate, unsigned ns)
329 {
330 const unsigned int ssp_timeout_mul = 4096;
331 /*
332 * Calculate ticks in ms since ns are large numbers
333 * and might overflow
334 */
335 const unsigned int clock_per_ms = clock_rate / 1000;
336 const unsigned int ms = ns / 1000;
337 const unsigned int ticks = ms * clock_per_ms;
338 const unsigned int ssp_ticks = ticks / ssp_timeout_mul;
339
340 WARN_ON(ssp_ticks == 0);
341 return ssp_ticks;
342 }
343
mxs_mmc_adtc(struct mxs_mmc_host * host)344 static void mxs_mmc_adtc(struct mxs_mmc_host *host)
345 {
346 struct mmc_command *cmd = host->cmd;
347 struct mmc_data *data = cmd->data;
348 struct dma_async_tx_descriptor *desc;
349 struct scatterlist *sgl = data->sg, *sg;
350 unsigned int sg_len = data->sg_len;
351 unsigned int i;
352
353 unsigned short dma_data_dir, timeout;
354 enum dma_transfer_direction slave_dirn;
355 unsigned int data_size = 0, log2_blksz;
356 unsigned int blocks = data->blocks;
357
358 struct mxs_ssp *ssp = &host->ssp;
359
360 u32 ignore_crc, get_resp, long_resp, read;
361 u32 ctrl0, cmd0, cmd1, val;
362
363 ignore_crc = (mmc_resp_type(cmd) & MMC_RSP_CRC) ?
364 0 : BM_SSP_CTRL0_IGNORE_CRC;
365 get_resp = (mmc_resp_type(cmd) & MMC_RSP_PRESENT) ?
366 BM_SSP_CTRL0_GET_RESP : 0;
367 long_resp = (mmc_resp_type(cmd) & MMC_RSP_136) ?
368 BM_SSP_CTRL0_LONG_RESP : 0;
369
370 if (data->flags & MMC_DATA_WRITE) {
371 dma_data_dir = DMA_TO_DEVICE;
372 slave_dirn = DMA_MEM_TO_DEV;
373 read = 0;
374 } else {
375 dma_data_dir = DMA_FROM_DEVICE;
376 slave_dirn = DMA_DEV_TO_MEM;
377 read = BM_SSP_CTRL0_READ;
378 }
379
380 ctrl0 = BF_SSP(host->bus_width, CTRL0_BUS_WIDTH) |
381 ignore_crc | get_resp | long_resp |
382 BM_SSP_CTRL0_DATA_XFER | read |
383 BM_SSP_CTRL0_WAIT_FOR_IRQ |
384 BM_SSP_CTRL0_ENABLE;
385
386 cmd0 = BF_SSP(cmd->opcode, CMD0_CMD);
387
388 /* get logarithm to base 2 of block size for setting register */
389 log2_blksz = ilog2(data->blksz);
390
391 /*
392 * take special care of the case that data size from data->sg
393 * is not equal to blocks x blksz
394 */
395 for_each_sg(sgl, sg, sg_len, i)
396 data_size += sg->length;
397
398 if (data_size != data->blocks * data->blksz)
399 blocks = 1;
400
401 /* xfer count, block size and count need to be set differently */
402 if (ssp_is_old(ssp)) {
403 ctrl0 |= BF_SSP(data_size, CTRL0_XFER_COUNT);
404 cmd0 |= BF_SSP(log2_blksz, CMD0_BLOCK_SIZE) |
405 BF_SSP(blocks - 1, CMD0_BLOCK_COUNT);
406 } else {
407 writel(data_size, ssp->base + HW_SSP_XFER_SIZE);
408 writel(BF_SSP(log2_blksz, BLOCK_SIZE_BLOCK_SIZE) |
409 BF_SSP(blocks - 1, BLOCK_SIZE_BLOCK_COUNT),
410 ssp->base + HW_SSP_BLOCK_SIZE);
411 }
412
413 if (cmd->opcode == SD_IO_RW_EXTENDED)
414 cmd0 |= BM_SSP_CMD0_APPEND_8CYC;
415
416 cmd1 = cmd->arg;
417
418 if (host->sdio_irq_en) {
419 ctrl0 |= BM_SSP_CTRL0_SDIO_IRQ_CHECK;
420 cmd0 |= BM_SSP_CMD0_CONT_CLKING_EN | BM_SSP_CMD0_SLOW_CLKING_EN;
421 }
422
423 /* set the timeout count */
424 timeout = mxs_ns_to_ssp_ticks(ssp->clk_rate, data->timeout_ns);
425 val = readl(ssp->base + HW_SSP_TIMING(ssp));
426 val &= ~(BM_SSP_TIMING_TIMEOUT);
427 val |= BF_SSP(timeout, TIMING_TIMEOUT);
428 writel(val, ssp->base + HW_SSP_TIMING(ssp));
429
430 /* pio */
431 ssp->ssp_pio_words[0] = ctrl0;
432 ssp->ssp_pio_words[1] = cmd0;
433 ssp->ssp_pio_words[2] = cmd1;
434 ssp->dma_dir = DMA_NONE;
435 ssp->slave_dirn = DMA_TRANS_NONE;
436 desc = mxs_mmc_prep_dma(host, 0);
437 if (!desc)
438 goto out;
439
440 /* append data sg */
441 WARN_ON(host->data != NULL);
442 host->data = data;
443 ssp->dma_dir = dma_data_dir;
444 ssp->slave_dirn = slave_dirn;
445 desc = mxs_mmc_prep_dma(host, DMA_PREP_INTERRUPT | MXS_DMA_CTRL_WAIT4END);
446 if (!desc)
447 goto out;
448
449 dmaengine_submit(desc);
450 dma_async_issue_pending(ssp->dmach);
451 return;
452 out:
453 dev_warn(mmc_dev(host->mmc),
454 "%s: failed to prep dma\n", __func__);
455 }
456
mxs_mmc_start_cmd(struct mxs_mmc_host * host,struct mmc_command * cmd)457 static void mxs_mmc_start_cmd(struct mxs_mmc_host *host,
458 struct mmc_command *cmd)
459 {
460 host->cmd = cmd;
461
462 switch (mmc_cmd_type(cmd)) {
463 case MMC_CMD_BC:
464 mxs_mmc_bc(host);
465 break;
466 case MMC_CMD_BCR:
467 mxs_mmc_ac(host);
468 break;
469 case MMC_CMD_AC:
470 mxs_mmc_ac(host);
471 break;
472 case MMC_CMD_ADTC:
473 mxs_mmc_adtc(host);
474 break;
475 default:
476 dev_warn(mmc_dev(host->mmc),
477 "%s: unknown MMC command\n", __func__);
478 break;
479 }
480 }
481
mxs_mmc_request(struct mmc_host * mmc,struct mmc_request * mrq)482 static void mxs_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq)
483 {
484 struct mxs_mmc_host *host = mmc_priv(mmc);
485
486 WARN_ON(host->mrq != NULL);
487 host->mrq = mrq;
488
489 if (mrq->sbc)
490 mxs_mmc_start_cmd(host, mrq->sbc);
491 else
492 mxs_mmc_start_cmd(host, mrq->cmd);
493 }
494
mxs_mmc_set_ios(struct mmc_host * mmc,struct mmc_ios * ios)495 static void mxs_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
496 {
497 struct mxs_mmc_host *host = mmc_priv(mmc);
498
499 if (ios->bus_width == MMC_BUS_WIDTH_8)
500 host->bus_width = 2;
501 else if (ios->bus_width == MMC_BUS_WIDTH_4)
502 host->bus_width = 1;
503 else
504 host->bus_width = 0;
505
506 if (ios->clock)
507 mxs_ssp_set_clk_rate(&host->ssp, ios->clock);
508 }
509
mxs_mmc_enable_sdio_irq(struct mmc_host * mmc,int enable)510 static void mxs_mmc_enable_sdio_irq(struct mmc_host *mmc, int enable)
511 {
512 struct mxs_mmc_host *host = mmc_priv(mmc);
513 struct mxs_ssp *ssp = &host->ssp;
514 unsigned long flags;
515
516 spin_lock_irqsave(&host->lock, flags);
517
518 host->sdio_irq_en = enable;
519
520 if (enable) {
521 writel(BM_SSP_CTRL0_SDIO_IRQ_CHECK,
522 ssp->base + HW_SSP_CTRL0 + STMP_OFFSET_REG_SET);
523 writel(BM_SSP_CTRL1_SDIO_IRQ_EN,
524 ssp->base + HW_SSP_CTRL1(ssp) + STMP_OFFSET_REG_SET);
525 } else {
526 writel(BM_SSP_CTRL0_SDIO_IRQ_CHECK,
527 ssp->base + HW_SSP_CTRL0 + STMP_OFFSET_REG_CLR);
528 writel(BM_SSP_CTRL1_SDIO_IRQ_EN,
529 ssp->base + HW_SSP_CTRL1(ssp) + STMP_OFFSET_REG_CLR);
530 }
531
532 spin_unlock_irqrestore(&host->lock, flags);
533
534 if (enable && readl(ssp->base + HW_SSP_STATUS(ssp)) &
535 BM_SSP_STATUS_SDIO_IRQ)
536 mmc_signal_sdio_irq(host->mmc);
537
538 }
539
540 static const struct mmc_host_ops mxs_mmc_ops = {
541 .request = mxs_mmc_request,
542 .get_ro = mmc_gpio_get_ro,
543 .get_cd = mxs_mmc_get_cd,
544 .set_ios = mxs_mmc_set_ios,
545 .enable_sdio_irq = mxs_mmc_enable_sdio_irq,
546 };
547
548 static const struct platform_device_id mxs_ssp_ids[] = {
549 {
550 .name = "imx23-mmc",
551 .driver_data = IMX23_SSP,
552 }, {
553 .name = "imx28-mmc",
554 .driver_data = IMX28_SSP,
555 }, {
556 /* sentinel */
557 }
558 };
559 MODULE_DEVICE_TABLE(platform, mxs_ssp_ids);
560
561 static const struct of_device_id mxs_mmc_dt_ids[] = {
562 { .compatible = "fsl,imx23-mmc", .data = (void *) IMX23_SSP, },
563 { .compatible = "fsl,imx28-mmc", .data = (void *) IMX28_SSP, },
564 { /* sentinel */ }
565 };
566 MODULE_DEVICE_TABLE(of, mxs_mmc_dt_ids);
567
mxs_mmc_regulator_disable(void * regulator)568 static void mxs_mmc_regulator_disable(void *regulator)
569 {
570 regulator_disable(regulator);
571 }
572
mxs_mmc_probe(struct platform_device * pdev)573 static int mxs_mmc_probe(struct platform_device *pdev)
574 {
575 const struct of_device_id *of_id =
576 of_match_device(mxs_mmc_dt_ids, &pdev->dev);
577 struct device_node *np = pdev->dev.of_node;
578 struct mxs_mmc_host *host;
579 struct mmc_host *mmc;
580 int ret = 0, irq_err;
581 struct regulator *reg_vmmc;
582 struct mxs_ssp *ssp;
583
584 irq_err = platform_get_irq(pdev, 0);
585 if (irq_err < 0)
586 return irq_err;
587
588 mmc = mmc_alloc_host(sizeof(struct mxs_mmc_host), &pdev->dev);
589 if (!mmc)
590 return -ENOMEM;
591
592 host = mmc_priv(mmc);
593 ssp = &host->ssp;
594 ssp->dev = &pdev->dev;
595 ssp->base = devm_platform_ioremap_resource(pdev, 0);
596 if (IS_ERR(ssp->base)) {
597 ret = PTR_ERR(ssp->base);
598 goto out_mmc_free;
599 }
600
601 ssp->devid = (enum mxs_ssp_id) of_id->data;
602
603 host->mmc = mmc;
604 host->sdio_irq_en = 0;
605
606 reg_vmmc = devm_regulator_get(&pdev->dev, "vmmc");
607 if (!IS_ERR(reg_vmmc)) {
608 ret = regulator_enable(reg_vmmc);
609 if (ret) {
610 dev_err(&pdev->dev,
611 "Failed to enable vmmc regulator: %d\n", ret);
612 goto out_mmc_free;
613 }
614
615 ret = devm_add_action_or_reset(&pdev->dev, mxs_mmc_regulator_disable,
616 reg_vmmc);
617 if (ret)
618 goto out_mmc_free;
619 }
620
621 ssp->clk = devm_clk_get(&pdev->dev, NULL);
622 if (IS_ERR(ssp->clk)) {
623 ret = PTR_ERR(ssp->clk);
624 goto out_mmc_free;
625 }
626 ret = clk_prepare_enable(ssp->clk);
627 if (ret)
628 goto out_mmc_free;
629
630 ret = mxs_mmc_reset(host);
631 if (ret) {
632 dev_err(&pdev->dev, "Failed to reset mmc: %d\n", ret);
633 goto out_clk_disable;
634 }
635
636 ssp->dmach = dma_request_chan(&pdev->dev, "rx-tx");
637 if (IS_ERR(ssp->dmach)) {
638 dev_err(mmc_dev(host->mmc),
639 "%s: failed to request dma\n", __func__);
640 ret = PTR_ERR(ssp->dmach);
641 goto out_clk_disable;
642 }
643
644 /* set mmc core parameters */
645 mmc->ops = &mxs_mmc_ops;
646 mmc->caps = MMC_CAP_SD_HIGHSPEED | MMC_CAP_MMC_HIGHSPEED |
647 MMC_CAP_SDIO_IRQ | MMC_CAP_NEEDS_POLL | MMC_CAP_CMD23;
648
649 host->broken_cd = of_property_read_bool(np, "broken-cd");
650
651 mmc->f_min = 400000;
652 mmc->f_max = 288000000;
653
654 ret = mmc_of_parse(mmc);
655 if (ret)
656 goto out_free_dma;
657
658 mmc->ocr_avail = MMC_VDD_32_33 | MMC_VDD_33_34;
659
660 mmc->max_segs = 52;
661 mmc->max_blk_size = 1 << 0xf;
662 mmc->max_blk_count = (ssp_is_old(ssp)) ? 0xff : 0xffffff;
663 mmc->max_req_size = (ssp_is_old(ssp)) ? 0xffff : 0xffffffff;
664 mmc->max_seg_size = dma_get_max_seg_size(ssp->dmach->device->dev);
665
666 platform_set_drvdata(pdev, mmc);
667
668 spin_lock_init(&host->lock);
669
670 ret = devm_request_irq(&pdev->dev, irq_err, mxs_mmc_irq_handler, 0,
671 dev_name(&pdev->dev), host);
672 if (ret)
673 goto out_free_dma;
674
675 ret = mmc_add_host(mmc);
676 if (ret)
677 goto out_free_dma;
678
679 dev_info(mmc_dev(host->mmc), "initialized\n");
680
681 return 0;
682
683 out_free_dma:
684 dma_release_channel(ssp->dmach);
685 out_clk_disable:
686 clk_disable_unprepare(ssp->clk);
687 out_mmc_free:
688 mmc_free_host(mmc);
689 return ret;
690 }
691
mxs_mmc_remove(struct platform_device * pdev)692 static int mxs_mmc_remove(struct platform_device *pdev)
693 {
694 struct mmc_host *mmc = platform_get_drvdata(pdev);
695 struct mxs_mmc_host *host = mmc_priv(mmc);
696 struct mxs_ssp *ssp = &host->ssp;
697
698 mmc_remove_host(mmc);
699
700 if (ssp->dmach)
701 dma_release_channel(ssp->dmach);
702
703 clk_disable_unprepare(ssp->clk);
704
705 mmc_free_host(mmc);
706
707 return 0;
708 }
709
710 #ifdef CONFIG_PM_SLEEP
mxs_mmc_suspend(struct device * dev)711 static int mxs_mmc_suspend(struct device *dev)
712 {
713 struct mmc_host *mmc = dev_get_drvdata(dev);
714 struct mxs_mmc_host *host = mmc_priv(mmc);
715 struct mxs_ssp *ssp = &host->ssp;
716
717 clk_disable_unprepare(ssp->clk);
718 return 0;
719 }
720
mxs_mmc_resume(struct device * dev)721 static int mxs_mmc_resume(struct device *dev)
722 {
723 struct mmc_host *mmc = dev_get_drvdata(dev);
724 struct mxs_mmc_host *host = mmc_priv(mmc);
725 struct mxs_ssp *ssp = &host->ssp;
726
727 return clk_prepare_enable(ssp->clk);
728 }
729 #endif
730
731 static SIMPLE_DEV_PM_OPS(mxs_mmc_pm_ops, mxs_mmc_suspend, mxs_mmc_resume);
732
733 static struct platform_driver mxs_mmc_driver = {
734 .probe = mxs_mmc_probe,
735 .remove = mxs_mmc_remove,
736 .id_table = mxs_ssp_ids,
737 .driver = {
738 .name = DRIVER_NAME,
739 .probe_type = PROBE_PREFER_ASYNCHRONOUS,
740 .pm = &mxs_mmc_pm_ops,
741 .of_match_table = mxs_mmc_dt_ids,
742 },
743 };
744
745 module_platform_driver(mxs_mmc_driver);
746
747 MODULE_DESCRIPTION("FREESCALE MXS MMC peripheral");
748 MODULE_AUTHOR("Freescale Semiconductor");
749 MODULE_LICENSE("GPL");
750 MODULE_ALIAS("platform:" DRIVER_NAME);
751