Lines Matching +full:no +full:- +full:mmc
2 * linux/drivers/mmc/host/sdhci.c - Secure Digital Host Controller Interface driver
4 * Copyright (C) 2005-2008 Pierre Ossman, All Rights Reserved.
13 * - JMicron (hardware and technical support)
21 #include <linux/dma-mapping.h>
32 #include <linux/mmc/mmc.h>
33 #include <linux/mmc/host.h>
34 #include <linux/mmc/card.h>
35 #include <linux/mmc/sdio.h>
36 #include <linux/mmc/slot-gpio.h>
43 pr_debug("%s: " DRIVER_NAME ": " f, mmc_hostname(host->mmc), ## x)
46 pr_err("%s: " DRIVER_NAME ": " f, mmc_hostname(host->mmc), ## x)
76 SDHCI_DUMP("Wake-up: 0x%08x | Clock: 0x%08x\n", in sdhci_dumpregs()
103 if (host->flags & SDHCI_USE_ADMA) { in sdhci_dumpregs()
104 if (host->flags & SDHCI_USE_64_BIT_DMA) { in sdhci_dumpregs()
128 return cmd->data || cmd->flags & MMC_RSP_BUSY; in sdhci_data_line_cmd()
135 if ((host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) || in sdhci_set_card_detection()
136 !mmc_card_is_removable(host->mmc) || mmc_can_gpio_cd(host->mmc)) in sdhci_set_card_detection()
143 host->ier |= present ? SDHCI_INT_CARD_REMOVE : in sdhci_set_card_detection()
146 host->ier &= ~(SDHCI_INT_CARD_REMOVE | SDHCI_INT_CARD_INSERT); in sdhci_set_card_detection()
149 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_card_detection()
150 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_card_detection()
165 if (host->bus_on) in sdhci_runtime_pm_bus_on()
167 host->bus_on = true; in sdhci_runtime_pm_bus_on()
168 pm_runtime_get_noresume(host->mmc->parent); in sdhci_runtime_pm_bus_on()
173 if (!host->bus_on) in sdhci_runtime_pm_bus_off()
175 host->bus_on = false; in sdhci_runtime_pm_bus_off()
176 pm_runtime_put_noidle(host->mmc->parent); in sdhci_runtime_pm_bus_off()
186 host->clock = 0; in sdhci_reset()
187 /* Reset-all turns off SD Bus Power */ in sdhci_reset()
188 if (host->quirks2 & SDHCI_QUIRK2_CARD_ON_NEEDS_BUS_ON) in sdhci_reset()
203 mmc_hostname(host->mmc), (int)mask); in sdhci_reset()
214 if (host->quirks & SDHCI_QUIRK_NO_CARD_NO_RESET) { in sdhci_do_reset()
215 struct mmc_host *mmc = host->mmc; in sdhci_do_reset() local
217 if (!mmc->ops->get_cd(mmc)) in sdhci_do_reset()
221 host->ops->reset(host, mask); in sdhci_do_reset()
224 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_do_reset()
225 if (host->ops->enable_dma) in sdhci_do_reset()
226 host->ops->enable_dma(host); in sdhci_do_reset()
230 host->preset_enabled = false; in sdhci_do_reset()
236 host->ier = SDHCI_INT_BUS_POWER | SDHCI_INT_DATA_END_BIT | in sdhci_set_default_irqs()
242 if (host->tuning_mode == SDHCI_TUNING_MODE_2 || in sdhci_set_default_irqs()
243 host->tuning_mode == SDHCI_TUNING_MODE_3) in sdhci_set_default_irqs()
244 host->ier |= SDHCI_INT_RETUNE; in sdhci_set_default_irqs()
246 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_default_irqs()
247 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_default_irqs()
252 struct mmc_host *mmc = host->mmc; in sdhci_init() local
261 host->cqe_on = false; in sdhci_init()
265 host->clock = 0; in sdhci_init()
266 mmc->ops->set_ios(mmc, &mmc->ios); in sdhci_init()
301 spin_lock_irqsave(&host->lock, flags); in sdhci_led_control()
303 if (host->runtime_suspended) in sdhci_led_control()
311 spin_unlock_irqrestore(&host->lock, flags); in sdhci_led_control()
316 struct mmc_host *mmc = host->mmc; in sdhci_led_register() local
318 snprintf(host->led_name, sizeof(host->led_name), in sdhci_led_register()
319 "%s::", mmc_hostname(mmc)); in sdhci_led_register()
321 host->led.name = host->led_name; in sdhci_led_register()
322 host->led.brightness = LED_OFF; in sdhci_led_register()
323 host->led.default_trigger = mmc_hostname(mmc); in sdhci_led_register()
324 host->led.brightness_set = sdhci_led_control; in sdhci_led_register()
326 return led_classdev_register(mmc_dev(mmc), &host->led); in sdhci_led_register()
331 led_classdev_unregister(&host->led); in sdhci_led_unregister()
380 blksize = host->data->blksz; in sdhci_read_block_pio()
386 BUG_ON(!sg_miter_next(&host->sg_miter)); in sdhci_read_block_pio()
388 len = min(host->sg_miter.length, blksize); in sdhci_read_block_pio()
390 blksize -= len; in sdhci_read_block_pio()
391 host->sg_miter.consumed = len; in sdhci_read_block_pio()
393 buf = host->sg_miter.addr; in sdhci_read_block_pio()
405 chunk--; in sdhci_read_block_pio()
406 len--; in sdhci_read_block_pio()
410 sg_miter_stop(&host->sg_miter); in sdhci_read_block_pio()
424 blksize = host->data->blksz; in sdhci_write_block_pio()
431 BUG_ON(!sg_miter_next(&host->sg_miter)); in sdhci_write_block_pio()
433 len = min(host->sg_miter.length, blksize); in sdhci_write_block_pio()
435 blksize -= len; in sdhci_write_block_pio()
436 host->sg_miter.consumed = len; in sdhci_write_block_pio()
438 buf = host->sg_miter.addr; in sdhci_write_block_pio()
445 len--; in sdhci_write_block_pio()
455 sg_miter_stop(&host->sg_miter); in sdhci_write_block_pio()
464 if (host->blocks == 0) in sdhci_transfer_pio()
467 if (host->data->flags & MMC_DATA_READ) in sdhci_transfer_pio()
477 if ((host->quirks & SDHCI_QUIRK_BROKEN_SMALL_PIO) && in sdhci_transfer_pio()
478 (host->data->blocks == 1)) in sdhci_transfer_pio()
482 if (host->quirks & SDHCI_QUIRK_PIO_NEEDS_DELAY) in sdhci_transfer_pio()
485 if (host->data->flags & MMC_DATA_READ) in sdhci_transfer_pio()
490 host->blocks--; in sdhci_transfer_pio()
491 if (host->blocks == 0) in sdhci_transfer_pio()
507 if (data->host_cookie == COOKIE_PRE_MAPPED) in sdhci_pre_dma_transfer()
508 return data->sg_count; in sdhci_pre_dma_transfer()
511 if (host->bounce_buffer) { in sdhci_pre_dma_transfer()
512 unsigned int length = data->blksz * data->blocks; in sdhci_pre_dma_transfer()
514 if (length > host->bounce_buffer_size) { in sdhci_pre_dma_transfer()
516 mmc_hostname(host->mmc), length, in sdhci_pre_dma_transfer()
517 host->bounce_buffer_size); in sdhci_pre_dma_transfer()
518 return -EIO; in sdhci_pre_dma_transfer()
522 sg_copy_to_buffer(data->sg, data->sg_len, in sdhci_pre_dma_transfer()
523 host->bounce_buffer, in sdhci_pre_dma_transfer()
527 dma_sync_single_for_device(host->mmc->parent, in sdhci_pre_dma_transfer()
528 host->bounce_addr, in sdhci_pre_dma_transfer()
529 host->bounce_buffer_size, in sdhci_pre_dma_transfer()
535 sg_count = dma_map_sg(mmc_dev(host->mmc), in sdhci_pre_dma_transfer()
536 data->sg, data->sg_len, in sdhci_pre_dma_transfer()
541 return -ENOSPC; in sdhci_pre_dma_transfer()
543 data->sg_count = sg_count; in sdhci_pre_dma_transfer()
544 data->host_cookie = cookie; in sdhci_pre_dma_transfer()
552 return kmap_atomic(sg_page(sg)) + sg->offset; in sdhci_kmap_atomic()
566 /* 32-bit and 64-bit descriptors have these members in same position */ in sdhci_adma_write_desc()
567 dma_desc->cmd = cpu_to_le16(cmd); in sdhci_adma_write_desc()
568 dma_desc->len = cpu_to_le16(len); in sdhci_adma_write_desc()
569 dma_desc->addr_lo = cpu_to_le32((u32)addr); in sdhci_adma_write_desc()
571 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_adma_write_desc()
572 dma_desc->addr_hi = cpu_to_le32((u64)addr >> 32); in sdhci_adma_write_desc()
579 /* 32-bit and 64-bit descriptors have 'cmd' in same position */ in sdhci_adma_mark_end()
580 dma_desc->cmd |= cpu_to_le16(ADMA2_END); in sdhci_adma_mark_end()
598 host->sg_count = sg_count; in sdhci_adma_table_pre()
600 desc = host->adma_table; in sdhci_adma_table_pre()
601 align = host->align_buffer; in sdhci_adma_table_pre()
603 align_addr = host->align_addr; in sdhci_adma_table_pre()
605 for_each_sg(data->sg, sg, host->sg_count, i) { in sdhci_adma_table_pre()
611 * be 32-bit aligned. If they aren't, then we use a bounce in sdhci_adma_table_pre()
615 offset = (SDHCI_ADMA2_ALIGN - (addr & SDHCI_ADMA2_MASK)) & in sdhci_adma_table_pre()
618 if (data->flags & MMC_DATA_WRITE) { in sdhci_adma_table_pre()
633 desc += host->desc_sz; in sdhci_adma_table_pre()
636 len -= offset; in sdhci_adma_table_pre()
645 desc += host->desc_sz; in sdhci_adma_table_pre()
652 WARN_ON((desc - host->adma_table) >= host->adma_table_sz); in sdhci_adma_table_pre()
655 if (host->quirks & SDHCI_QUIRK_NO_ENDATTR_IN_NOPDESC) { in sdhci_adma_table_pre()
657 if (desc != host->adma_table) { in sdhci_adma_table_pre()
658 desc -= host->desc_sz; in sdhci_adma_table_pre()
662 /* Add a terminating entry - nop, end, valid */ in sdhci_adma_table_pre()
676 if (data->flags & MMC_DATA_READ) { in sdhci_adma_table_post()
680 for_each_sg(data->sg, sg, host->sg_count, i) in sdhci_adma_table_post()
687 dma_sync_sg_for_cpu(mmc_dev(host->mmc), data->sg, in sdhci_adma_table_post()
688 data->sg_len, DMA_FROM_DEVICE); in sdhci_adma_table_post()
690 align = host->align_buffer; in sdhci_adma_table_post()
692 for_each_sg(data->sg, sg, host->sg_count, i) { in sdhci_adma_table_post()
694 size = SDHCI_ADMA2_ALIGN - in sdhci_adma_table_post()
710 if (host->bounce_buffer) in sdhci_sdma_address()
711 return host->bounce_addr; in sdhci_sdma_address()
713 return sg_dma_address(host->data->sg); in sdhci_sdma_address()
724 target_timeout = cmd->busy_timeout * 1000; in sdhci_target_timeout()
726 target_timeout = DIV_ROUND_UP(data->timeout_ns, 1000); in sdhci_target_timeout()
727 if (host->clock && data->timeout_clks) { in sdhci_target_timeout()
731 * data->timeout_clks is in units of clock cycles. in sdhci_target_timeout()
732 * host->clock is in Hz. target_timeout is in us. in sdhci_target_timeout()
735 val = 1000000ULL * data->timeout_clks; in sdhci_target_timeout()
736 if (do_div(val, host->clock)) in sdhci_target_timeout()
748 struct mmc_data *data = cmd->data; in sdhci_calc_sw_timeout()
749 struct mmc_host *mmc = host->mmc; in sdhci_calc_sw_timeout() local
750 struct mmc_ios *ios = &mmc->ios; in sdhci_calc_sw_timeout()
751 unsigned char bus_width = 1 << ios->bus_width; in sdhci_calc_sw_timeout()
761 blksz = data->blksz; in sdhci_calc_sw_timeout()
762 freq = host->mmc->actual_clock ? : host->clock; in sdhci_calc_sw_timeout()
768 host->data_timeout = data->blocks * target_timeout + in sdhci_calc_sw_timeout()
771 host->data_timeout = target_timeout; in sdhci_calc_sw_timeout()
774 if (host->data_timeout) in sdhci_calc_sw_timeout()
775 host->data_timeout += MMC_CMD_TRANSFER_TIME; in sdhci_calc_sw_timeout()
782 struct mmc_data *data = cmd->data; in sdhci_calc_timeout()
790 * longer to time out, but that's much better than having a too-short in sdhci_calc_timeout()
793 if (host->quirks & SDHCI_QUIRK_BROKEN_TIMEOUT_VAL) in sdhci_calc_timeout()
797 if (!data && !cmd->busy_timeout) in sdhci_calc_timeout()
809 * (2) host->timeout_clk < 2^16 in sdhci_calc_timeout()
814 current_timeout = (1 << 13) * 1000 / host->timeout_clk; in sdhci_calc_timeout()
823 if (!(host->quirks2 & SDHCI_QUIRK2_DISABLE_HW_TIMEOUT)) in sdhci_calc_timeout()
825 count, cmd->opcode); in sdhci_calc_timeout()
839 if (host->flags & SDHCI_REQ_USE_DMA) in sdhci_set_transfer_irqs()
840 host->ier = (host->ier & ~pio_irqs) | dma_irqs; in sdhci_set_transfer_irqs()
842 host->ier = (host->ier & ~dma_irqs) | pio_irqs; in sdhci_set_transfer_irqs()
844 if (host->flags & (SDHCI_AUTO_CMD23 | SDHCI_AUTO_CMD12)) in sdhci_set_transfer_irqs()
845 host->ier |= SDHCI_INT_AUTO_CMD_ERR; in sdhci_set_transfer_irqs()
847 host->ier &= ~SDHCI_INT_AUTO_CMD_ERR; in sdhci_set_transfer_irqs()
849 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_transfer_irqs()
850 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_transfer_irqs()
856 host->ier |= SDHCI_INT_DATA_TIMEOUT; in sdhci_set_data_timeout_irq()
858 host->ier &= ~SDHCI_INT_DATA_TIMEOUT; in sdhci_set_data_timeout_irq()
859 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_data_timeout_irq()
860 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_data_timeout_irq()
867 if (host->ops->set_timeout) { in sdhci_set_timeout()
868 host->ops->set_timeout(host, cmd); in sdhci_set_timeout()
875 host->quirks2 & SDHCI_QUIRK2_DISABLE_HW_TIMEOUT) { in sdhci_set_timeout()
878 } else if (!(host->ier & SDHCI_INT_DATA_TIMEOUT)) { in sdhci_set_timeout()
889 struct mmc_data *data = cmd->data; in sdhci_prepare_data()
891 host->data_timeout = 0; in sdhci_prepare_data()
899 WARN_ON(host->data); in sdhci_prepare_data()
902 BUG_ON(data->blksz * data->blocks > 524288); in sdhci_prepare_data()
903 BUG_ON(data->blksz > host->mmc->max_blk_size); in sdhci_prepare_data()
904 BUG_ON(data->blocks > 65535); in sdhci_prepare_data()
906 host->data = data; in sdhci_prepare_data()
907 host->data_early = 0; in sdhci_prepare_data()
908 host->data->bytes_xfered = 0; in sdhci_prepare_data()
910 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_prepare_data()
915 host->flags |= SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
926 if (host->flags & SDHCI_USE_ADMA) { in sdhci_prepare_data()
927 if (host->quirks & SDHCI_QUIRK_32BIT_ADMA_SIZE) { in sdhci_prepare_data()
937 if (host->quirks & SDHCI_QUIRK_32BIT_DMA_SIZE) in sdhci_prepare_data()
939 if (host->quirks & SDHCI_QUIRK_32BIT_DMA_ADDR) in sdhci_prepare_data()
944 for_each_sg(data->sg, sg, data->sg_len, i) { in sdhci_prepare_data()
945 if (sg->length & length_mask) { in sdhci_prepare_data()
947 sg->length); in sdhci_prepare_data()
948 host->flags &= ~SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
951 if (sg->offset & offset_mask) { in sdhci_prepare_data()
953 host->flags &= ~SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
960 if (host->flags & SDHCI_REQ_USE_DMA) { in sdhci_prepare_data()
969 host->flags &= ~SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
970 } else if (host->flags & SDHCI_USE_ADMA) { in sdhci_prepare_data()
973 sdhci_writel(host, host->adma_addr, SDHCI_ADMA_ADDRESS); in sdhci_prepare_data()
974 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_prepare_data()
976 (u64)host->adma_addr >> 32, in sdhci_prepare_data()
990 if (host->version >= SDHCI_SPEC_200) { in sdhci_prepare_data()
993 if ((host->flags & SDHCI_REQ_USE_DMA) && in sdhci_prepare_data()
994 (host->flags & SDHCI_USE_ADMA)) { in sdhci_prepare_data()
995 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_prepare_data()
1005 if (!(host->flags & SDHCI_REQ_USE_DMA)) { in sdhci_prepare_data()
1009 if (host->data->flags & MMC_DATA_READ) in sdhci_prepare_data()
1013 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); in sdhci_prepare_data()
1014 host->blocks = data->blocks; in sdhci_prepare_data()
1020 sdhci_writew(host, SDHCI_MAKE_BLKSZ(host->sdma_boundary, data->blksz), in sdhci_prepare_data()
1022 sdhci_writew(host, data->blocks, SDHCI_BLOCK_COUNT); in sdhci_prepare_data()
1028 return !mrq->sbc && (host->flags & SDHCI_AUTO_CMD12) && in sdhci_auto_cmd12()
1029 !mrq->cap_cmd_during_tfr; in sdhci_auto_cmd12()
1036 struct mmc_data *data = cmd->data; in sdhci_set_transfer_mode()
1039 if (host->quirks2 & in sdhci_set_transfer_mode()
1042 if (cmd->opcode != MMC_SEND_TUNING_BLOCK_HS200) in sdhci_set_transfer_mode()
1045 /* clear Auto CMD settings for no data CMDs */ in sdhci_set_transfer_mode()
1053 WARN_ON(!host->data); in sdhci_set_transfer_mode()
1055 if (!(host->quirks2 & SDHCI_QUIRK2_SUPPORT_SINGLE)) in sdhci_set_transfer_mode()
1058 if (mmc_op_multi(cmd->opcode) || data->blocks > 1) { in sdhci_set_transfer_mode()
1062 * on successful completion (so no Auto-CMD12). in sdhci_set_transfer_mode()
1064 if (sdhci_auto_cmd12(host, cmd->mrq) && in sdhci_set_transfer_mode()
1065 (cmd->opcode != SD_IO_RW_EXTENDED)) in sdhci_set_transfer_mode()
1067 else if (cmd->mrq->sbc && (host->flags & SDHCI_AUTO_CMD23)) { in sdhci_set_transfer_mode()
1069 sdhci_writel(host, cmd->mrq->sbc->arg, SDHCI_ARGUMENT2); in sdhci_set_transfer_mode()
1073 if (data->flags & MMC_DATA_READ) in sdhci_set_transfer_mode()
1075 if (host->flags & SDHCI_REQ_USE_DMA) in sdhci_set_transfer_mode()
1083 return (!(host->flags & SDHCI_DEVICE_DEAD) && in sdhci_needs_reset()
1084 ((mrq->cmd && mrq->cmd->error) || in sdhci_needs_reset()
1085 (mrq->sbc && mrq->sbc->error) || in sdhci_needs_reset()
1086 (mrq->data && mrq->data->stop && mrq->data->stop->error) || in sdhci_needs_reset()
1087 (host->quirks & SDHCI_QUIRK_RESET_AFTER_REQUEST))); in sdhci_needs_reset()
1095 if (host->mrqs_done[i] == mrq) { in __sdhci_finish_mrq()
1102 if (!host->mrqs_done[i]) { in __sdhci_finish_mrq()
1103 host->mrqs_done[i] = mrq; in __sdhci_finish_mrq()
1110 tasklet_schedule(&host->finish_tasklet); in __sdhci_finish_mrq()
1115 if (host->cmd && host->cmd->mrq == mrq) in sdhci_finish_mrq()
1116 host->cmd = NULL; in sdhci_finish_mrq()
1118 if (host->data_cmd && host->data_cmd->mrq == mrq) in sdhci_finish_mrq()
1119 host->data_cmd = NULL; in sdhci_finish_mrq()
1121 if (host->data && host->data->mrq == mrq) in sdhci_finish_mrq()
1122 host->data = NULL; in sdhci_finish_mrq()
1125 host->pending_reset = true; in sdhci_finish_mrq()
1132 struct mmc_command *data_cmd = host->data_cmd; in sdhci_finish_data()
1133 struct mmc_data *data = host->data; in sdhci_finish_data()
1135 host->data = NULL; in sdhci_finish_data()
1136 host->data_cmd = NULL; in sdhci_finish_data()
1142 if (data->error) { in sdhci_finish_data()
1143 if (!host->cmd || host->cmd == data_cmd) in sdhci_finish_data()
1148 if ((host->flags & (SDHCI_REQ_USE_DMA | SDHCI_USE_ADMA)) == in sdhci_finish_data()
1159 if (data->error) in sdhci_finish_data()
1160 data->bytes_xfered = 0; in sdhci_finish_data()
1162 data->bytes_xfered = data->blksz * data->blocks; in sdhci_finish_data()
1165 * Need to send CMD12 if - in sdhci_finish_data()
1166 * a) open-ended multiblock transfer (no CMD23) in sdhci_finish_data()
1169 if (data->stop && in sdhci_finish_data()
1170 (data->error || in sdhci_finish_data()
1171 !data->mrq->sbc)) { in sdhci_finish_data()
1177 if (data->mrq->cap_cmd_during_tfr) { in sdhci_finish_data()
1178 sdhci_finish_mrq(host, data->mrq); in sdhci_finish_data()
1181 host->cmd = NULL; in sdhci_finish_data()
1182 sdhci_send_command(host, data->stop); in sdhci_finish_data()
1185 sdhci_finish_mrq(host, data->mrq); in sdhci_finish_data()
1192 if (sdhci_data_line_cmd(mrq->cmd)) in sdhci_mod_timer()
1193 mod_timer(&host->data_timer, timeout); in sdhci_mod_timer()
1195 mod_timer(&host->timer, timeout); in sdhci_mod_timer()
1200 if (sdhci_data_line_cmd(mrq->cmd)) in sdhci_del_timer()
1201 del_timer(&host->data_timer); in sdhci_del_timer()
1203 del_timer(&host->timer); in sdhci_del_timer()
1212 WARN_ON(host->cmd); in sdhci_send_command()
1214 /* Initially, a command has no error */ in sdhci_send_command()
1215 cmd->error = 0; in sdhci_send_command()
1217 if ((host->quirks2 & SDHCI_QUIRK2_STOP_WITH_TC) && in sdhci_send_command()
1218 cmd->opcode == MMC_STOP_TRANSMISSION) in sdhci_send_command()
1219 cmd->flags |= MMC_RSP_BUSY; in sdhci_send_command()
1230 if (cmd->mrq->data && (cmd == cmd->mrq->data->stop)) in sdhci_send_command()
1236 mmc_hostname(host->mmc)); in sdhci_send_command()
1238 cmd->error = -EIO; in sdhci_send_command()
1239 sdhci_finish_mrq(host, cmd->mrq); in sdhci_send_command()
1242 timeout--; in sdhci_send_command()
1246 host->cmd = cmd; in sdhci_send_command()
1248 WARN_ON(host->data_cmd); in sdhci_send_command()
1249 host->data_cmd = cmd; in sdhci_send_command()
1254 sdhci_writel(host, cmd->arg, SDHCI_ARGUMENT); in sdhci_send_command()
1258 if ((cmd->flags & MMC_RSP_136) && (cmd->flags & MMC_RSP_BUSY)) { in sdhci_send_command()
1260 mmc_hostname(host->mmc)); in sdhci_send_command()
1261 cmd->error = -EINVAL; in sdhci_send_command()
1262 sdhci_finish_mrq(host, cmd->mrq); in sdhci_send_command()
1266 if (!(cmd->flags & MMC_RSP_PRESENT)) in sdhci_send_command()
1268 else if (cmd->flags & MMC_RSP_136) in sdhci_send_command()
1270 else if (cmd->flags & MMC_RSP_BUSY) in sdhci_send_command()
1275 if (cmd->flags & MMC_RSP_CRC) in sdhci_send_command()
1277 if (cmd->flags & MMC_RSP_OPCODE) in sdhci_send_command()
1281 if (cmd->data || cmd->opcode == MMC_SEND_TUNING_BLOCK || in sdhci_send_command()
1282 cmd->opcode == MMC_SEND_TUNING_BLOCK_HS200) in sdhci_send_command()
1286 if (host->data_timeout) in sdhci_send_command()
1287 timeout += nsecs_to_jiffies(host->data_timeout); in sdhci_send_command()
1288 else if (!cmd->data && cmd->busy_timeout > 9000) in sdhci_send_command()
1289 timeout += DIV_ROUND_UP(cmd->busy_timeout, 1000) * HZ + HZ; in sdhci_send_command()
1292 sdhci_mod_timer(host, cmd->mrq, timeout); in sdhci_send_command()
1294 sdhci_writew(host, SDHCI_MAKE_CMD(cmd->opcode, flags), SDHCI_COMMAND); in sdhci_send_command()
1303 reg = SDHCI_RESPONSE + (3 - i) * 4; in sdhci_read_rsp_136()
1304 cmd->resp[i] = sdhci_readl(host, reg); in sdhci_read_rsp_136()
1307 if (host->quirks2 & SDHCI_QUIRK2_RSP_136_HAS_CRC) in sdhci_read_rsp_136()
1312 cmd->resp[i] <<= 8; in sdhci_read_rsp_136()
1314 cmd->resp[i] |= cmd->resp[i + 1] >> 24; in sdhci_read_rsp_136()
1320 struct mmc_command *cmd = host->cmd; in sdhci_finish_command()
1322 host->cmd = NULL; in sdhci_finish_command()
1324 if (cmd->flags & MMC_RSP_PRESENT) { in sdhci_finish_command()
1325 if (cmd->flags & MMC_RSP_136) { in sdhci_finish_command()
1328 cmd->resp[0] = sdhci_readl(host, SDHCI_RESPONSE); in sdhci_finish_command()
1332 if (cmd->mrq->cap_cmd_during_tfr && cmd == cmd->mrq->cmd) in sdhci_finish_command()
1333 mmc_command_done(host->mmc, cmd->mrq); in sdhci_finish_command()
1345 if (cmd->flags & MMC_RSP_BUSY) { in sdhci_finish_command()
1346 if (cmd->data) { in sdhci_finish_command()
1348 } else if (!(host->quirks & SDHCI_QUIRK_NO_BUSY_IRQ) && in sdhci_finish_command()
1349 cmd == host->data_cmd) { in sdhci_finish_command()
1356 if (cmd == cmd->mrq->sbc) { in sdhci_finish_command()
1357 sdhci_send_command(host, cmd->mrq->cmd); in sdhci_finish_command()
1361 if (host->data && host->data_early) in sdhci_finish_command()
1364 if (!cmd->data) in sdhci_finish_command()
1365 sdhci_finish_mrq(host, cmd->mrq); in sdhci_finish_command()
1373 switch (host->timing) { in sdhci_get_preset_value()
1395 pr_warn("%s: Invalid UHS-I mode selected\n", in sdhci_get_preset_value()
1396 mmc_hostname(host->mmc)); in sdhci_get_preset_value()
1411 if (host->version >= SDHCI_SPEC_300) { in sdhci_calc_clk()
1412 if (host->preset_enabled) { in sdhci_calc_clk()
1419 if (host->clk_mul && in sdhci_calc_clk()
1423 clk_mul = host->clk_mul; in sdhci_calc_clk()
1434 if (host->clk_mul) { in sdhci_calc_clk()
1436 if ((host->max_clk * host->clk_mul / div) in sdhci_calc_clk()
1440 if ((host->max_clk * host->clk_mul / div) <= clock) { in sdhci_calc_clk()
1447 clk_mul = host->clk_mul; in sdhci_calc_clk()
1448 div--; in sdhci_calc_clk()
1458 if (!host->clk_mul || switch_base_clk) { in sdhci_calc_clk()
1460 if (host->max_clk <= clock) in sdhci_calc_clk()
1465 if ((host->max_clk / div) <= clock) in sdhci_calc_clk()
1471 if ((host->quirks2 & SDHCI_QUIRK2_CLOCK_DIV_ZERO_BROKEN) in sdhci_calc_clk()
1472 && !div && host->max_clk <= 25000000) in sdhci_calc_clk()
1478 if ((host->max_clk / div) <= clock) in sdhci_calc_clk()
1487 *actual_clock = (host->max_clk * clk_mul) / real_div; in sdhci_calc_clk()
1513 mmc_hostname(host->mmc)); in sdhci_enable_clk()
1529 host->mmc->actual_clock = 0; in sdhci_set_clock()
1536 clk = sdhci_calc_clk(host, clock, &host->mmc->actual_clock); in sdhci_set_clock()
1544 struct mmc_host *mmc = host->mmc; in sdhci_set_power_reg() local
1546 mmc_regulator_set_ocr(mmc, mmc->supply.vmmc, vdd); in sdhci_set_power_reg()
1581 mmc_hostname(host->mmc), vdd); in sdhci_set_power_noreg()
1586 if (host->pwr == pwr) in sdhci_set_power_noreg()
1589 host->pwr = pwr; in sdhci_set_power_noreg()
1593 if (host->quirks2 & SDHCI_QUIRK2_CARD_ON_NEEDS_BUS_ON) in sdhci_set_power_noreg()
1600 if (!(host->quirks & SDHCI_QUIRK_SINGLE_POWER_WRITE)) in sdhci_set_power_noreg()
1608 if (host->quirks & SDHCI_QUIRK_NO_SIMULT_VDD_AND_POWER) in sdhci_set_power_noreg()
1615 if (host->quirks2 & SDHCI_QUIRK2_CARD_ON_NEEDS_BUS_ON) in sdhci_set_power_noreg()
1622 if (host->quirks & SDHCI_QUIRK_DELAY_AFTER_POWER) in sdhci_set_power_noreg()
1631 if (IS_ERR(host->mmc->supply.vmmc)) in sdhci_set_power()
1640 * MMC callbacks *
1644 static void sdhci_request(struct mmc_host *mmc, struct mmc_request *mrq) in sdhci_request() argument
1650 host = mmc_priv(mmc); in sdhci_request()
1653 present = mmc->ops->get_cd(mmc); in sdhci_request()
1655 spin_lock_irqsave(&host->lock, flags); in sdhci_request()
1660 * Ensure we don't send the STOP for non-SET_BLOCK_COUNTED in sdhci_request()
1661 * requests if Auto-CMD12 is enabled. in sdhci_request()
1664 if (mrq->stop) { in sdhci_request()
1665 mrq->data->stop = NULL; in sdhci_request()
1666 mrq->stop = NULL; in sdhci_request()
1670 if (!present || host->flags & SDHCI_DEVICE_DEAD) { in sdhci_request()
1671 mrq->cmd->error = -ENOMEDIUM; in sdhci_request()
1674 if (mrq->sbc && !(host->flags & SDHCI_AUTO_CMD23)) in sdhci_request()
1675 sdhci_send_command(host, mrq->sbc); in sdhci_request()
1677 sdhci_send_command(host, mrq->cmd); in sdhci_request()
1681 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request()
1693 if (host->mmc->caps & MMC_CAP_8_BIT_DATA) in sdhci_set_bus_width()
1724 ctrl_2 |= SDHCI_CTRL_HS400; /* Non-standard */ in sdhci_set_uhs_signaling()
1729 void sdhci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) in sdhci_set_ios() argument
1731 struct sdhci_host *host = mmc_priv(mmc); in sdhci_set_ios()
1734 if (ios->power_mode == MMC_POWER_UNDEFINED) in sdhci_set_ios()
1737 if (host->flags & SDHCI_DEVICE_DEAD) { in sdhci_set_ios()
1738 if (!IS_ERR(mmc->supply.vmmc) && in sdhci_set_ios()
1739 ios->power_mode == MMC_POWER_OFF) in sdhci_set_ios()
1740 mmc_regulator_set_ocr(mmc, mmc->supply.vmmc, 0); in sdhci_set_ios()
1748 if (ios->power_mode == MMC_POWER_OFF) { in sdhci_set_ios()
1753 if (host->version >= SDHCI_SPEC_300 && in sdhci_set_ios()
1754 (ios->power_mode == MMC_POWER_UP) && in sdhci_set_ios()
1755 !(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN)) in sdhci_set_ios()
1758 if (!ios->clock || ios->clock != host->clock) { in sdhci_set_ios()
1759 host->ops->set_clock(host, ios->clock); in sdhci_set_ios()
1760 host->clock = ios->clock; in sdhci_set_ios()
1762 if (host->quirks & SDHCI_QUIRK_DATA_TIMEOUT_USES_SDCLK && in sdhci_set_ios()
1763 host->clock) { in sdhci_set_ios()
1764 host->timeout_clk = host->mmc->actual_clock ? in sdhci_set_ios()
1765 host->mmc->actual_clock / 1000 : in sdhci_set_ios()
1766 host->clock / 1000; in sdhci_set_ios()
1767 host->mmc->max_busy_timeout = in sdhci_set_ios()
1768 host->ops->get_max_timeout_count ? in sdhci_set_ios()
1769 host->ops->get_max_timeout_count(host) : in sdhci_set_ios()
1771 host->mmc->max_busy_timeout /= host->timeout_clk; in sdhci_set_ios()
1775 if (host->ops->set_power) in sdhci_set_ios()
1776 host->ops->set_power(host, ios->power_mode, ios->vdd); in sdhci_set_ios()
1778 sdhci_set_power(host, ios->power_mode, ios->vdd); in sdhci_set_ios()
1780 if (host->ops->platform_send_init_74_clocks) in sdhci_set_ios()
1781 host->ops->platform_send_init_74_clocks(host, ios->power_mode); in sdhci_set_ios()
1783 host->ops->set_bus_width(host, ios->bus_width); in sdhci_set_ios()
1787 if (!(host->quirks & SDHCI_QUIRK_NO_HISPD_BIT)) { in sdhci_set_ios()
1788 if (ios->timing == MMC_TIMING_SD_HS || in sdhci_set_ios()
1789 ios->timing == MMC_TIMING_MMC_HS || in sdhci_set_ios()
1790 ios->timing == MMC_TIMING_MMC_HS400 || in sdhci_set_ios()
1791 ios->timing == MMC_TIMING_MMC_HS200 || in sdhci_set_ios()
1792 ios->timing == MMC_TIMING_MMC_DDR52 || in sdhci_set_ios()
1793 ios->timing == MMC_TIMING_UHS_SDR50 || in sdhci_set_ios()
1794 ios->timing == MMC_TIMING_UHS_SDR104 || in sdhci_set_ios()
1795 ios->timing == MMC_TIMING_UHS_DDR50 || in sdhci_set_ios()
1796 ios->timing == MMC_TIMING_UHS_SDR25) in sdhci_set_ios()
1802 if (host->version >= SDHCI_SPEC_300) { in sdhci_set_ios()
1805 if (!host->preset_enabled) { in sdhci_set_ios()
1813 if (ios->drv_type == MMC_SET_DRIVER_TYPE_A) in sdhci_set_ios()
1815 else if (ios->drv_type == MMC_SET_DRIVER_TYPE_B) in sdhci_set_ios()
1817 else if (ios->drv_type == MMC_SET_DRIVER_TYPE_C) in sdhci_set_ios()
1819 else if (ios->drv_type == MMC_SET_DRIVER_TYPE_D) in sdhci_set_ios()
1823 mmc_hostname(mmc)); in sdhci_set_ios()
1843 /* Re-enable SD Clock */ in sdhci_set_ios()
1844 host->ops->set_clock(host, host->clock); in sdhci_set_ios()
1852 host->ops->set_uhs_signaling(host, ios->timing); in sdhci_set_ios()
1853 host->timing = ios->timing; in sdhci_set_ios()
1855 if (!(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN) && in sdhci_set_ios()
1856 ((ios->timing == MMC_TIMING_UHS_SDR12) || in sdhci_set_ios()
1857 (ios->timing == MMC_TIMING_UHS_SDR25) || in sdhci_set_ios()
1858 (ios->timing == MMC_TIMING_UHS_SDR50) || in sdhci_set_ios()
1859 (ios->timing == MMC_TIMING_UHS_SDR104) || in sdhci_set_ios()
1860 (ios->timing == MMC_TIMING_UHS_DDR50) || in sdhci_set_ios()
1861 (ios->timing == MMC_TIMING_MMC_DDR52))) { in sdhci_set_ios()
1866 ios->drv_type = (preset & SDHCI_PRESET_DRV_MASK) in sdhci_set_ios()
1870 /* Re-enable SD Clock */ in sdhci_set_ios()
1871 host->ops->set_clock(host, host->clock); in sdhci_set_ios()
1880 if (host->quirks & SDHCI_QUIRK_RESET_CMD_DATA_ON_IOS) in sdhci_set_ios()
1887 static int sdhci_get_cd(struct mmc_host *mmc) in sdhci_get_cd() argument
1889 struct sdhci_host *host = mmc_priv(mmc); in sdhci_get_cd()
1890 int gpio_cd = mmc_gpio_get_cd(mmc); in sdhci_get_cd()
1892 if (host->flags & SDHCI_DEVICE_DEAD) in sdhci_get_cd()
1896 if (!mmc_card_is_removable(host->mmc)) in sdhci_get_cd()
1907 if (host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) in sdhci_get_cd()
1919 spin_lock_irqsave(&host->lock, flags); in sdhci_check_ro()
1921 if (host->flags & SDHCI_DEVICE_DEAD) in sdhci_check_ro()
1923 else if (host->ops->get_ro) in sdhci_check_ro()
1924 is_readonly = host->ops->get_ro(host); in sdhci_check_ro()
1929 spin_unlock_irqrestore(&host->lock, flags); in sdhci_check_ro()
1931 /* This quirk needs to be replaced by a callback-function later */ in sdhci_check_ro()
1932 return host->quirks & SDHCI_QUIRK_INVERTED_WRITE_PROTECT ? in sdhci_check_ro()
1938 static int sdhci_get_ro(struct mmc_host *mmc) in sdhci_get_ro() argument
1940 struct sdhci_host *host = mmc_priv(mmc); in sdhci_get_ro()
1943 if (!(host->quirks & SDHCI_QUIRK_UNSTABLE_RO_DETECT)) in sdhci_get_ro()
1957 static void sdhci_hw_reset(struct mmc_host *mmc) in sdhci_hw_reset() argument
1959 struct sdhci_host *host = mmc_priv(mmc); in sdhci_hw_reset()
1961 if (host->ops && host->ops->hw_reset) in sdhci_hw_reset()
1962 host->ops->hw_reset(host); in sdhci_hw_reset()
1967 if (!(host->flags & SDHCI_DEVICE_DEAD)) { in sdhci_enable_sdio_irq_nolock()
1969 host->ier |= SDHCI_INT_CARD_INT; in sdhci_enable_sdio_irq_nolock()
1971 host->ier &= ~SDHCI_INT_CARD_INT; in sdhci_enable_sdio_irq_nolock()
1973 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_enable_sdio_irq_nolock()
1974 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_enable_sdio_irq_nolock()
1979 void sdhci_enable_sdio_irq(struct mmc_host *mmc, int enable) in sdhci_enable_sdio_irq() argument
1981 struct sdhci_host *host = mmc_priv(mmc); in sdhci_enable_sdio_irq()
1985 pm_runtime_get_noresume(host->mmc->parent); in sdhci_enable_sdio_irq()
1987 spin_lock_irqsave(&host->lock, flags); in sdhci_enable_sdio_irq()
1989 host->flags |= SDHCI_SDIO_IRQ_ENABLED; in sdhci_enable_sdio_irq()
1991 host->flags &= ~SDHCI_SDIO_IRQ_ENABLED; in sdhci_enable_sdio_irq()
1994 spin_unlock_irqrestore(&host->lock, flags); in sdhci_enable_sdio_irq()
1997 pm_runtime_put_noidle(host->mmc->parent); in sdhci_enable_sdio_irq()
2001 int sdhci_start_signal_voltage_switch(struct mmc_host *mmc, in sdhci_start_signal_voltage_switch() argument
2004 struct sdhci_host *host = mmc_priv(mmc); in sdhci_start_signal_voltage_switch()
2012 if (host->version < SDHCI_SPEC_300) in sdhci_start_signal_voltage_switch()
2017 switch (ios->signal_voltage) { in sdhci_start_signal_voltage_switch()
2019 if (!(host->flags & SDHCI_SIGNALING_330)) in sdhci_start_signal_voltage_switch()
2020 return -EINVAL; in sdhci_start_signal_voltage_switch()
2025 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_start_signal_voltage_switch()
2026 ret = mmc_regulator_set_vqmmc(mmc, ios); in sdhci_start_signal_voltage_switch()
2029 mmc_hostname(mmc)); in sdhci_start_signal_voltage_switch()
2030 return -EIO; in sdhci_start_signal_voltage_switch()
2042 mmc_hostname(mmc)); in sdhci_start_signal_voltage_switch()
2044 return -EAGAIN; in sdhci_start_signal_voltage_switch()
2046 if (!(host->flags & SDHCI_SIGNALING_180)) in sdhci_start_signal_voltage_switch()
2047 return -EINVAL; in sdhci_start_signal_voltage_switch()
2048 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_start_signal_voltage_switch()
2049 ret = mmc_regulator_set_vqmmc(mmc, ios); in sdhci_start_signal_voltage_switch()
2052 mmc_hostname(mmc)); in sdhci_start_signal_voltage_switch()
2053 return -EIO; in sdhci_start_signal_voltage_switch()
2065 if (host->ops->voltage_switch) in sdhci_start_signal_voltage_switch()
2066 host->ops->voltage_switch(host); in sdhci_start_signal_voltage_switch()
2074 mmc_hostname(mmc)); in sdhci_start_signal_voltage_switch()
2076 return -EAGAIN; in sdhci_start_signal_voltage_switch()
2078 if (!(host->flags & SDHCI_SIGNALING_120)) in sdhci_start_signal_voltage_switch()
2079 return -EINVAL; in sdhci_start_signal_voltage_switch()
2080 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_start_signal_voltage_switch()
2081 ret = mmc_regulator_set_vqmmc(mmc, ios); in sdhci_start_signal_voltage_switch()
2084 mmc_hostname(mmc)); in sdhci_start_signal_voltage_switch()
2085 return -EIO; in sdhci_start_signal_voltage_switch()
2090 /* No signal voltage switch required */ in sdhci_start_signal_voltage_switch()
2096 static int sdhci_card_busy(struct mmc_host *mmc) in sdhci_card_busy() argument
2098 struct sdhci_host *host = mmc_priv(mmc); in sdhci_card_busy()
2107 static int sdhci_prepare_hs400_tuning(struct mmc_host *mmc, struct mmc_ios *ios) in sdhci_prepare_hs400_tuning() argument
2109 struct sdhci_host *host = mmc_priv(mmc); in sdhci_prepare_hs400_tuning()
2112 spin_lock_irqsave(&host->lock, flags); in sdhci_prepare_hs400_tuning()
2113 host->flags |= SDHCI_HS400_TUNING; in sdhci_prepare_hs400_tuning()
2114 spin_unlock_irqrestore(&host->lock, flags); in sdhci_prepare_hs400_tuning()
2125 if (host->quirks2 & SDHCI_QUIRK2_TUNING_WORK_AROUND) in sdhci_start_tuning()
2146 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_end_tuning()
2147 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_end_tuning()
2171 mmc_abort_tuning(host->mmc, opcode); in sdhci_abort_tuning()
2177 * automatically) so mmc_send_tuning() will return -EIO. Also the tuning command
2178 * interrupt setup is different to other commands and there is no timeout
2183 struct mmc_host *mmc = host->mmc; in sdhci_send_tuning() local
2187 u32 b = host->sdma_boundary; in sdhci_send_tuning()
2189 spin_lock_irqsave(&host->lock, flags); in sdhci_send_tuning()
2202 mmc->ios.bus_width == MMC_BUS_WIDTH_8) in sdhci_send_tuning()
2217 host->cmd = NULL; in sdhci_send_tuning()
2221 host->tuning_done = 0; in sdhci_send_tuning()
2224 spin_unlock_irqrestore(&host->lock, flags); in sdhci_send_tuning()
2227 wait_event_timeout(host->buf_ready_int, (host->tuning_done == 1), in sdhci_send_tuning()
2246 if (!host->tuning_done) { in __sdhci_execute_tuning()
2248 mmc_hostname(host->mmc)); in __sdhci_execute_tuning()
2261 if (host->tuning_delay > 0) in __sdhci_execute_tuning()
2262 mdelay(host->tuning_delay); in __sdhci_execute_tuning()
2266 mmc_hostname(host->mmc)); in __sdhci_execute_tuning()
2270 int sdhci_execute_tuning(struct mmc_host *mmc, u32 opcode) in sdhci_execute_tuning() argument
2272 struct sdhci_host *host = mmc_priv(mmc); in sdhci_execute_tuning()
2277 hs400_tuning = host->flags & SDHCI_HS400_TUNING; in sdhci_execute_tuning()
2279 if (host->tuning_mode == SDHCI_TUNING_MODE_1) in sdhci_execute_tuning()
2280 tuning_count = host->tuning_count; in sdhci_execute_tuning()
2289 switch (host->timing) { in sdhci_execute_tuning()
2292 err = -EINVAL; in sdhci_execute_tuning()
2297 * Periodic re-tuning for HS400 is not expected to be needed, so in sdhci_execute_tuning()
2309 if (host->flags & SDHCI_SDR50_NEEDS_TUNING) in sdhci_execute_tuning()
2317 if (host->ops->platform_execute_tuning) { in sdhci_execute_tuning()
2318 err = host->ops->platform_execute_tuning(host, opcode); in sdhci_execute_tuning()
2322 host->mmc->retune_period = tuning_count; in sdhci_execute_tuning()
2324 if (host->tuning_delay < 0) in sdhci_execute_tuning()
2325 host->tuning_delay = opcode == MMC_SEND_TUNING_BLOCK; in sdhci_execute_tuning()
2333 host->flags &= ~SDHCI_HS400_TUNING; in sdhci_execute_tuning()
2342 if (host->version < SDHCI_SPEC_300) in sdhci_enable_preset_value()
2349 if (host->preset_enabled != enable) { in sdhci_enable_preset_value()
2360 host->flags |= SDHCI_PV_ENABLED; in sdhci_enable_preset_value()
2362 host->flags &= ~SDHCI_PV_ENABLED; in sdhci_enable_preset_value()
2364 host->preset_enabled = enable; in sdhci_enable_preset_value()
2368 static void sdhci_post_req(struct mmc_host *mmc, struct mmc_request *mrq, in sdhci_post_req() argument
2371 struct sdhci_host *host = mmc_priv(mmc); in sdhci_post_req()
2372 struct mmc_data *data = mrq->data; in sdhci_post_req()
2374 if (data->host_cookie != COOKIE_UNMAPPED) in sdhci_post_req()
2375 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in sdhci_post_req()
2378 data->host_cookie = COOKIE_UNMAPPED; in sdhci_post_req()
2381 static void sdhci_pre_req(struct mmc_host *mmc, struct mmc_request *mrq) in sdhci_pre_req() argument
2383 struct sdhci_host *host = mmc_priv(mmc); in sdhci_pre_req()
2385 mrq->data->host_cookie = COOKIE_UNMAPPED; in sdhci_pre_req()
2388 * No pre-mapping in the pre hook if we're using the bounce buffer, in sdhci_pre_req()
2392 if (host->flags & SDHCI_REQ_USE_DMA && !host->bounce_buffer) in sdhci_pre_req()
2393 sdhci_pre_dma_transfer(host, mrq->data, COOKIE_PRE_MAPPED); in sdhci_pre_req()
2398 return host->cmd || host->data_cmd; in sdhci_has_requests()
2403 if (host->data_cmd) { in sdhci_error_out_mrqs()
2404 host->data_cmd->error = err; in sdhci_error_out_mrqs()
2405 sdhci_finish_mrq(host, host->data_cmd->mrq); in sdhci_error_out_mrqs()
2408 if (host->cmd) { in sdhci_error_out_mrqs()
2409 host->cmd->error = err; in sdhci_error_out_mrqs()
2410 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_error_out_mrqs()
2414 static void sdhci_card_event(struct mmc_host *mmc) in sdhci_card_event() argument
2416 struct sdhci_host *host = mmc_priv(mmc); in sdhci_card_event()
2421 if (host->ops->card_event) in sdhci_card_event()
2422 host->ops->card_event(host); in sdhci_card_event()
2424 present = mmc->ops->get_cd(mmc); in sdhci_card_event()
2426 spin_lock_irqsave(&host->lock, flags); in sdhci_card_event()
2431 mmc_hostname(host->mmc)); in sdhci_card_event()
2433 mmc_hostname(host->mmc)); in sdhci_card_event()
2438 sdhci_error_out_mrqs(host, -ENOMEDIUM); in sdhci_card_event()
2441 spin_unlock_irqrestore(&host->lock, flags); in sdhci_card_event()
2472 spin_lock_irqsave(&host->lock, flags); in sdhci_request_done()
2475 mrq = host->mrqs_done[i]; in sdhci_request_done()
2481 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
2492 if (host->flags & SDHCI_REQ_USE_DMA) { in sdhci_request_done()
2493 struct mmc_data *data = mrq->data; in sdhci_request_done()
2495 if (data && data->host_cookie == COOKIE_MAPPED) { in sdhci_request_done()
2496 if (host->bounce_buffer) { in sdhci_request_done()
2502 unsigned int length = data->bytes_xfered; in sdhci_request_done()
2504 if (length > host->bounce_buffer_size) { in sdhci_request_done()
2506 mmc_hostname(host->mmc), in sdhci_request_done()
2507 host->bounce_buffer_size, in sdhci_request_done()
2508 data->bytes_xfered); in sdhci_request_done()
2510 length = host->bounce_buffer_size; in sdhci_request_done()
2513 host->mmc->parent, in sdhci_request_done()
2514 host->bounce_addr, in sdhci_request_done()
2515 host->bounce_buffer_size, in sdhci_request_done()
2517 sg_copy_from_buffer(data->sg, in sdhci_request_done()
2518 data->sg_len, in sdhci_request_done()
2519 host->bounce_buffer, in sdhci_request_done()
2522 /* No copying, just switch ownership */ in sdhci_request_done()
2524 host->mmc->parent, in sdhci_request_done()
2525 host->bounce_addr, in sdhci_request_done()
2526 host->bounce_buffer_size, in sdhci_request_done()
2531 dma_unmap_sg(mmc_dev(host->mmc), data->sg, in sdhci_request_done()
2532 data->sg_len, in sdhci_request_done()
2535 data->host_cookie = COOKIE_UNMAPPED; in sdhci_request_done()
2547 * also be in mrqs_done, otherwise host->cmd and host->data_cmd in sdhci_request_done()
2550 if (host->cmd || host->data_cmd) { in sdhci_request_done()
2551 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
2556 if (host->quirks & SDHCI_QUIRK_CLOCK_BEFORE_RESET) in sdhci_request_done()
2558 host->ops->set_clock(host, host->clock); in sdhci_request_done()
2565 host->pending_reset = false; in sdhci_request_done()
2571 host->mrqs_done[i] = NULL; in sdhci_request_done()
2574 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
2576 mmc_request_done(host->mmc, mrq); in sdhci_request_done()
2596 spin_lock_irqsave(&host->lock, flags); in sdhci_timeout_timer()
2598 if (host->cmd && !sdhci_data_line_cmd(host->cmd)) { in sdhci_timeout_timer()
2600 mmc_hostname(host->mmc)); in sdhci_timeout_timer()
2603 host->cmd->error = -ETIMEDOUT; in sdhci_timeout_timer()
2604 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_timeout_timer()
2608 spin_unlock_irqrestore(&host->lock, flags); in sdhci_timeout_timer()
2618 spin_lock_irqsave(&host->lock, flags); in sdhci_timeout_data_timer()
2620 if (host->data || host->data_cmd || in sdhci_timeout_data_timer()
2621 (host->cmd && sdhci_data_line_cmd(host->cmd))) { in sdhci_timeout_data_timer()
2623 mmc_hostname(host->mmc)); in sdhci_timeout_data_timer()
2626 if (host->data) { in sdhci_timeout_data_timer()
2627 host->data->error = -ETIMEDOUT; in sdhci_timeout_data_timer()
2629 } else if (host->data_cmd) { in sdhci_timeout_data_timer()
2630 host->data_cmd->error = -ETIMEDOUT; in sdhci_timeout_data_timer()
2631 sdhci_finish_mrq(host, host->data_cmd->mrq); in sdhci_timeout_data_timer()
2633 host->cmd->error = -ETIMEDOUT; in sdhci_timeout_data_timer()
2634 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_timeout_data_timer()
2639 spin_unlock_irqrestore(&host->lock, flags); in sdhci_timeout_data_timer()
2650 /* Handle auto-CMD12 error */ in sdhci_cmd_irq()
2651 if (intmask & SDHCI_INT_AUTO_CMD_ERR && host->data_cmd) { in sdhci_cmd_irq()
2652 struct mmc_request *mrq = host->data_cmd->mrq; in sdhci_cmd_irq()
2658 /* Treat auto-CMD12 error the same as data error */ in sdhci_cmd_irq()
2659 if (!mrq->sbc && (host->flags & SDHCI_AUTO_CMD12)) { in sdhci_cmd_irq()
2665 if (!host->cmd) { in sdhci_cmd_irq()
2671 if (host->pending_reset) in sdhci_cmd_irq()
2673 pr_err("%s: Got command interrupt 0x%08x even though no command operation was in progress.\n", in sdhci_cmd_irq()
2674 mmc_hostname(host->mmc), (unsigned)intmask); in sdhci_cmd_irq()
2682 host->cmd->error = -ETIMEDOUT; in sdhci_cmd_irq()
2684 host->cmd->error = -EILSEQ; in sdhci_cmd_irq()
2687 if (host->cmd->data && in sdhci_cmd_irq()
2690 host->cmd = NULL; in sdhci_cmd_irq()
2695 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_cmd_irq()
2699 /* Handle auto-CMD23 error */ in sdhci_cmd_irq()
2701 struct mmc_request *mrq = host->cmd->mrq; in sdhci_cmd_irq()
2704 -ETIMEDOUT : in sdhci_cmd_irq()
2705 -EILSEQ; in sdhci_cmd_irq()
2707 if (mrq->sbc && (host->flags & SDHCI_AUTO_CMD23)) { in sdhci_cmd_irq()
2708 mrq->sbc->error = err; in sdhci_cmd_irq()
2720 void *desc = host->adma_table; in sdhci_adma_show_error()
2721 dma_addr_t dma = host->adma_addr; in sdhci_adma_show_error()
2728 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_adma_show_error()
2731 le32_to_cpu(dma_desc->addr_hi), in sdhci_adma_show_error()
2732 le32_to_cpu(dma_desc->addr_lo), in sdhci_adma_show_error()
2733 le16_to_cpu(dma_desc->len), in sdhci_adma_show_error()
2734 le16_to_cpu(dma_desc->cmd)); in sdhci_adma_show_error()
2738 le32_to_cpu(dma_desc->addr_lo), in sdhci_adma_show_error()
2739 le16_to_cpu(dma_desc->len), in sdhci_adma_show_error()
2740 le16_to_cpu(dma_desc->cmd)); in sdhci_adma_show_error()
2742 desc += host->desc_sz; in sdhci_adma_show_error()
2743 dma += host->desc_sz; in sdhci_adma_show_error()
2745 if (dma_desc->cmd & cpu_to_le16(ADMA2_END)) in sdhci_adma_show_error()
2759 host->tuning_done = 1; in sdhci_data_irq()
2760 wake_up(&host->buf_ready_int); in sdhci_data_irq()
2765 if (!host->data) { in sdhci_data_irq()
2766 struct mmc_command *data_cmd = host->data_cmd; in sdhci_data_irq()
2773 if (data_cmd && (data_cmd->flags & MMC_RSP_BUSY)) { in sdhci_data_irq()
2775 host->data_cmd = NULL; in sdhci_data_irq()
2776 data_cmd->error = -ETIMEDOUT; in sdhci_data_irq()
2777 sdhci_finish_mrq(host, data_cmd->mrq); in sdhci_data_irq()
2781 host->data_cmd = NULL; in sdhci_data_irq()
2783 * Some cards handle busy-end interrupt in sdhci_data_irq()
2787 if (host->cmd == data_cmd) in sdhci_data_irq()
2790 sdhci_finish_mrq(host, data_cmd->mrq); in sdhci_data_irq()
2800 if (host->pending_reset) in sdhci_data_irq()
2803 pr_err("%s: Got data interrupt 0x%08x even though no data operation was in progress.\n", in sdhci_data_irq()
2804 mmc_hostname(host->mmc), (unsigned)intmask); in sdhci_data_irq()
2811 host->data->error = -ETIMEDOUT; in sdhci_data_irq()
2813 host->data->error = -EILSEQ; in sdhci_data_irq()
2817 host->data->error = -EILSEQ; in sdhci_data_irq()
2819 pr_err("%s: ADMA error: 0x%08x\n", mmc_hostname(host->mmc), in sdhci_data_irq()
2822 host->data->error = -EIO; in sdhci_data_irq()
2823 if (host->ops->adma_workaround) in sdhci_data_irq()
2824 host->ops->adma_workaround(host, intmask); in sdhci_data_irq()
2827 if (host->data->error) in sdhci_data_irq()
2846 dmanow = dmastart + host->data->bytes_xfered; in sdhci_data_irq()
2851 ~(SDHCI_DEFAULT_BOUNDARY_SIZE - 1)) + in sdhci_data_irq()
2853 host->data->bytes_xfered = dmanow - dmastart; in sdhci_data_irq()
2855 dmastart, host->data->bytes_xfered, dmanow); in sdhci_data_irq()
2860 if (host->cmd == host->data_cmd) { in sdhci_data_irq()
2866 host->data_early = 1; in sdhci_data_irq()
2881 spin_lock(&host->lock); in sdhci_irq()
2883 if (host->runtime_suspended && !sdhci_sdio_irq_enabled(host)) { in sdhci_irq()
2884 spin_unlock(&host->lock); in sdhci_irq()
2897 if (host->ops->irq) { in sdhci_irq()
2898 intmask = host->ops->irq(host, intmask); in sdhci_irq()
2923 host->ier &= ~(SDHCI_INT_CARD_INSERT | in sdhci_irq()
2925 host->ier |= present ? SDHCI_INT_CARD_REMOVE : in sdhci_irq()
2927 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_irq()
2928 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_irq()
2933 host->thread_isr |= intmask & (SDHCI_INT_CARD_INSERT | in sdhci_irq()
2946 mmc_hostname(host->mmc)); in sdhci_irq()
2949 mmc_retune_needed(host->mmc); in sdhci_irq()
2952 (host->ier & SDHCI_INT_CARD_INT)) { in sdhci_irq()
2954 host->thread_isr |= SDHCI_INT_CARD_INT; in sdhci_irq()
2972 } while (intmask && --max_loops); in sdhci_irq()
2974 spin_unlock(&host->lock); in sdhci_irq()
2978 mmc_hostname(host->mmc), unexpected); in sdhci_irq()
2991 spin_lock_irqsave(&host->lock, flags); in sdhci_thread_irq()
2992 isr = host->thread_isr; in sdhci_thread_irq()
2993 host->thread_isr = 0; in sdhci_thread_irq()
2994 spin_unlock_irqrestore(&host->lock, flags); in sdhci_thread_irq()
2997 struct mmc_host *mmc = host->mmc; in sdhci_thread_irq() local
2999 mmc->ops->card_event(mmc); in sdhci_thread_irq()
3000 mmc_detect_change(mmc, msecs_to_jiffies(200)); in sdhci_thread_irq()
3004 sdio_run_irqs(host->mmc); in sdhci_thread_irq()
3006 spin_lock_irqsave(&host->lock, flags); in sdhci_thread_irq()
3007 if (host->flags & SDHCI_SDIO_IRQ_ENABLED) in sdhci_thread_irq()
3009 spin_unlock_irqrestore(&host->lock, flags); in sdhci_thread_irq()
3025 return mmc_card_is_removable(host->mmc) && in sdhci_cd_irq_can_wakeup()
3026 !(host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) && in sdhci_cd_irq_can_wakeup()
3027 !mmc_can_gpio_cd(host->mmc); in sdhci_cd_irq_can_wakeup()
3032 * the Interrupt Status Enable register too. See 'Table 1-6: Wakeup Signal
3051 if (mmc_card_wake_sdio_irq(host->mmc)) { in sdhci_enable_irq_wakeups()
3066 host->irq_wake_enabled = !enable_irq_wake(host->irq); in sdhci_enable_irq_wakeups()
3068 return host->irq_wake_enabled; in sdhci_enable_irq_wakeups()
3081 disable_irq_wake(host->irq); in sdhci_disable_irq_wakeups()
3083 host->irq_wake_enabled = false; in sdhci_disable_irq_wakeups()
3090 mmc_retune_timer_stop(host->mmc); in sdhci_suspend_host()
3092 if (!device_may_wakeup(mmc_dev(host->mmc)) || in sdhci_suspend_host()
3094 host->ier = 0; in sdhci_suspend_host()
3097 free_irq(host->irq, host); in sdhci_suspend_host()
3107 struct mmc_host *mmc = host->mmc; in sdhci_resume_host() local
3110 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_resume_host()
3111 if (host->ops->enable_dma) in sdhci_resume_host()
3112 host->ops->enable_dma(host); in sdhci_resume_host()
3115 if ((host->mmc->pm_flags & MMC_PM_KEEP_POWER) && in sdhci_resume_host()
3116 (host->quirks2 & SDHCI_QUIRK2_HOST_OFF_CARD_ON)) { in sdhci_resume_host()
3119 host->pwr = 0; in sdhci_resume_host()
3120 host->clock = 0; in sdhci_resume_host()
3121 mmc->ops->set_ios(mmc, &mmc->ios); in sdhci_resume_host()
3123 sdhci_init(host, (host->mmc->pm_flags & MMC_PM_KEEP_POWER)); in sdhci_resume_host()
3127 if (host->irq_wake_enabled) { in sdhci_resume_host()
3130 ret = request_threaded_irq(host->irq, sdhci_irq, in sdhci_resume_host()
3132 mmc_hostname(host->mmc), host); in sdhci_resume_host()
3148 mmc_retune_timer_stop(host->mmc); in sdhci_runtime_suspend_host()
3150 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_suspend_host()
3151 host->ier &= SDHCI_INT_CARD_INT; in sdhci_runtime_suspend_host()
3152 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_runtime_suspend_host()
3153 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_runtime_suspend_host()
3154 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_suspend_host()
3156 synchronize_hardirq(host->irq); in sdhci_runtime_suspend_host()
3158 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_suspend_host()
3159 host->runtime_suspended = true; in sdhci_runtime_suspend_host()
3160 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_suspend_host()
3168 struct mmc_host *mmc = host->mmc; in sdhci_runtime_resume_host() local
3170 int host_flags = host->flags; in sdhci_runtime_resume_host()
3173 if (host->ops->enable_dma) in sdhci_runtime_resume_host()
3174 host->ops->enable_dma(host); in sdhci_runtime_resume_host()
3179 if (mmc->ios.power_mode != MMC_POWER_UNDEFINED && in sdhci_runtime_resume_host()
3180 mmc->ios.power_mode != MMC_POWER_OFF) { in sdhci_runtime_resume_host()
3181 /* Force clock and power re-program */ in sdhci_runtime_resume_host()
3182 host->pwr = 0; in sdhci_runtime_resume_host()
3183 host->clock = 0; in sdhci_runtime_resume_host()
3184 mmc->ops->start_signal_voltage_switch(mmc, &mmc->ios); in sdhci_runtime_resume_host()
3185 mmc->ops->set_ios(mmc, &mmc->ios); in sdhci_runtime_resume_host()
3188 !(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN)) { in sdhci_runtime_resume_host()
3189 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_resume_host()
3191 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_resume_host()
3194 if ((mmc->caps2 & MMC_CAP2_HS400_ES) && in sdhci_runtime_resume_host()
3195 mmc->ops->hs400_enhanced_strobe) in sdhci_runtime_resume_host()
3196 mmc->ops->hs400_enhanced_strobe(mmc, &mmc->ios); in sdhci_runtime_resume_host()
3199 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_resume_host()
3201 host->runtime_suspended = false; in sdhci_runtime_resume_host()
3204 if (host->flags & SDHCI_SDIO_IRQ_ENABLED) in sdhci_runtime_resume_host()
3210 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_resume_host()
3224 void sdhci_cqe_enable(struct mmc_host *mmc) in sdhci_cqe_enable() argument
3226 struct sdhci_host *host = mmc_priv(mmc); in sdhci_cqe_enable()
3230 spin_lock_irqsave(&host->lock, flags); in sdhci_cqe_enable()
3234 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_cqe_enable()
3240 sdhci_writew(host, SDHCI_MAKE_BLKSZ(host->sdma_boundary, 512), in sdhci_cqe_enable()
3246 host->ier = host->cqe_ier; in sdhci_cqe_enable()
3248 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_cqe_enable()
3249 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_cqe_enable()
3251 host->cqe_on = true; in sdhci_cqe_enable()
3254 mmc_hostname(mmc), host->ier, in sdhci_cqe_enable()
3258 spin_unlock_irqrestore(&host->lock, flags); in sdhci_cqe_enable()
3262 void sdhci_cqe_disable(struct mmc_host *mmc, bool recovery) in sdhci_cqe_disable() argument
3264 struct sdhci_host *host = mmc_priv(mmc); in sdhci_cqe_disable()
3267 spin_lock_irqsave(&host->lock, flags); in sdhci_cqe_disable()
3271 host->cqe_on = false; in sdhci_cqe_disable()
3279 mmc_hostname(mmc), host->ier, in sdhci_cqe_disable()
3283 spin_unlock_irqrestore(&host->lock, flags); in sdhci_cqe_disable()
3292 if (!host->cqe_on) in sdhci_cqe_irq()
3296 *cmd_error = -EILSEQ; in sdhci_cqe_irq()
3298 *cmd_error = -ETIMEDOUT; in sdhci_cqe_irq()
3303 *data_error = -EILSEQ; in sdhci_cqe_irq()
3305 *data_error = -ETIMEDOUT; in sdhci_cqe_irq()
3307 *data_error = -EIO; in sdhci_cqe_irq()
3312 mask = intmask & host->cqe_ier; in sdhci_cqe_irq()
3317 mmc_hostname(host->mmc)); in sdhci_cqe_irq()
3319 intmask &= ~(host->cqe_ier | SDHCI_INT_ERROR); in sdhci_cqe_irq()
3323 mmc_hostname(host->mmc), intmask); in sdhci_cqe_irq()
3340 struct mmc_host *mmc; in sdhci_alloc_host() local
3345 mmc = mmc_alloc_host(sizeof(struct sdhci_host) + priv_size, dev); in sdhci_alloc_host()
3346 if (!mmc) in sdhci_alloc_host()
3347 return ERR_PTR(-ENOMEM); in sdhci_alloc_host()
3349 host = mmc_priv(mmc); in sdhci_alloc_host()
3350 host->mmc = mmc; in sdhci_alloc_host()
3351 host->mmc_host_ops = sdhci_ops; in sdhci_alloc_host()
3352 mmc->ops = &host->mmc_host_ops; in sdhci_alloc_host()
3354 host->flags = SDHCI_SIGNALING_330; in sdhci_alloc_host()
3356 host->cqe_ier = SDHCI_CQE_INT_MASK; in sdhci_alloc_host()
3357 host->cqe_err_ier = SDHCI_CQE_INT_ERR_MASK; in sdhci_alloc_host()
3359 host->tuning_delay = -1; in sdhci_alloc_host()
3361 host->sdma_boundary = SDHCI_DEFAULT_BOUNDARY_ARG; in sdhci_alloc_host()
3370 struct mmc_host *mmc = host->mmc; in sdhci_set_dma_mask() local
3371 struct device *dev = mmc_dev(mmc); in sdhci_set_dma_mask()
3372 int ret = -EINVAL; in sdhci_set_dma_mask()
3374 if (host->quirks2 & SDHCI_QUIRK2_BROKEN_64_BIT_DMA) in sdhci_set_dma_mask()
3375 host->flags &= ~SDHCI_USE_64_BIT_DMA; in sdhci_set_dma_mask()
3377 /* Try 64-bit mask if hardware is capable of it */ in sdhci_set_dma_mask()
3378 if (host->flags & SDHCI_USE_64_BIT_DMA) { in sdhci_set_dma_mask()
3381 pr_warn("%s: Failed to set 64-bit DMA mask.\n", in sdhci_set_dma_mask()
3382 mmc_hostname(mmc)); in sdhci_set_dma_mask()
3383 host->flags &= ~SDHCI_USE_64_BIT_DMA; in sdhci_set_dma_mask()
3387 /* 32-bit mask as default & fallback */ in sdhci_set_dma_mask()
3391 pr_warn("%s: Failed to set 32-bit DMA mask.\n", in sdhci_set_dma_mask()
3392 mmc_hostname(mmc)); in sdhci_set_dma_mask()
3404 if (host->read_caps) in __sdhci_read_caps()
3407 host->read_caps = true; in __sdhci_read_caps()
3410 host->quirks = debug_quirks; in __sdhci_read_caps()
3413 host->quirks2 = debug_quirks2; in __sdhci_read_caps()
3417 of_property_read_u64(mmc_dev(host->mmc)->of_node, in __sdhci_read_caps()
3418 "sdhci-caps-mask", &dt_caps_mask); in __sdhci_read_caps()
3419 of_property_read_u64(mmc_dev(host->mmc)->of_node, in __sdhci_read_caps()
3420 "sdhci-caps", &dt_caps); in __sdhci_read_caps()
3423 host->version = (v & SDHCI_SPEC_VER_MASK) >> SDHCI_SPEC_VER_SHIFT; in __sdhci_read_caps()
3425 if (host->quirks & SDHCI_QUIRK_MISSING_CAPS) in __sdhci_read_caps()
3429 host->caps = *caps; in __sdhci_read_caps()
3431 host->caps = sdhci_readl(host, SDHCI_CAPABILITIES); in __sdhci_read_caps()
3432 host->caps &= ~lower_32_bits(dt_caps_mask); in __sdhci_read_caps()
3433 host->caps |= lower_32_bits(dt_caps); in __sdhci_read_caps()
3436 if (host->version < SDHCI_SPEC_300) in __sdhci_read_caps()
3440 host->caps1 = *caps1; in __sdhci_read_caps()
3442 host->caps1 = sdhci_readl(host, SDHCI_CAPABILITIES_1); in __sdhci_read_caps()
3443 host->caps1 &= ~upper_32_bits(dt_caps_mask); in __sdhci_read_caps()
3444 host->caps1 |= upper_32_bits(dt_caps); in __sdhci_read_caps()
3451 struct mmc_host *mmc = host->mmc; in sdhci_allocate_bounce_buffer() local
3458 * has diminishing returns, this is probably because SD/MMC in sdhci_allocate_bounce_buffer()
3467 if (mmc->max_req_size < bounce_size) in sdhci_allocate_bounce_buffer()
3468 bounce_size = mmc->max_req_size; in sdhci_allocate_bounce_buffer()
3476 host->bounce_buffer = devm_kmalloc(mmc->parent, in sdhci_allocate_bounce_buffer()
3479 if (!host->bounce_buffer) { in sdhci_allocate_bounce_buffer()
3481 mmc_hostname(mmc), in sdhci_allocate_bounce_buffer()
3485 * mmc->max_segs == 1. in sdhci_allocate_bounce_buffer()
3490 host->bounce_addr = dma_map_single(mmc->parent, in sdhci_allocate_bounce_buffer()
3491 host->bounce_buffer, in sdhci_allocate_bounce_buffer()
3494 ret = dma_mapping_error(mmc->parent, host->bounce_addr); in sdhci_allocate_bounce_buffer()
3498 host->bounce_buffer_size = bounce_size; in sdhci_allocate_bounce_buffer()
3501 mmc->max_segs = max_blocks; in sdhci_allocate_bounce_buffer()
3502 mmc->max_seg_size = bounce_size; in sdhci_allocate_bounce_buffer()
3503 mmc->max_req_size = bounce_size; in sdhci_allocate_bounce_buffer()
3506 mmc_hostname(mmc), max_blocks, bounce_size); in sdhci_allocate_bounce_buffer()
3513 struct mmc_host *mmc; in sdhci_setup_host() local
3522 return -EINVAL; in sdhci_setup_host()
3524 mmc = host->mmc; in sdhci_setup_host()
3532 ret = mmc_regulator_get_supply(mmc); in sdhci_setup_host()
3545 override_timeout_clk = host->timeout_clk; in sdhci_setup_host()
3547 if (host->version > SDHCI_SPEC_300) { in sdhci_setup_host()
3549 mmc_hostname(mmc), host->version); in sdhci_setup_host()
3552 if (host->quirks & SDHCI_QUIRK_BROKEN_CQE) in sdhci_setup_host()
3553 mmc->caps2 &= ~MMC_CAP2_CQE; in sdhci_setup_host()
3555 if (host->quirks & SDHCI_QUIRK_FORCE_DMA) in sdhci_setup_host()
3556 host->flags |= SDHCI_USE_SDMA; in sdhci_setup_host()
3557 else if (!(host->caps & SDHCI_CAN_DO_SDMA)) in sdhci_setup_host()
3560 host->flags |= SDHCI_USE_SDMA; in sdhci_setup_host()
3562 if ((host->quirks & SDHCI_QUIRK_BROKEN_DMA) && in sdhci_setup_host()
3563 (host->flags & SDHCI_USE_SDMA)) { in sdhci_setup_host()
3565 host->flags &= ~SDHCI_USE_SDMA; in sdhci_setup_host()
3568 if ((host->version >= SDHCI_SPEC_200) && in sdhci_setup_host()
3569 (host->caps & SDHCI_CAN_DO_ADMA2)) in sdhci_setup_host()
3570 host->flags |= SDHCI_USE_ADMA; in sdhci_setup_host()
3572 if ((host->quirks & SDHCI_QUIRK_BROKEN_ADMA) && in sdhci_setup_host()
3573 (host->flags & SDHCI_USE_ADMA)) { in sdhci_setup_host()
3575 host->flags &= ~SDHCI_USE_ADMA; in sdhci_setup_host()
3579 * It is assumed that a 64-bit capable device has set a 64-bit DMA mask in sdhci_setup_host()
3580 * and *must* do 64-bit DMA. A driver has the opportunity to change in sdhci_setup_host()
3581 * that during the first call to ->enable_dma(). Similarly in sdhci_setup_host()
3585 if (host->caps & SDHCI_CAN_64BIT) in sdhci_setup_host()
3586 host->flags |= SDHCI_USE_64_BIT_DMA; in sdhci_setup_host()
3588 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_setup_host()
3591 if (!ret && host->ops->enable_dma) in sdhci_setup_host()
3592 ret = host->ops->enable_dma(host); in sdhci_setup_host()
3595 pr_warn("%s: No suitable DMA available - falling back to PIO\n", in sdhci_setup_host()
3596 mmc_hostname(mmc)); in sdhci_setup_host()
3597 host->flags &= ~(SDHCI_USE_SDMA | SDHCI_USE_ADMA); in sdhci_setup_host()
3603 /* SDMA does not support 64-bit DMA */ in sdhci_setup_host()
3604 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_setup_host()
3605 host->flags &= ~SDHCI_USE_SDMA; in sdhci_setup_host()
3607 if (host->flags & SDHCI_USE_ADMA) { in sdhci_setup_host()
3617 if (host->flags & SDHCI_USE_64_BIT_DMA) { in sdhci_setup_host()
3618 host->adma_table_sz = (SDHCI_MAX_SEGS * 2 + 1) * in sdhci_setup_host()
3620 host->desc_sz = SDHCI_ADMA2_64_DESC_SZ; in sdhci_setup_host()
3622 host->adma_table_sz = (SDHCI_MAX_SEGS * 2 + 1) * in sdhci_setup_host()
3624 host->desc_sz = SDHCI_ADMA2_32_DESC_SZ; in sdhci_setup_host()
3627 host->align_buffer_sz = SDHCI_MAX_SEGS * SDHCI_ADMA2_ALIGN; in sdhci_setup_host()
3628 buf = dma_alloc_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_setup_host()
3629 host->adma_table_sz, &dma, GFP_KERNEL); in sdhci_setup_host()
3631 pr_warn("%s: Unable to allocate ADMA buffers - falling back to standard DMA\n", in sdhci_setup_host()
3632 mmc_hostname(mmc)); in sdhci_setup_host()
3633 host->flags &= ~SDHCI_USE_ADMA; in sdhci_setup_host()
3634 } else if ((dma + host->align_buffer_sz) & in sdhci_setup_host()
3635 (SDHCI_ADMA2_DESC_ALIGN - 1)) { in sdhci_setup_host()
3637 mmc_hostname(mmc)); in sdhci_setup_host()
3638 host->flags &= ~SDHCI_USE_ADMA; in sdhci_setup_host()
3639 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_setup_host()
3640 host->adma_table_sz, buf, dma); in sdhci_setup_host()
3642 host->align_buffer = buf; in sdhci_setup_host()
3643 host->align_addr = dma; in sdhci_setup_host()
3645 host->adma_table = buf + host->align_buffer_sz; in sdhci_setup_host()
3646 host->adma_addr = dma + host->align_buffer_sz; in sdhci_setup_host()
3655 if (!(host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA))) { in sdhci_setup_host()
3656 host->dma_mask = DMA_BIT_MASK(64); in sdhci_setup_host()
3657 mmc_dev(mmc)->dma_mask = &host->dma_mask; in sdhci_setup_host()
3660 if (host->version >= SDHCI_SPEC_300) in sdhci_setup_host()
3661 host->max_clk = (host->caps & SDHCI_CLOCK_V3_BASE_MASK) in sdhci_setup_host()
3664 host->max_clk = (host->caps & SDHCI_CLOCK_BASE_MASK) in sdhci_setup_host()
3667 host->max_clk *= 1000000; in sdhci_setup_host()
3668 if (host->max_clk == 0 || host->quirks & in sdhci_setup_host()
3670 if (!host->ops->get_max_clock) { in sdhci_setup_host()
3672 mmc_hostname(mmc)); in sdhci_setup_host()
3673 ret = -ENODEV; in sdhci_setup_host()
3676 host->max_clk = host->ops->get_max_clock(host); in sdhci_setup_host()
3683 host->clk_mul = (host->caps1 & SDHCI_CLOCK_MUL_MASK) >> in sdhci_setup_host()
3692 if (host->clk_mul) in sdhci_setup_host()
3693 host->clk_mul += 1; in sdhci_setup_host()
3698 max_clk = host->max_clk; in sdhci_setup_host()
3700 if (host->ops->get_min_clock) in sdhci_setup_host()
3701 mmc->f_min = host->ops->get_min_clock(host); in sdhci_setup_host()
3702 else if (host->version >= SDHCI_SPEC_300) { in sdhci_setup_host()
3703 if (host->clk_mul) in sdhci_setup_host()
3704 max_clk = host->max_clk * host->clk_mul; in sdhci_setup_host()
3709 mmc->f_min = host->max_clk / SDHCI_MAX_DIV_SPEC_300; in sdhci_setup_host()
3711 mmc->f_min = host->max_clk / SDHCI_MAX_DIV_SPEC_200; in sdhci_setup_host()
3713 if (!mmc->f_max || mmc->f_max > max_clk) in sdhci_setup_host()
3714 mmc->f_max = max_clk; in sdhci_setup_host()
3716 if (!(host->quirks & SDHCI_QUIRK_DATA_TIMEOUT_USES_SDCLK)) { in sdhci_setup_host()
3717 host->timeout_clk = (host->caps & SDHCI_TIMEOUT_CLK_MASK) >> in sdhci_setup_host()
3720 if (host->caps & SDHCI_TIMEOUT_CLK_UNIT) in sdhci_setup_host()
3721 host->timeout_clk *= 1000; in sdhci_setup_host()
3723 if (host->timeout_clk == 0) { in sdhci_setup_host()
3724 if (!host->ops->get_timeout_clock) { in sdhci_setup_host()
3726 mmc_hostname(mmc)); in sdhci_setup_host()
3727 ret = -ENODEV; in sdhci_setup_host()
3731 host->timeout_clk = in sdhci_setup_host()
3732 DIV_ROUND_UP(host->ops->get_timeout_clock(host), in sdhci_setup_host()
3737 host->timeout_clk = override_timeout_clk; in sdhci_setup_host()
3739 mmc->max_busy_timeout = host->ops->get_max_timeout_count ? in sdhci_setup_host()
3740 host->ops->get_max_timeout_count(host) : 1 << 27; in sdhci_setup_host()
3741 mmc->max_busy_timeout /= host->timeout_clk; in sdhci_setup_host()
3744 if (host->quirks2 & SDHCI_QUIRK2_DISABLE_HW_TIMEOUT && in sdhci_setup_host()
3745 !host->ops->get_max_timeout_count) in sdhci_setup_host()
3746 mmc->max_busy_timeout = 0; in sdhci_setup_host()
3748 mmc->caps |= MMC_CAP_SDIO_IRQ | MMC_CAP_ERASE | MMC_CAP_CMD23; in sdhci_setup_host()
3749 mmc->caps2 |= MMC_CAP2_SDIO_IRQ_NOTHREAD; in sdhci_setup_host()
3751 if (host->quirks & SDHCI_QUIRK_MULTIBLOCK_READ_ACMD12) in sdhci_setup_host()
3752 host->flags |= SDHCI_AUTO_CMD12; in sdhci_setup_host()
3754 /* Auto-CMD23 stuff only works in ADMA or PIO. */ in sdhci_setup_host()
3755 if ((host->version >= SDHCI_SPEC_300) && in sdhci_setup_host()
3756 ((host->flags & SDHCI_USE_ADMA) || in sdhci_setup_host()
3757 !(host->flags & SDHCI_USE_SDMA)) && in sdhci_setup_host()
3758 !(host->quirks2 & SDHCI_QUIRK2_ACMD23_BROKEN)) { in sdhci_setup_host()
3759 host->flags |= SDHCI_AUTO_CMD23; in sdhci_setup_host()
3760 DBG("Auto-CMD23 available\n"); in sdhci_setup_host()
3762 DBG("Auto-CMD23 unavailable\n"); in sdhci_setup_host()
3766 * A controller may support 8-bit width, but the board itself in sdhci_setup_host()
3768 * 8-bit width must set "mmc->caps |= MMC_CAP_8_BIT_DATA;" in in sdhci_setup_host()
3770 * won't assume 8-bit width for hosts without that CAP. in sdhci_setup_host()
3772 if (!(host->quirks & SDHCI_QUIRK_FORCE_1_BIT_DATA)) in sdhci_setup_host()
3773 mmc->caps |= MMC_CAP_4_BIT_DATA; in sdhci_setup_host()
3775 if (host->quirks2 & SDHCI_QUIRK2_HOST_NO_CMD23) in sdhci_setup_host()
3776 mmc->caps &= ~MMC_CAP_CMD23; in sdhci_setup_host()
3778 if (host->caps & SDHCI_CAN_DO_HISPD) in sdhci_setup_host()
3779 mmc->caps |= MMC_CAP_SD_HIGHSPEED | MMC_CAP_MMC_HIGHSPEED; in sdhci_setup_host()
3781 if ((host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) && in sdhci_setup_host()
3782 mmc_card_is_removable(mmc) && in sdhci_setup_host()
3783 mmc_gpio_get_cd(host->mmc) < 0) in sdhci_setup_host()
3784 mmc->caps |= MMC_CAP_NEEDS_POLL; in sdhci_setup_host()
3786 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_setup_host()
3787 ret = regulator_enable(mmc->supply.vqmmc); in sdhci_setup_host()
3789 /* If vqmmc provides no 1.8V signalling, then there's no UHS */ in sdhci_setup_host()
3790 if (!regulator_is_supported_voltage(mmc->supply.vqmmc, 1700000, in sdhci_setup_host()
3792 host->caps1 &= ~(SDHCI_SUPPORT_SDR104 | in sdhci_setup_host()
3797 if (!regulator_is_supported_voltage(mmc->supply.vqmmc, 2700000, in sdhci_setup_host()
3799 host->flags &= ~SDHCI_SIGNALING_330; in sdhci_setup_host()
3803 mmc_hostname(mmc), ret); in sdhci_setup_host()
3804 mmc->supply.vqmmc = ERR_PTR(-EINVAL); in sdhci_setup_host()
3808 if (host->quirks2 & SDHCI_QUIRK2_NO_1_8_V) { in sdhci_setup_host()
3809 host->caps1 &= ~(SDHCI_SUPPORT_SDR104 | SDHCI_SUPPORT_SDR50 | in sdhci_setup_host()
3813 * (indicated using mmc-hs200-1_8v/mmc-hs400-1_8v dt property), in sdhci_setup_host()
3819 mmc->caps2 &= ~(MMC_CAP2_HSX00_1_8V | MMC_CAP2_HS400_ES); in sdhci_setup_host()
3820 mmc->caps &= ~(MMC_CAP_1_8V_DDR | MMC_CAP_UHS); in sdhci_setup_host()
3823 /* Any UHS-I mode in caps implies SDR12 and SDR25 support. */ in sdhci_setup_host()
3824 if (host->caps1 & (SDHCI_SUPPORT_SDR104 | SDHCI_SUPPORT_SDR50 | in sdhci_setup_host()
3826 mmc->caps |= MMC_CAP_UHS_SDR12 | MMC_CAP_UHS_SDR25; in sdhci_setup_host()
3829 if (host->caps1 & SDHCI_SUPPORT_SDR104) { in sdhci_setup_host()
3830 mmc->caps |= MMC_CAP_UHS_SDR104 | MMC_CAP_UHS_SDR50; in sdhci_setup_host()
3834 if (!(host->quirks2 & SDHCI_QUIRK2_BROKEN_HS200)) in sdhci_setup_host()
3835 mmc->caps2 |= MMC_CAP2_HS200; in sdhci_setup_host()
3836 } else if (host->caps1 & SDHCI_SUPPORT_SDR50) { in sdhci_setup_host()
3837 mmc->caps |= MMC_CAP_UHS_SDR50; in sdhci_setup_host()
3840 if (host->quirks2 & SDHCI_QUIRK2_CAPS_BIT63_FOR_HS400 && in sdhci_setup_host()
3841 (host->caps1 & SDHCI_SUPPORT_HS400)) in sdhci_setup_host()
3842 mmc->caps2 |= MMC_CAP2_HS400; in sdhci_setup_host()
3844 if ((mmc->caps2 & MMC_CAP2_HSX00_1_2V) && in sdhci_setup_host()
3845 (IS_ERR(mmc->supply.vqmmc) || in sdhci_setup_host()
3846 !regulator_is_supported_voltage(mmc->supply.vqmmc, 1100000, in sdhci_setup_host()
3848 mmc->caps2 &= ~MMC_CAP2_HSX00_1_2V; in sdhci_setup_host()
3850 if ((host->caps1 & SDHCI_SUPPORT_DDR50) && in sdhci_setup_host()
3851 !(host->quirks2 & SDHCI_QUIRK2_BROKEN_DDR50)) in sdhci_setup_host()
3852 mmc->caps |= MMC_CAP_UHS_DDR50; in sdhci_setup_host()
3855 if (host->caps1 & SDHCI_USE_SDR50_TUNING) in sdhci_setup_host()
3856 host->flags |= SDHCI_SDR50_NEEDS_TUNING; in sdhci_setup_host()
3859 if (host->caps1 & SDHCI_DRIVER_TYPE_A) in sdhci_setup_host()
3860 mmc->caps |= MMC_CAP_DRIVER_TYPE_A; in sdhci_setup_host()
3861 if (host->caps1 & SDHCI_DRIVER_TYPE_C) in sdhci_setup_host()
3862 mmc->caps |= MMC_CAP_DRIVER_TYPE_C; in sdhci_setup_host()
3863 if (host->caps1 & SDHCI_DRIVER_TYPE_D) in sdhci_setup_host()
3864 mmc->caps |= MMC_CAP_DRIVER_TYPE_D; in sdhci_setup_host()
3866 /* Initial value for re-tuning timer count */ in sdhci_setup_host()
3867 host->tuning_count = (host->caps1 & SDHCI_RETUNING_TIMER_COUNT_MASK) >> in sdhci_setup_host()
3871 * In case Re-tuning Timer is not disabled, the actual value of in sdhci_setup_host()
3872 * re-tuning timer will be 2 ^ (n - 1). in sdhci_setup_host()
3874 if (host->tuning_count) in sdhci_setup_host()
3875 host->tuning_count = 1 << (host->tuning_count - 1); in sdhci_setup_host()
3877 /* Re-tuning mode supported by the Host Controller */ in sdhci_setup_host()
3878 host->tuning_mode = (host->caps1 & SDHCI_RETUNING_MODE_MASK) >> in sdhci_setup_host()
3891 if (!max_current_caps && !IS_ERR(mmc->supply.vmmc)) { in sdhci_setup_host()
3892 int curr = regulator_get_current_limit(mmc->supply.vmmc); in sdhci_setup_host()
3907 if (host->caps & SDHCI_CAN_VDD_330) { in sdhci_setup_host()
3910 mmc->max_current_330 = ((max_current_caps & in sdhci_setup_host()
3915 if (host->caps & SDHCI_CAN_VDD_300) { in sdhci_setup_host()
3918 mmc->max_current_300 = ((max_current_caps & in sdhci_setup_host()
3923 if (host->caps & SDHCI_CAN_VDD_180) { in sdhci_setup_host()
3926 mmc->max_current_180 = ((max_current_caps & in sdhci_setup_host()
3933 if (host->ocr_mask) in sdhci_setup_host()
3934 ocr_avail = host->ocr_mask; in sdhci_setup_host()
3937 if (mmc->ocr_avail) in sdhci_setup_host()
3938 ocr_avail = mmc->ocr_avail; in sdhci_setup_host()
3940 mmc->ocr_avail = ocr_avail; in sdhci_setup_host()
3941 mmc->ocr_avail_sdio = ocr_avail; in sdhci_setup_host()
3942 if (host->ocr_avail_sdio) in sdhci_setup_host()
3943 mmc->ocr_avail_sdio &= host->ocr_avail_sdio; in sdhci_setup_host()
3944 mmc->ocr_avail_sd = ocr_avail; in sdhci_setup_host()
3945 if (host->ocr_avail_sd) in sdhci_setup_host()
3946 mmc->ocr_avail_sd &= host->ocr_avail_sd; in sdhci_setup_host()
3948 mmc->ocr_avail_sd &= ~MMC_VDD_165_195; in sdhci_setup_host()
3949 mmc->ocr_avail_mmc = ocr_avail; in sdhci_setup_host()
3950 if (host->ocr_avail_mmc) in sdhci_setup_host()
3951 mmc->ocr_avail_mmc &= host->ocr_avail_mmc; in sdhci_setup_host()
3953 if (mmc->ocr_avail == 0) { in sdhci_setup_host()
3955 mmc_hostname(mmc)); in sdhci_setup_host()
3956 ret = -ENODEV; in sdhci_setup_host()
3960 if ((mmc->caps & (MMC_CAP_UHS_SDR12 | MMC_CAP_UHS_SDR25 | in sdhci_setup_host()
3963 (mmc->caps2 & (MMC_CAP2_HS200_1_8V_SDR | MMC_CAP2_HS400_1_8V))) in sdhci_setup_host()
3964 host->flags |= SDHCI_SIGNALING_180; in sdhci_setup_host()
3966 if (mmc->caps2 & MMC_CAP2_HSX00_1_2V) in sdhci_setup_host()
3967 host->flags |= SDHCI_SIGNALING_120; in sdhci_setup_host()
3969 spin_lock_init(&host->lock); in sdhci_setup_host()
3976 mmc->max_req_size = 524288; in sdhci_setup_host()
3982 if (host->flags & SDHCI_USE_ADMA) { in sdhci_setup_host()
3983 mmc->max_segs = SDHCI_MAX_SEGS; in sdhci_setup_host()
3984 } else if (host->flags & SDHCI_USE_SDMA) { in sdhci_setup_host()
3985 mmc->max_segs = 1; in sdhci_setup_host()
3989 mmc->max_req_size = min(mmc->max_req_size, in sdhci_setup_host()
3993 mmc->max_segs = SDHCI_MAX_SEGS; in sdhci_setup_host()
4001 if (host->flags & SDHCI_USE_ADMA) { in sdhci_setup_host()
4002 if (host->quirks & SDHCI_QUIRK_BROKEN_ADMA_ZEROLEN_DESC) in sdhci_setup_host()
4003 mmc->max_seg_size = 65535; in sdhci_setup_host()
4005 mmc->max_seg_size = 65536; in sdhci_setup_host()
4007 mmc->max_seg_size = mmc->max_req_size; in sdhci_setup_host()
4014 if (host->quirks & SDHCI_QUIRK_FORCE_BLK_SZ_2048) { in sdhci_setup_host()
4015 mmc->max_blk_size = 2; in sdhci_setup_host()
4017 mmc->max_blk_size = (host->caps & SDHCI_MAX_BLOCK_MASK) >> in sdhci_setup_host()
4019 if (mmc->max_blk_size >= 3) { in sdhci_setup_host()
4021 mmc_hostname(mmc)); in sdhci_setup_host()
4022 mmc->max_blk_size = 0; in sdhci_setup_host()
4026 mmc->max_blk_size = 512 << mmc->max_blk_size; in sdhci_setup_host()
4031 mmc->max_blk_count = (host->quirks & SDHCI_QUIRK_NO_MULTIBLOCK) ? 1 : 65535; in sdhci_setup_host()
4033 if (mmc->max_segs == 1) { in sdhci_setup_host()
4034 /* This may alter mmc->*_blk_* parameters */ in sdhci_setup_host()
4043 if (!IS_ERR(mmc->supply.vqmmc)) in sdhci_setup_host()
4044 regulator_disable(mmc->supply.vqmmc); in sdhci_setup_host()
4046 if (host->align_buffer) in sdhci_setup_host()
4047 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_setup_host()
4048 host->adma_table_sz, host->align_buffer, in sdhci_setup_host()
4049 host->align_addr); in sdhci_setup_host()
4050 host->adma_table = NULL; in sdhci_setup_host()
4051 host->align_buffer = NULL; in sdhci_setup_host()
4059 struct mmc_host *mmc = host->mmc; in sdhci_cleanup_host() local
4061 if (!IS_ERR(mmc->supply.vqmmc)) in sdhci_cleanup_host()
4062 regulator_disable(mmc->supply.vqmmc); in sdhci_cleanup_host()
4064 if (host->align_buffer) in sdhci_cleanup_host()
4065 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_cleanup_host()
4066 host->adma_table_sz, host->align_buffer, in sdhci_cleanup_host()
4067 host->align_addr); in sdhci_cleanup_host()
4068 host->adma_table = NULL; in sdhci_cleanup_host()
4069 host->align_buffer = NULL; in sdhci_cleanup_host()
4075 struct mmc_host *mmc = host->mmc; in __sdhci_add_host() local
4081 tasklet_init(&host->finish_tasklet, in __sdhci_add_host()
4084 timer_setup(&host->timer, sdhci_timeout_timer, 0); in __sdhci_add_host()
4085 timer_setup(&host->data_timer, sdhci_timeout_data_timer, 0); in __sdhci_add_host()
4087 init_waitqueue_head(&host->buf_ready_int); in __sdhci_add_host()
4091 ret = request_threaded_irq(host->irq, sdhci_irq, sdhci_thread_irq, in __sdhci_add_host()
4092 IRQF_SHARED, mmc_hostname(mmc), host); in __sdhci_add_host()
4095 mmc_hostname(mmc), host->irq, ret); in __sdhci_add_host()
4102 mmc_hostname(mmc), ret); in __sdhci_add_host()
4108 ret = mmc_add_host(mmc); in __sdhci_add_host()
4113 mmc_hostname(mmc), host->hw_name, dev_name(mmc_dev(mmc)), in __sdhci_add_host()
4114 (host->flags & SDHCI_USE_ADMA) ? in __sdhci_add_host()
4115 (host->flags & SDHCI_USE_64_BIT_DMA) ? "ADMA 64-bit" : "ADMA" : in __sdhci_add_host()
4116 (host->flags & SDHCI_USE_SDMA) ? "DMA" : "PIO"); in __sdhci_add_host()
4128 free_irq(host->irq, host); in __sdhci_add_host()
4130 tasklet_kill(&host->finish_tasklet); in __sdhci_add_host()
4159 struct mmc_host *mmc = host->mmc; in sdhci_remove_host() local
4163 spin_lock_irqsave(&host->lock, flags); in sdhci_remove_host()
4165 host->flags |= SDHCI_DEVICE_DEAD; in sdhci_remove_host()
4169 " transfer!\n", mmc_hostname(mmc)); in sdhci_remove_host()
4170 sdhci_error_out_mrqs(host, -ENOMEDIUM); in sdhci_remove_host()
4173 spin_unlock_irqrestore(&host->lock, flags); in sdhci_remove_host()
4178 mmc_remove_host(mmc); in sdhci_remove_host()
4187 free_irq(host->irq, host); in sdhci_remove_host()
4189 del_timer_sync(&host->timer); in sdhci_remove_host()
4190 del_timer_sync(&host->data_timer); in sdhci_remove_host()
4192 tasklet_kill(&host->finish_tasklet); in sdhci_remove_host()
4194 if (!IS_ERR(mmc->supply.vqmmc)) in sdhci_remove_host()
4195 regulator_disable(mmc->supply.vqmmc); in sdhci_remove_host()
4197 if (host->align_buffer) in sdhci_remove_host()
4198 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_remove_host()
4199 host->adma_table_sz, host->align_buffer, in sdhci_remove_host()
4200 host->align_addr); in sdhci_remove_host()
4202 host->adma_table = NULL; in sdhci_remove_host()
4203 host->align_buffer = NULL; in sdhci_remove_host()
4210 mmc_free_host(host->mmc); in sdhci_free_host()