Lines Matching +full:cmd +full:- +full:dat +full:- +full:pins
1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * linux/drivers/mmc/host/sdhci.c - Secure Digital Host Controller Interface driver
5 * Copyright (C) 2005-2008 Pierre Ossman, All Rights Reserved.
9 * - JMicron (hardware and technical support)
19 #include <linux/dma-mapping.h>
34 #include <linux/mmc/slot-gpio.h>
41 pr_debug("%s: " DRIVER_NAME ": " f, mmc_hostname(host->mmc), ## x)
44 pr_err("%s: " DRIVER_NAME ": " f, mmc_hostname(host->mmc), ## x)
53 static bool sdhci_send_command(struct sdhci_host *host, struct mmc_command *cmd);
74 SDHCI_DUMP("Wake-up: 0x%08x | Clock: 0x%08x\n", in sdhci_dumpregs()
89 SDHCI_DUMP("Cmd: 0x%08x | Max curr: 0x%08x\n", in sdhci_dumpregs()
101 if (host->flags & SDHCI_USE_ADMA) { in sdhci_dumpregs()
102 if (host->flags & SDHCI_USE_64_BIT_DMA) { in sdhci_dumpregs()
114 if (host->ops->dump_vendor_regs) in sdhci_dumpregs()
115 host->ops->dump_vendor_regs(host); in sdhci_dumpregs()
145 host->v4_mode = true; in sdhci_enable_v4_mode()
150 static inline bool sdhci_data_line_cmd(struct mmc_command *cmd) in sdhci_data_line_cmd() argument
152 return cmd->data || cmd->flags & MMC_RSP_BUSY; in sdhci_data_line_cmd()
159 if ((host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) || in sdhci_set_card_detection()
160 !mmc_card_is_removable(host->mmc) || mmc_can_gpio_cd(host->mmc)) in sdhci_set_card_detection()
167 host->ier |= present ? SDHCI_INT_CARD_REMOVE : in sdhci_set_card_detection()
170 host->ier &= ~(SDHCI_INT_CARD_REMOVE | SDHCI_INT_CARD_INSERT); in sdhci_set_card_detection()
173 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_card_detection()
174 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_card_detection()
189 if (host->bus_on) in sdhci_runtime_pm_bus_on()
191 host->bus_on = true; in sdhci_runtime_pm_bus_on()
192 pm_runtime_get_noresume(host->mmc->parent); in sdhci_runtime_pm_bus_on()
197 if (!host->bus_on) in sdhci_runtime_pm_bus_off()
199 host->bus_on = false; in sdhci_runtime_pm_bus_off()
200 pm_runtime_put_noidle(host->mmc->parent); in sdhci_runtime_pm_bus_off()
210 host->clock = 0; in sdhci_reset()
211 /* Reset-all turns off SD Bus Power */ in sdhci_reset()
212 if (host->quirks2 & SDHCI_QUIRK2_CARD_ON_NEEDS_BUS_ON) in sdhci_reset()
227 mmc_hostname(host->mmc), (int)mask); in sdhci_reset()
238 if (host->quirks & SDHCI_QUIRK_NO_CARD_NO_RESET) { in sdhci_do_reset()
239 struct mmc_host *mmc = host->mmc; in sdhci_do_reset()
241 if (!mmc->ops->get_cd(mmc)) in sdhci_do_reset()
245 host->ops->reset(host, mask); in sdhci_do_reset()
248 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_do_reset()
249 if (host->ops->enable_dma) in sdhci_do_reset()
250 host->ops->enable_dma(host); in sdhci_do_reset()
254 host->preset_enabled = false; in sdhci_do_reset()
260 host->ier = SDHCI_INT_BUS_POWER | SDHCI_INT_DATA_END_BIT | in sdhci_set_default_irqs()
266 if (host->tuning_mode == SDHCI_TUNING_MODE_2 || in sdhci_set_default_irqs()
267 host->tuning_mode == SDHCI_TUNING_MODE_3) in sdhci_set_default_irqs()
268 host->ier |= SDHCI_INT_RETUNE; in sdhci_set_default_irqs()
270 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_default_irqs()
271 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_default_irqs()
279 if (host->version < SDHCI_SPEC_200) in sdhci_config_dma()
290 if (!(host->flags & SDHCI_REQ_USE_DMA)) in sdhci_config_dma()
294 if (host->flags & SDHCI_USE_ADMA) in sdhci_config_dma()
297 if (host->flags & SDHCI_USE_64_BIT_DMA) { in sdhci_config_dma()
299 * If v4 mode, all supported DMA can be 64-bit addressing if in sdhci_config_dma()
300 * controller supports 64-bit system address, otherwise only in sdhci_config_dma()
301 * ADMA can support 64-bit addressing. in sdhci_config_dma()
303 if (host->v4_mode) { in sdhci_config_dma()
307 } else if (host->flags & SDHCI_USE_ADMA) { in sdhci_config_dma()
322 struct mmc_host *mmc = host->mmc; in sdhci_init()
330 if (host->v4_mode) in sdhci_init()
333 spin_lock_irqsave(&host->lock, flags); in sdhci_init()
335 spin_unlock_irqrestore(&host->lock, flags); in sdhci_init()
337 host->cqe_on = false; in sdhci_init()
341 host->clock = 0; in sdhci_init()
342 mmc->ops->set_ios(mmc, &mmc->ios); in sdhci_init()
348 u32 cd = host->ier & (SDHCI_INT_CARD_REMOVE | SDHCI_INT_CARD_INSERT); in sdhci_reinit()
359 if (cd != (host->ier & (SDHCI_INT_CARD_REMOVE | SDHCI_INT_CARD_INSERT))) in sdhci_reinit()
360 mmc_detect_change(host->mmc, msecs_to_jiffies(200)); in sdhci_reinit()
367 if (host->quirks & SDHCI_QUIRK_NO_LED) in __sdhci_led_activate()
379 if (host->quirks & SDHCI_QUIRK_NO_LED) in __sdhci_led_deactivate()
394 spin_lock_irqsave(&host->lock, flags); in sdhci_led_control()
396 if (host->runtime_suspended) in sdhci_led_control()
404 spin_unlock_irqrestore(&host->lock, flags); in sdhci_led_control()
409 struct mmc_host *mmc = host->mmc; in sdhci_led_register()
411 if (host->quirks & SDHCI_QUIRK_NO_LED) in sdhci_led_register()
414 snprintf(host->led_name, sizeof(host->led_name), in sdhci_led_register()
417 host->led.name = host->led_name; in sdhci_led_register()
418 host->led.brightness = LED_OFF; in sdhci_led_register()
419 host->led.default_trigger = mmc_hostname(mmc); in sdhci_led_register()
420 host->led.brightness_set = sdhci_led_control; in sdhci_led_register()
422 return led_classdev_register(mmc_dev(mmc), &host->led); in sdhci_led_register()
427 if (host->quirks & SDHCI_QUIRK_NO_LED) in sdhci_led_unregister()
430 led_classdev_unregister(&host->led); in sdhci_led_unregister()
467 if (sdhci_data_line_cmd(mrq->cmd)) in sdhci_mod_timer()
468 mod_timer(&host->data_timer, timeout); in sdhci_mod_timer()
470 mod_timer(&host->timer, timeout); in sdhci_mod_timer()
475 if (sdhci_data_line_cmd(mrq->cmd)) in sdhci_del_timer()
476 del_timer(&host->data_timer); in sdhci_del_timer()
478 del_timer(&host->timer); in sdhci_del_timer()
483 return host->cmd || host->data_cmd; in sdhci_has_requests()
501 blksize = host->data->blksz; in sdhci_read_block_pio()
507 BUG_ON(!sg_miter_next(&host->sg_miter)); in sdhci_read_block_pio()
509 len = min(host->sg_miter.length, blksize); in sdhci_read_block_pio()
511 blksize -= len; in sdhci_read_block_pio()
512 host->sg_miter.consumed = len; in sdhci_read_block_pio()
514 buf = host->sg_miter.addr; in sdhci_read_block_pio()
526 chunk--; in sdhci_read_block_pio()
527 len--; in sdhci_read_block_pio()
531 sg_miter_stop(&host->sg_miter); in sdhci_read_block_pio()
545 blksize = host->data->blksz; in sdhci_write_block_pio()
552 BUG_ON(!sg_miter_next(&host->sg_miter)); in sdhci_write_block_pio()
554 len = min(host->sg_miter.length, blksize); in sdhci_write_block_pio()
556 blksize -= len; in sdhci_write_block_pio()
557 host->sg_miter.consumed = len; in sdhci_write_block_pio()
559 buf = host->sg_miter.addr; in sdhci_write_block_pio()
566 len--; in sdhci_write_block_pio()
576 sg_miter_stop(&host->sg_miter); in sdhci_write_block_pio()
585 if (host->blocks == 0) in sdhci_transfer_pio()
588 if (host->data->flags & MMC_DATA_READ) in sdhci_transfer_pio()
598 if ((host->quirks & SDHCI_QUIRK_BROKEN_SMALL_PIO) && in sdhci_transfer_pio()
599 (host->data->blocks == 1)) in sdhci_transfer_pio()
603 if (host->quirks & SDHCI_QUIRK_PIO_NEEDS_DELAY) in sdhci_transfer_pio()
606 if (host->data->flags & MMC_DATA_READ) in sdhci_transfer_pio()
611 host->blocks--; in sdhci_transfer_pio()
612 if (host->blocks == 0) in sdhci_transfer_pio()
628 if (data->host_cookie == COOKIE_PRE_MAPPED) in sdhci_pre_dma_transfer()
629 return data->sg_count; in sdhci_pre_dma_transfer()
632 if (host->bounce_buffer) { in sdhci_pre_dma_transfer()
633 unsigned int length = data->blksz * data->blocks; in sdhci_pre_dma_transfer()
635 if (length > host->bounce_buffer_size) { in sdhci_pre_dma_transfer()
637 mmc_hostname(host->mmc), length, in sdhci_pre_dma_transfer()
638 host->bounce_buffer_size); in sdhci_pre_dma_transfer()
639 return -EIO; in sdhci_pre_dma_transfer()
643 if (host->ops->copy_to_bounce_buffer) { in sdhci_pre_dma_transfer()
644 host->ops->copy_to_bounce_buffer(host, in sdhci_pre_dma_transfer()
647 sg_copy_to_buffer(data->sg, data->sg_len, in sdhci_pre_dma_transfer()
648 host->bounce_buffer, length); in sdhci_pre_dma_transfer()
652 dma_sync_single_for_device(host->mmc->parent, in sdhci_pre_dma_transfer()
653 host->bounce_addr, in sdhci_pre_dma_transfer()
654 host->bounce_buffer_size, in sdhci_pre_dma_transfer()
660 sg_count = dma_map_sg(mmc_dev(host->mmc), in sdhci_pre_dma_transfer()
661 data->sg, data->sg_len, in sdhci_pre_dma_transfer()
666 return -ENOSPC; in sdhci_pre_dma_transfer()
668 data->sg_count = sg_count; in sdhci_pre_dma_transfer()
669 data->host_cookie = cookie; in sdhci_pre_dma_transfer()
677 return kmap_atomic(sg_page(sg)) + sg->offset; in sdhci_kmap_atomic()
687 dma_addr_t addr, int len, unsigned int cmd) in sdhci_adma_write_desc() argument
691 /* 32-bit and 64-bit descriptors have these members in same position */ in sdhci_adma_write_desc()
692 dma_desc->cmd = cpu_to_le16(cmd); in sdhci_adma_write_desc()
693 dma_desc->len = cpu_to_le16(len); in sdhci_adma_write_desc()
694 dma_desc->addr_lo = cpu_to_le32(lower_32_bits(addr)); in sdhci_adma_write_desc()
696 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_adma_write_desc()
697 dma_desc->addr_hi = cpu_to_le32(upper_32_bits(addr)); in sdhci_adma_write_desc()
699 *desc += host->desc_sz; in sdhci_adma_write_desc()
705 int len, unsigned int cmd) in __sdhci_adma_write_desc() argument
707 if (host->ops->adma_write_desc) in __sdhci_adma_write_desc()
708 host->ops->adma_write_desc(host, desc, addr, len, cmd); in __sdhci_adma_write_desc()
710 sdhci_adma_write_desc(host, desc, addr, len, cmd); in __sdhci_adma_write_desc()
717 /* 32-bit and 64-bit descriptors have 'cmd' in same position */ in sdhci_adma_mark_end()
718 dma_desc->cmd |= cpu_to_le16(ADMA2_END); in sdhci_adma_mark_end()
736 host->sg_count = sg_count; in sdhci_adma_table_pre()
738 desc = host->adma_table; in sdhci_adma_table_pre()
739 align = host->align_buffer; in sdhci_adma_table_pre()
741 align_addr = host->align_addr; in sdhci_adma_table_pre()
743 for_each_sg(data->sg, sg, host->sg_count, i) { in sdhci_adma_table_pre()
749 * be 32-bit aligned. If they aren't, then we use a bounce in sdhci_adma_table_pre()
753 offset = (SDHCI_ADMA2_ALIGN - (addr & SDHCI_ADMA2_MASK)) & in sdhci_adma_table_pre()
756 if (data->flags & MMC_DATA_WRITE) { in sdhci_adma_table_pre()
772 len -= offset; in sdhci_adma_table_pre()
786 WARN_ON((desc - host->adma_table) >= host->adma_table_sz); in sdhci_adma_table_pre()
789 if (host->quirks & SDHCI_QUIRK_NO_ENDATTR_IN_NOPDESC) { in sdhci_adma_table_pre()
791 if (desc != host->adma_table) { in sdhci_adma_table_pre()
792 desc -= host->desc_sz; in sdhci_adma_table_pre()
796 /* Add a terminating entry - nop, end, valid */ in sdhci_adma_table_pre()
810 if (data->flags & MMC_DATA_READ) { in sdhci_adma_table_post()
814 for_each_sg(data->sg, sg, host->sg_count, i) in sdhci_adma_table_post()
821 dma_sync_sg_for_cpu(mmc_dev(host->mmc), data->sg, in sdhci_adma_table_post()
822 data->sg_len, DMA_FROM_DEVICE); in sdhci_adma_table_post()
824 align = host->align_buffer; in sdhci_adma_table_post()
826 for_each_sg(data->sg, sg, host->sg_count, i) { in sdhci_adma_table_post()
828 size = SDHCI_ADMA2_ALIGN - in sdhci_adma_table_post()
845 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_set_adma_addr()
851 if (host->bounce_buffer) in sdhci_sdma_address()
852 return host->bounce_addr; in sdhci_sdma_address()
854 return sg_dma_address(host->data->sg); in sdhci_sdma_address()
859 if (host->v4_mode) in sdhci_set_sdma_addr()
866 struct mmc_command *cmd, in sdhci_target_timeout() argument
873 target_timeout = cmd->busy_timeout * 1000; in sdhci_target_timeout()
875 target_timeout = DIV_ROUND_UP(data->timeout_ns, 1000); in sdhci_target_timeout()
876 if (host->clock && data->timeout_clks) { in sdhci_target_timeout()
880 * data->timeout_clks is in units of clock cycles. in sdhci_target_timeout()
881 * host->clock is in Hz. target_timeout is in us. in sdhci_target_timeout()
884 val = 1000000ULL * data->timeout_clks; in sdhci_target_timeout()
885 if (do_div(val, host->clock)) in sdhci_target_timeout()
895 struct mmc_command *cmd) in sdhci_calc_sw_timeout() argument
897 struct mmc_data *data = cmd->data; in sdhci_calc_sw_timeout()
898 struct mmc_host *mmc = host->mmc; in sdhci_calc_sw_timeout()
899 struct mmc_ios *ios = &mmc->ios; in sdhci_calc_sw_timeout()
900 unsigned char bus_width = 1 << ios->bus_width; in sdhci_calc_sw_timeout()
906 target_timeout = sdhci_target_timeout(host, cmd, data); in sdhci_calc_sw_timeout()
910 blksz = data->blksz; in sdhci_calc_sw_timeout()
911 freq = host->mmc->actual_clock ? : host->clock; in sdhci_calc_sw_timeout()
917 host->data_timeout = data->blocks * target_timeout + in sdhci_calc_sw_timeout()
920 host->data_timeout = target_timeout; in sdhci_calc_sw_timeout()
923 if (host->data_timeout) in sdhci_calc_sw_timeout()
924 host->data_timeout += MMC_CMD_TRANSFER_TIME; in sdhci_calc_sw_timeout()
927 static u8 sdhci_calc_timeout(struct sdhci_host *host, struct mmc_command *cmd, in sdhci_calc_timeout() argument
939 * longer to time out, but that's much better than having a too-short in sdhci_calc_timeout()
942 if (host->quirks & SDHCI_QUIRK_BROKEN_TIMEOUT_VAL) in sdhci_calc_timeout()
946 if (cmd == NULL) in sdhci_calc_timeout()
949 data = cmd->data; in sdhci_calc_timeout()
951 if (!data && !cmd->busy_timeout) in sdhci_calc_timeout()
955 target_timeout = sdhci_target_timeout(host, cmd, data); in sdhci_calc_timeout()
963 * (2) host->timeout_clk < 2^16 in sdhci_calc_timeout()
968 current_timeout = (1 << 13) * 1000 / host->timeout_clk; in sdhci_calc_timeout()
977 if (!(host->quirks2 & SDHCI_QUIRK2_DISABLE_HW_TIMEOUT)) in sdhci_calc_timeout()
978 DBG("Too large timeout 0x%x requested for CMD%d!\n", in sdhci_calc_timeout()
979 count, cmd->opcode); in sdhci_calc_timeout()
993 if (host->flags & SDHCI_REQ_USE_DMA) in sdhci_set_transfer_irqs()
994 host->ier = (host->ier & ~pio_irqs) | dma_irqs; in sdhci_set_transfer_irqs()
996 host->ier = (host->ier & ~dma_irqs) | pio_irqs; in sdhci_set_transfer_irqs()
998 if (host->flags & (SDHCI_AUTO_CMD23 | SDHCI_AUTO_CMD12)) in sdhci_set_transfer_irqs()
999 host->ier |= SDHCI_INT_AUTO_CMD_ERR; in sdhci_set_transfer_irqs()
1001 host->ier &= ~SDHCI_INT_AUTO_CMD_ERR; in sdhci_set_transfer_irqs()
1003 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_transfer_irqs()
1004 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_transfer_irqs()
1010 host->ier |= SDHCI_INT_DATA_TIMEOUT; in sdhci_set_data_timeout_irq()
1012 host->ier &= ~SDHCI_INT_DATA_TIMEOUT; in sdhci_set_data_timeout_irq()
1013 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_set_data_timeout_irq()
1014 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_set_data_timeout_irq()
1018 void __sdhci_set_timeout(struct sdhci_host *host, struct mmc_command *cmd) in __sdhci_set_timeout() argument
1021 u8 count = sdhci_calc_timeout(host, cmd, &too_big); in __sdhci_set_timeout()
1024 host->quirks2 & SDHCI_QUIRK2_DISABLE_HW_TIMEOUT) { in __sdhci_set_timeout()
1025 sdhci_calc_sw_timeout(host, cmd); in __sdhci_set_timeout()
1027 } else if (!(host->ier & SDHCI_INT_DATA_TIMEOUT)) { in __sdhci_set_timeout()
1035 static void sdhci_set_timeout(struct sdhci_host *host, struct mmc_command *cmd) in sdhci_set_timeout() argument
1037 if (host->ops->set_timeout) in sdhci_set_timeout()
1038 host->ops->set_timeout(host, cmd); in sdhci_set_timeout()
1040 __sdhci_set_timeout(host, cmd); in sdhci_set_timeout()
1046 WARN_ON(host->data); in sdhci_initialize_data()
1049 BUG_ON(data->blksz * data->blocks > 524288); in sdhci_initialize_data()
1050 BUG_ON(data->blksz > host->mmc->max_blk_size); in sdhci_initialize_data()
1051 BUG_ON(data->blocks > 65535); in sdhci_initialize_data()
1053 host->data = data; in sdhci_initialize_data()
1054 host->data_early = 0; in sdhci_initialize_data()
1055 host->data->bytes_xfered = 0; in sdhci_initialize_data()
1063 SDHCI_MAKE_BLKSZ(host->sdma_boundary, data->blksz), in sdhci_set_block_info()
1066 * For Version 4.10 onwards, if v4 mode is enabled, 32-bit Block Count in sdhci_set_block_info()
1067 * can be supported, in that case 16-bit block count register must be 0. in sdhci_set_block_info()
1069 if (host->version >= SDHCI_SPEC_410 && host->v4_mode && in sdhci_set_block_info()
1070 (host->quirks2 & SDHCI_QUIRK2_USE_32BIT_BLK_CNT)) { in sdhci_set_block_info()
1073 sdhci_writew(host, data->blocks, SDHCI_32BIT_BLK_CNT); in sdhci_set_block_info()
1075 sdhci_writew(host, data->blocks, SDHCI_BLOCK_COUNT); in sdhci_set_block_info()
1079 static void sdhci_prepare_data(struct sdhci_host *host, struct mmc_command *cmd) in sdhci_prepare_data() argument
1081 struct mmc_data *data = cmd->data; in sdhci_prepare_data()
1085 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_prepare_data()
1090 host->flags |= SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
1101 if (host->flags & SDHCI_USE_ADMA) { in sdhci_prepare_data()
1102 if (host->quirks & SDHCI_QUIRK_32BIT_ADMA_SIZE) { in sdhci_prepare_data()
1112 if (host->quirks & SDHCI_QUIRK_32BIT_DMA_SIZE) in sdhci_prepare_data()
1114 if (host->quirks & SDHCI_QUIRK_32BIT_DMA_ADDR) in sdhci_prepare_data()
1119 for_each_sg(data->sg, sg, data->sg_len, i) { in sdhci_prepare_data()
1120 if (sg->length & length_mask) { in sdhci_prepare_data()
1122 sg->length); in sdhci_prepare_data()
1123 host->flags &= ~SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
1126 if (sg->offset & offset_mask) { in sdhci_prepare_data()
1128 host->flags &= ~SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
1135 if (host->flags & SDHCI_REQ_USE_DMA) { in sdhci_prepare_data()
1144 host->flags &= ~SDHCI_REQ_USE_DMA; in sdhci_prepare_data()
1145 } else if (host->flags & SDHCI_USE_ADMA) { in sdhci_prepare_data()
1147 sdhci_set_adma_addr(host, host->adma_addr); in sdhci_prepare_data()
1156 if (!(host->flags & SDHCI_REQ_USE_DMA)) { in sdhci_prepare_data()
1160 if (host->data->flags & MMC_DATA_READ) in sdhci_prepare_data()
1164 sg_miter_start(&host->sg_miter, data->sg, data->sg_len, flags); in sdhci_prepare_data()
1165 host->blocks = data->blocks; in sdhci_prepare_data()
1178 struct mmc_host *mmc = host->mmc; in sdhci_external_dma_init()
1180 host->tx_chan = dma_request_chan(mmc->parent, "tx"); in sdhci_external_dma_init()
1181 if (IS_ERR(host->tx_chan)) { in sdhci_external_dma_init()
1182 ret = PTR_ERR(host->tx_chan); in sdhci_external_dma_init()
1183 if (ret != -EPROBE_DEFER) in sdhci_external_dma_init()
1185 host->tx_chan = NULL; in sdhci_external_dma_init()
1189 host->rx_chan = dma_request_chan(mmc->parent, "rx"); in sdhci_external_dma_init()
1190 if (IS_ERR(host->rx_chan)) { in sdhci_external_dma_init()
1191 if (host->tx_chan) { in sdhci_external_dma_init()
1192 dma_release_channel(host->tx_chan); in sdhci_external_dma_init()
1193 host->tx_chan = NULL; in sdhci_external_dma_init()
1196 ret = PTR_ERR(host->rx_chan); in sdhci_external_dma_init()
1197 if (ret != -EPROBE_DEFER) in sdhci_external_dma_init()
1199 host->rx_chan = NULL; in sdhci_external_dma_init()
1208 return data->flags & MMC_DATA_WRITE ? host->tx_chan : host->rx_chan; in sdhci_external_dma_channel()
1212 struct mmc_command *cmd) in sdhci_external_dma_setup() argument
1217 struct mmc_data *data = cmd->data; in sdhci_external_dma_setup()
1223 if (!host->mapbase) in sdhci_external_dma_setup()
1224 return -EINVAL; in sdhci_external_dma_setup()
1227 cfg.src_addr = host->mapbase + SDHCI_BUFFER; in sdhci_external_dma_setup()
1228 cfg.dst_addr = host->mapbase + SDHCI_BUFFER; in sdhci_external_dma_setup()
1231 cfg.src_maxburst = data->blksz / 4; in sdhci_external_dma_setup()
1232 cfg.dst_maxburst = data->blksz / 4; in sdhci_external_dma_setup()
1235 for (i = 0; i < data->sg_len; i++) { in sdhci_external_dma_setup()
1236 if ((data->sg + i)->length % data->blksz) in sdhci_external_dma_setup()
1237 return -EINVAL; in sdhci_external_dma_setup()
1248 return -EINVAL; in sdhci_external_dma_setup()
1250 dir = data->flags & MMC_DATA_WRITE ? DMA_MEM_TO_DEV : DMA_DEV_TO_MEM; in sdhci_external_dma_setup()
1251 desc = dmaengine_prep_slave_sg(chan, data->sg, data->sg_len, dir, in sdhci_external_dma_setup()
1254 return -EINVAL; in sdhci_external_dma_setup()
1256 desc->callback = NULL; in sdhci_external_dma_setup()
1257 desc->callback_param = NULL; in sdhci_external_dma_setup()
1268 if (host->tx_chan) { in sdhci_external_dma_release()
1269 dma_release_channel(host->tx_chan); in sdhci_external_dma_release()
1270 host->tx_chan = NULL; in sdhci_external_dma_release()
1273 if (host->rx_chan) { in sdhci_external_dma_release()
1274 dma_release_channel(host->rx_chan); in sdhci_external_dma_release()
1275 host->rx_chan = NULL; in sdhci_external_dma_release()
1282 struct mmc_command *cmd) in __sdhci_external_dma_prepare_data() argument
1284 struct mmc_data *data = cmd->data; in __sdhci_external_dma_prepare_data()
1288 host->flags |= SDHCI_REQ_USE_DMA; in __sdhci_external_dma_prepare_data()
1295 struct mmc_command *cmd) in sdhci_external_dma_prepare_data() argument
1297 if (!sdhci_external_dma_setup(host, cmd)) { in sdhci_external_dma_prepare_data()
1298 __sdhci_external_dma_prepare_data(host, cmd); in sdhci_external_dma_prepare_data()
1302 mmc_hostname(host->mmc)); in sdhci_external_dma_prepare_data()
1303 sdhci_prepare_data(host, cmd); in sdhci_external_dma_prepare_data()
1308 struct mmc_command *cmd) in sdhci_external_dma_pre_transfer() argument
1312 if (!cmd->data) in sdhci_external_dma_pre_transfer()
1315 chan = sdhci_external_dma_channel(host, cmd->data); in sdhci_external_dma_pre_transfer()
1324 return -EOPNOTSUPP; in sdhci_external_dma_init()
1332 struct mmc_command *cmd) in sdhci_external_dma_prepare_data() argument
1339 struct mmc_command *cmd) in sdhci_external_dma_pre_transfer() argument
1353 host->use_external_dma = en; in sdhci_switch_external_dma()
1360 return !mrq->sbc && (host->flags & SDHCI_AUTO_CMD12) && in sdhci_auto_cmd12()
1361 !mrq->cap_cmd_during_tfr; in sdhci_auto_cmd12()
1367 return mrq->sbc && (host->flags & SDHCI_AUTO_CMD23); in sdhci_auto_cmd23()
1373 return mrq->sbc && !(host->flags & SDHCI_AUTO_CMD23); in sdhci_manual_cmd23()
1377 struct mmc_command *cmd, in sdhci_auto_cmd_select() argument
1380 bool use_cmd12 = sdhci_auto_cmd12(host, cmd->mrq) && in sdhci_auto_cmd_select()
1381 (cmd->opcode != SD_IO_RW_EXTENDED); in sdhci_auto_cmd_select()
1382 bool use_cmd23 = sdhci_auto_cmd23(host, cmd->mrq); in sdhci_auto_cmd_select()
1386 * In case of Version 4.10 or later, use of 'Auto CMD Auto in sdhci_auto_cmd_select()
1389 * here because some controllers (e.g sdhci-of-dwmshc) expect it. in sdhci_auto_cmd_select()
1391 if (host->version >= SDHCI_SPEC_410 && host->v4_mode && in sdhci_auto_cmd_select()
1407 * on successful completion (so no Auto-CMD12). in sdhci_auto_cmd_select()
1416 struct mmc_command *cmd) in sdhci_set_transfer_mode() argument
1419 struct mmc_data *data = cmd->data; in sdhci_set_transfer_mode()
1422 if (host->quirks2 & in sdhci_set_transfer_mode()
1425 if (cmd->opcode != MMC_SEND_TUNING_BLOCK_HS200) in sdhci_set_transfer_mode()
1428 /* clear Auto CMD settings for no data CMDs */ in sdhci_set_transfer_mode()
1436 WARN_ON(!host->data); in sdhci_set_transfer_mode()
1438 if (!(host->quirks2 & SDHCI_QUIRK2_SUPPORT_SINGLE)) in sdhci_set_transfer_mode()
1441 if (mmc_op_multi(cmd->opcode) || data->blocks > 1) { in sdhci_set_transfer_mode()
1443 sdhci_auto_cmd_select(host, cmd, &mode); in sdhci_set_transfer_mode()
1444 if (sdhci_auto_cmd23(host, cmd->mrq)) in sdhci_set_transfer_mode()
1445 sdhci_writel(host, cmd->mrq->sbc->arg, SDHCI_ARGUMENT2); in sdhci_set_transfer_mode()
1448 if (data->flags & MMC_DATA_READ) in sdhci_set_transfer_mode()
1450 if (host->flags & SDHCI_REQ_USE_DMA) in sdhci_set_transfer_mode()
1458 return (!(host->flags & SDHCI_DEVICE_DEAD) && in sdhci_needs_reset()
1459 ((mrq->cmd && mrq->cmd->error) || in sdhci_needs_reset()
1460 (mrq->sbc && mrq->sbc->error) || in sdhci_needs_reset()
1461 (mrq->data && mrq->data->stop && mrq->data->stop->error) || in sdhci_needs_reset()
1462 (host->quirks & SDHCI_QUIRK_RESET_AFTER_REQUEST))); in sdhci_needs_reset()
1470 if (host->mrqs_done[i] == mrq) { in sdhci_set_mrq_done()
1477 if (!host->mrqs_done[i]) { in sdhci_set_mrq_done()
1478 host->mrqs_done[i] = mrq; in sdhci_set_mrq_done()
1488 if (host->cmd && host->cmd->mrq == mrq) in __sdhci_finish_mrq()
1489 host->cmd = NULL; in __sdhci_finish_mrq()
1491 if (host->data_cmd && host->data_cmd->mrq == mrq) in __sdhci_finish_mrq()
1492 host->data_cmd = NULL; in __sdhci_finish_mrq()
1494 if (host->deferred_cmd && host->deferred_cmd->mrq == mrq) in __sdhci_finish_mrq()
1495 host->deferred_cmd = NULL; in __sdhci_finish_mrq()
1497 if (host->data && host->data->mrq == mrq) in __sdhci_finish_mrq()
1498 host->data = NULL; in __sdhci_finish_mrq()
1501 host->pending_reset = true; in __sdhci_finish_mrq()
1515 queue_work(host->complete_wq, &host->complete_work); in sdhci_finish_mrq()
1520 struct mmc_command *data_cmd = host->data_cmd; in __sdhci_finish_data()
1521 struct mmc_data *data = host->data; in __sdhci_finish_data()
1523 host->data = NULL; in __sdhci_finish_data()
1524 host->data_cmd = NULL; in __sdhci_finish_data()
1530 if (data->error) { in __sdhci_finish_data()
1531 if (!host->cmd || host->cmd == data_cmd) in __sdhci_finish_data()
1536 if ((host->flags & (SDHCI_REQ_USE_DMA | SDHCI_USE_ADMA)) == in __sdhci_finish_data()
1547 if (data->error) in __sdhci_finish_data()
1548 data->bytes_xfered = 0; in __sdhci_finish_data()
1550 data->bytes_xfered = data->blksz * data->blocks; in __sdhci_finish_data()
1553 * Need to send CMD12 if - in __sdhci_finish_data()
1554 * a) open-ended multiblock transfer not using auto CMD12 (no CMD23) in __sdhci_finish_data()
1557 if (data->stop && in __sdhci_finish_data()
1558 ((!data->mrq->sbc && !sdhci_auto_cmd12(host, data->mrq)) || in __sdhci_finish_data()
1559 data->error)) { in __sdhci_finish_data()
1565 if (data->mrq->cap_cmd_during_tfr) { in __sdhci_finish_data()
1566 __sdhci_finish_mrq(host, data->mrq); in __sdhci_finish_data()
1569 host->cmd = NULL; in __sdhci_finish_data()
1570 if (!sdhci_send_command(host, data->stop)) { in __sdhci_finish_data()
1576 data->stop->error = -EIO; in __sdhci_finish_data()
1577 __sdhci_finish_mrq(host, data->mrq); in __sdhci_finish_data()
1579 WARN_ON(host->deferred_cmd); in __sdhci_finish_data()
1580 host->deferred_cmd = data->stop; in __sdhci_finish_data()
1585 __sdhci_finish_mrq(host, data->mrq); in __sdhci_finish_data()
1594 static bool sdhci_send_command(struct sdhci_host *host, struct mmc_command *cmd) in sdhci_send_command() argument
1600 WARN_ON(host->cmd); in sdhci_send_command()
1603 cmd->error = 0; in sdhci_send_command()
1605 if ((host->quirks2 & SDHCI_QUIRK2_STOP_WITH_TC) && in sdhci_send_command()
1606 cmd->opcode == MMC_STOP_TRANSMISSION) in sdhci_send_command()
1607 cmd->flags |= MMC_RSP_BUSY; in sdhci_send_command()
1610 if (sdhci_data_line_cmd(cmd)) in sdhci_send_command()
1615 if (cmd->mrq->data && (cmd == cmd->mrq->data->stop)) in sdhci_send_command()
1621 host->cmd = cmd; in sdhci_send_command()
1622 host->data_timeout = 0; in sdhci_send_command()
1623 if (sdhci_data_line_cmd(cmd)) { in sdhci_send_command()
1624 WARN_ON(host->data_cmd); in sdhci_send_command()
1625 host->data_cmd = cmd; in sdhci_send_command()
1626 sdhci_set_timeout(host, cmd); in sdhci_send_command()
1629 if (cmd->data) { in sdhci_send_command()
1630 if (host->use_external_dma) in sdhci_send_command()
1631 sdhci_external_dma_prepare_data(host, cmd); in sdhci_send_command()
1633 sdhci_prepare_data(host, cmd); in sdhci_send_command()
1636 sdhci_writel(host, cmd->arg, SDHCI_ARGUMENT); in sdhci_send_command()
1638 sdhci_set_transfer_mode(host, cmd); in sdhci_send_command()
1640 if ((cmd->flags & MMC_RSP_136) && (cmd->flags & MMC_RSP_BUSY)) { in sdhci_send_command()
1643 * This does not happen in practice because 136-bit response in sdhci_send_command()
1647 cmd->flags &= ~MMC_RSP_BUSY; in sdhci_send_command()
1650 if (!(cmd->flags & MMC_RSP_PRESENT)) in sdhci_send_command()
1652 else if (cmd->flags & MMC_RSP_136) in sdhci_send_command()
1654 else if (cmd->flags & MMC_RSP_BUSY) in sdhci_send_command()
1659 if (cmd->flags & MMC_RSP_CRC) in sdhci_send_command()
1661 if (cmd->flags & MMC_RSP_OPCODE) in sdhci_send_command()
1665 if (cmd->data || cmd->opcode == MMC_SEND_TUNING_BLOCK || in sdhci_send_command()
1666 cmd->opcode == MMC_SEND_TUNING_BLOCK_HS200) in sdhci_send_command()
1670 if (host->data_timeout) in sdhci_send_command()
1671 timeout += nsecs_to_jiffies(host->data_timeout); in sdhci_send_command()
1672 else if (!cmd->data && cmd->busy_timeout > 9000) in sdhci_send_command()
1673 timeout += DIV_ROUND_UP(cmd->busy_timeout, 1000) * HZ + HZ; in sdhci_send_command()
1676 sdhci_mod_timer(host, cmd->mrq, timeout); in sdhci_send_command()
1678 if (host->use_external_dma) in sdhci_send_command()
1679 sdhci_external_dma_pre_transfer(host, cmd); in sdhci_send_command()
1681 sdhci_writew(host, SDHCI_MAKE_CMD(cmd->opcode, flags), SDHCI_COMMAND); in sdhci_send_command()
1687 struct mmc_command *cmd, bool present) in sdhci_present_error() argument
1689 if (!present || host->flags & SDHCI_DEVICE_DEAD) { in sdhci_present_error()
1690 cmd->error = -ENOMEDIUM; in sdhci_present_error()
1698 struct mmc_command *cmd, in sdhci_send_command_retry() argument
1700 __releases(host->lock) in sdhci_send_command_retry()
1701 __acquires(host->lock) in sdhci_send_command_retry()
1703 struct mmc_command *deferred_cmd = host->deferred_cmd; in sdhci_send_command_retry()
1707 while (!sdhci_send_command(host, cmd)) { in sdhci_send_command_retry()
1708 if (!timeout--) { in sdhci_send_command_retry()
1710 mmc_hostname(host->mmc)); in sdhci_send_command_retry()
1712 cmd->error = -EIO; in sdhci_send_command_retry()
1716 spin_unlock_irqrestore(&host->lock, flags); in sdhci_send_command_retry()
1720 present = host->mmc->ops->get_cd(host->mmc); in sdhci_send_command_retry()
1722 spin_lock_irqsave(&host->lock, flags); in sdhci_send_command_retry()
1725 if (cmd == deferred_cmd && cmd != host->deferred_cmd) in sdhci_send_command_retry()
1728 if (sdhci_present_error(host, cmd, present)) in sdhci_send_command_retry()
1732 if (cmd == host->deferred_cmd) in sdhci_send_command_retry()
1733 host->deferred_cmd = NULL; in sdhci_send_command_retry()
1738 static void sdhci_read_rsp_136(struct sdhci_host *host, struct mmc_command *cmd) in sdhci_read_rsp_136() argument
1743 reg = SDHCI_RESPONSE + (3 - i) * 4; in sdhci_read_rsp_136()
1744 cmd->resp[i] = sdhci_readl(host, reg); in sdhci_read_rsp_136()
1747 if (host->quirks2 & SDHCI_QUIRK2_RSP_136_HAS_CRC) in sdhci_read_rsp_136()
1752 cmd->resp[i] <<= 8; in sdhci_read_rsp_136()
1754 cmd->resp[i] |= cmd->resp[i + 1] >> 24; in sdhci_read_rsp_136()
1760 struct mmc_command *cmd = host->cmd; in sdhci_finish_command() local
1762 host->cmd = NULL; in sdhci_finish_command()
1764 if (cmd->flags & MMC_RSP_PRESENT) { in sdhci_finish_command()
1765 if (cmd->flags & MMC_RSP_136) { in sdhci_finish_command()
1766 sdhci_read_rsp_136(host, cmd); in sdhci_finish_command()
1768 cmd->resp[0] = sdhci_readl(host, SDHCI_RESPONSE); in sdhci_finish_command()
1772 if (cmd->mrq->cap_cmd_during_tfr && cmd == cmd->mrq->cmd) in sdhci_finish_command()
1773 mmc_command_done(host->mmc, cmd->mrq); in sdhci_finish_command()
1785 if (cmd->flags & MMC_RSP_BUSY) { in sdhci_finish_command()
1786 if (cmd->data) { in sdhci_finish_command()
1788 } else if (!(host->quirks & SDHCI_QUIRK_NO_BUSY_IRQ) && in sdhci_finish_command()
1789 cmd == host->data_cmd) { in sdhci_finish_command()
1796 if (cmd == cmd->mrq->sbc) { in sdhci_finish_command()
1797 if (!sdhci_send_command(host, cmd->mrq->cmd)) { in sdhci_finish_command()
1798 WARN_ON(host->deferred_cmd); in sdhci_finish_command()
1799 host->deferred_cmd = cmd->mrq->cmd; in sdhci_finish_command()
1804 if (host->data && host->data_early) in sdhci_finish_command()
1807 if (!cmd->data) in sdhci_finish_command()
1808 __sdhci_finish_mrq(host, cmd->mrq); in sdhci_finish_command()
1816 switch (host->timing) { in sdhci_get_preset_value()
1842 pr_warn("%s: Invalid UHS-I mode selected\n", in sdhci_get_preset_value()
1843 mmc_hostname(host->mmc)); in sdhci_get_preset_value()
1858 if (host->version >= SDHCI_SPEC_300) { in sdhci_calc_clk()
1859 if (host->preset_enabled) { in sdhci_calc_clk()
1865 if (host->clk_mul && in sdhci_calc_clk()
1869 clk_mul = host->clk_mul; in sdhci_calc_clk()
1880 if (host->clk_mul) { in sdhci_calc_clk()
1882 if ((host->max_clk * host->clk_mul / div) in sdhci_calc_clk()
1886 if ((host->max_clk * host->clk_mul / div) <= clock) { in sdhci_calc_clk()
1893 clk_mul = host->clk_mul; in sdhci_calc_clk()
1894 div--; in sdhci_calc_clk()
1904 if (!host->clk_mul || switch_base_clk) { in sdhci_calc_clk()
1906 if (host->max_clk <= clock) in sdhci_calc_clk()
1911 if ((host->max_clk / div) <= clock) in sdhci_calc_clk()
1917 if ((host->quirks2 & SDHCI_QUIRK2_CLOCK_DIV_ZERO_BROKEN) in sdhci_calc_clk()
1918 && !div && host->max_clk <= 25000000) in sdhci_calc_clk()
1924 if ((host->max_clk / div) <= clock) in sdhci_calc_clk()
1933 *actual_clock = (host->max_clk * clk_mul) / real_div; in sdhci_calc_clk()
1959 mmc_hostname(host->mmc)); in sdhci_enable_clk()
1966 if (host->version >= SDHCI_SPEC_410 && host->v4_mode) { in sdhci_enable_clk()
1981 mmc_hostname(host->mmc)); in sdhci_enable_clk()
1998 host->mmc->actual_clock = 0; in sdhci_set_clock()
2005 clk = sdhci_calc_clk(host, clock, &host->mmc->actual_clock); in sdhci_set_clock()
2013 struct mmc_host *mmc = host->mmc; in sdhci_set_power_reg()
2015 mmc_regulator_set_ocr(mmc, mmc->supply.vmmc, vdd); in sdhci_set_power_reg()
2056 mmc_hostname(host->mmc), vdd); in sdhci_set_power_noreg()
2061 if (host->pwr == pwr) in sdhci_set_power_noreg()
2064 host->pwr = pwr; in sdhci_set_power_noreg()
2068 if (host->quirks2 & SDHCI_QUIRK2_CARD_ON_NEEDS_BUS_ON) in sdhci_set_power_noreg()
2075 if (!(host->quirks & SDHCI_QUIRK_SINGLE_POWER_WRITE)) in sdhci_set_power_noreg()
2083 if (host->quirks & SDHCI_QUIRK_NO_SIMULT_VDD_AND_POWER) in sdhci_set_power_noreg()
2090 if (host->quirks2 & SDHCI_QUIRK2_CARD_ON_NEEDS_BUS_ON) in sdhci_set_power_noreg()
2097 if (host->quirks & SDHCI_QUIRK_DELAY_AFTER_POWER) in sdhci_set_power_noreg()
2106 if (IS_ERR(host->mmc->supply.vmmc)) in sdhci_set_power()
2123 if (!IS_ERR(host->mmc->supply.vmmc)) { in sdhci_set_power_and_bus_voltage()
2124 struct mmc_host *mmc = host->mmc; in sdhci_set_power_and_bus_voltage()
2126 mmc_regulator_set_ocr(mmc, mmc->supply.vmmc, vdd); in sdhci_set_power_and_bus_voltage()
2141 struct mmc_command *cmd; in sdhci_request() local
2146 present = mmc->ops->get_cd(mmc); in sdhci_request()
2148 spin_lock_irqsave(&host->lock, flags); in sdhci_request()
2152 if (sdhci_present_error(host, mrq->cmd, present)) in sdhci_request()
2155 cmd = sdhci_manual_cmd23(host, mrq) ? mrq->sbc : mrq->cmd; in sdhci_request()
2157 if (!sdhci_send_command_retry(host, cmd, flags)) in sdhci_request()
2160 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request()
2166 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request()
2173 struct mmc_command *cmd; in sdhci_request_atomic() local
2177 spin_lock_irqsave(&host->lock, flags); in sdhci_request_atomic()
2179 if (sdhci_present_error(host, mrq->cmd, true)) { in sdhci_request_atomic()
2184 cmd = sdhci_manual_cmd23(host, mrq) ? mrq->sbc : mrq->cmd; in sdhci_request_atomic()
2190 * again in non-atomic context. So we should not finish this request in sdhci_request_atomic()
2193 if (!sdhci_send_command(host, cmd)) in sdhci_request_atomic()
2194 ret = -EBUSY; in sdhci_request_atomic()
2199 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_atomic()
2213 if (host->mmc->caps & MMC_CAP_8_BIT_DATA) in sdhci_set_bus_width()
2244 ctrl_2 |= SDHCI_CTRL_HS400; /* Non-standard */ in sdhci_set_uhs_signaling()
2254 if (ios->power_mode == MMC_POWER_UNDEFINED) in sdhci_set_ios()
2257 if (host->flags & SDHCI_DEVICE_DEAD) { in sdhci_set_ios()
2258 if (!IS_ERR(mmc->supply.vmmc) && in sdhci_set_ios()
2259 ios->power_mode == MMC_POWER_OFF) in sdhci_set_ios()
2260 mmc_regulator_set_ocr(mmc, mmc->supply.vmmc, 0); in sdhci_set_ios()
2268 if (ios->power_mode == MMC_POWER_OFF) { in sdhci_set_ios()
2273 if (host->version >= SDHCI_SPEC_300 && in sdhci_set_ios()
2274 (ios->power_mode == MMC_POWER_UP) && in sdhci_set_ios()
2275 !(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN)) in sdhci_set_ios()
2278 if (!ios->clock || ios->clock != host->clock) { in sdhci_set_ios()
2279 host->ops->set_clock(host, ios->clock); in sdhci_set_ios()
2280 host->clock = ios->clock; in sdhci_set_ios()
2282 if (host->quirks & SDHCI_QUIRK_DATA_TIMEOUT_USES_SDCLK && in sdhci_set_ios()
2283 host->clock) { in sdhci_set_ios()
2284 host->timeout_clk = host->mmc->actual_clock ? in sdhci_set_ios()
2285 host->mmc->actual_clock / 1000 : in sdhci_set_ios()
2286 host->clock / 1000; in sdhci_set_ios()
2287 host->mmc->max_busy_timeout = in sdhci_set_ios()
2288 host->ops->get_max_timeout_count ? in sdhci_set_ios()
2289 host->ops->get_max_timeout_count(host) : in sdhci_set_ios()
2291 host->mmc->max_busy_timeout /= host->timeout_clk; in sdhci_set_ios()
2295 if (host->ops->set_power) in sdhci_set_ios()
2296 host->ops->set_power(host, ios->power_mode, ios->vdd); in sdhci_set_ios()
2298 sdhci_set_power(host, ios->power_mode, ios->vdd); in sdhci_set_ios()
2300 if (host->ops->platform_send_init_74_clocks) in sdhci_set_ios()
2301 host->ops->platform_send_init_74_clocks(host, ios->power_mode); in sdhci_set_ios()
2303 host->ops->set_bus_width(host, ios->bus_width); in sdhci_set_ios()
2307 if (!(host->quirks & SDHCI_QUIRK_NO_HISPD_BIT)) { in sdhci_set_ios()
2308 if (ios->timing == MMC_TIMING_SD_HS || in sdhci_set_ios()
2309 ios->timing == MMC_TIMING_MMC_HS || in sdhci_set_ios()
2310 ios->timing == MMC_TIMING_MMC_HS400 || in sdhci_set_ios()
2311 ios->timing == MMC_TIMING_MMC_HS200 || in sdhci_set_ios()
2312 ios->timing == MMC_TIMING_MMC_DDR52 || in sdhci_set_ios()
2313 ios->timing == MMC_TIMING_UHS_SDR50 || in sdhci_set_ios()
2314 ios->timing == MMC_TIMING_UHS_SDR104 || in sdhci_set_ios()
2315 ios->timing == MMC_TIMING_UHS_DDR50 || in sdhci_set_ios()
2316 ios->timing == MMC_TIMING_UHS_SDR25) in sdhci_set_ios()
2322 if (host->version >= SDHCI_SPEC_300) { in sdhci_set_ios()
2325 if (!host->preset_enabled) { in sdhci_set_ios()
2333 if (ios->drv_type == MMC_SET_DRIVER_TYPE_A) in sdhci_set_ios()
2335 else if (ios->drv_type == MMC_SET_DRIVER_TYPE_B) in sdhci_set_ios()
2337 else if (ios->drv_type == MMC_SET_DRIVER_TYPE_C) in sdhci_set_ios()
2339 else if (ios->drv_type == MMC_SET_DRIVER_TYPE_D) in sdhci_set_ios()
2363 /* Re-enable SD Clock */ in sdhci_set_ios()
2364 host->ops->set_clock(host, host->clock); in sdhci_set_ios()
2372 host->ops->set_uhs_signaling(host, ios->timing); in sdhci_set_ios()
2373 host->timing = ios->timing; in sdhci_set_ios()
2375 if (!(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN) && in sdhci_set_ios()
2376 ((ios->timing == MMC_TIMING_UHS_SDR12) || in sdhci_set_ios()
2377 (ios->timing == MMC_TIMING_UHS_SDR25) || in sdhci_set_ios()
2378 (ios->timing == MMC_TIMING_UHS_SDR50) || in sdhci_set_ios()
2379 (ios->timing == MMC_TIMING_UHS_SDR104) || in sdhci_set_ios()
2380 (ios->timing == MMC_TIMING_UHS_DDR50) || in sdhci_set_ios()
2381 (ios->timing == MMC_TIMING_MMC_DDR52))) { in sdhci_set_ios()
2386 ios->drv_type = FIELD_GET(SDHCI_PRESET_DRV_MASK, in sdhci_set_ios()
2390 /* Re-enable SD Clock */ in sdhci_set_ios()
2391 host->ops->set_clock(host, host->clock); in sdhci_set_ios()
2400 if (host->quirks & SDHCI_QUIRK_RESET_CMD_DATA_ON_IOS) in sdhci_set_ios()
2410 if (host->flags & SDHCI_DEVICE_DEAD) in sdhci_get_cd()
2414 if (!mmc_card_is_removable(host->mmc)) in sdhci_get_cd()
2425 if (host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) in sdhci_get_cd()
2437 spin_lock_irqsave(&host->lock, flags); in sdhci_check_ro()
2439 if (host->flags & SDHCI_DEVICE_DEAD) in sdhci_check_ro()
2441 else if (host->ops->get_ro) in sdhci_check_ro()
2442 is_readonly = host->ops->get_ro(host); in sdhci_check_ro()
2443 else if (mmc_can_gpio_ro(host->mmc)) in sdhci_check_ro()
2444 is_readonly = mmc_gpio_get_ro(host->mmc); in sdhci_check_ro()
2449 spin_unlock_irqrestore(&host->lock, flags); in sdhci_check_ro()
2451 /* This quirk needs to be replaced by a callback-function later */ in sdhci_check_ro()
2452 return host->quirks & SDHCI_QUIRK_INVERTED_WRITE_PROTECT ? in sdhci_check_ro()
2463 if (!(host->quirks & SDHCI_QUIRK_UNSTABLE_RO_DETECT)) in sdhci_get_ro()
2481 if (host->ops && host->ops->hw_reset) in sdhci_hw_reset()
2482 host->ops->hw_reset(host); in sdhci_hw_reset()
2487 if (!(host->flags & SDHCI_DEVICE_DEAD)) { in sdhci_enable_sdio_irq_nolock()
2489 host->ier |= SDHCI_INT_CARD_INT; in sdhci_enable_sdio_irq_nolock()
2491 host->ier &= ~SDHCI_INT_CARD_INT; in sdhci_enable_sdio_irq_nolock()
2493 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_enable_sdio_irq_nolock()
2494 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_enable_sdio_irq_nolock()
2504 pm_runtime_get_noresume(host->mmc->parent); in sdhci_enable_sdio_irq()
2506 spin_lock_irqsave(&host->lock, flags); in sdhci_enable_sdio_irq()
2508 spin_unlock_irqrestore(&host->lock, flags); in sdhci_enable_sdio_irq()
2511 pm_runtime_put_noidle(host->mmc->parent); in sdhci_enable_sdio_irq()
2520 spin_lock_irqsave(&host->lock, flags); in sdhci_ack_sdio_irq()
2522 spin_unlock_irqrestore(&host->lock, flags); in sdhci_ack_sdio_irq()
2536 if (host->version < SDHCI_SPEC_300) in sdhci_start_signal_voltage_switch()
2541 switch (ios->signal_voltage) { in sdhci_start_signal_voltage_switch()
2543 if (!(host->flags & SDHCI_SIGNALING_330)) in sdhci_start_signal_voltage_switch()
2544 return -EINVAL; in sdhci_start_signal_voltage_switch()
2549 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_start_signal_voltage_switch()
2554 return -EIO; in sdhci_start_signal_voltage_switch()
2568 return -EAGAIN; in sdhci_start_signal_voltage_switch()
2570 if (!(host->flags & SDHCI_SIGNALING_180)) in sdhci_start_signal_voltage_switch()
2571 return -EINVAL; in sdhci_start_signal_voltage_switch()
2572 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_start_signal_voltage_switch()
2577 return -EIO; in sdhci_start_signal_voltage_switch()
2589 if (host->ops->voltage_switch) in sdhci_start_signal_voltage_switch()
2590 host->ops->voltage_switch(host); in sdhci_start_signal_voltage_switch()
2600 return -EAGAIN; in sdhci_start_signal_voltage_switch()
2602 if (!(host->flags & SDHCI_SIGNALING_120)) in sdhci_start_signal_voltage_switch()
2603 return -EINVAL; in sdhci_start_signal_voltage_switch()
2604 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_start_signal_voltage_switch()
2609 return -EIO; in sdhci_start_signal_voltage_switch()
2625 /* Check whether DAT[0] is 0 */ in sdhci_card_busy()
2636 spin_lock_irqsave(&host->lock, flags); in sdhci_prepare_hs400_tuning()
2637 host->flags |= SDHCI_HS400_TUNING; in sdhci_prepare_hs400_tuning()
2638 spin_unlock_irqrestore(&host->lock, flags); in sdhci_prepare_hs400_tuning()
2649 if (host->quirks2 & SDHCI_QUIRK2_TUNING_WORK_AROUND) in sdhci_start_tuning()
2670 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_end_tuning()
2671 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_end_tuning()
2695 mmc_abort_tuning(host->mmc, opcode); in sdhci_abort_tuning()
2702 * automatically) so mmc_send_tuning() will return -EIO. Also the tuning command
2708 struct mmc_host *mmc = host->mmc; in sdhci_send_tuning()
2709 struct mmc_command cmd = {}; in sdhci_send_tuning() local
2712 u32 b = host->sdma_boundary; in sdhci_send_tuning()
2714 spin_lock_irqsave(&host->lock, flags); in sdhci_send_tuning()
2716 cmd.opcode = opcode; in sdhci_send_tuning()
2717 cmd.flags = MMC_RSP_R1 | MMC_CMD_ADTC; in sdhci_send_tuning()
2718 cmd.mrq = &mrq; in sdhci_send_tuning()
2720 mrq.cmd = &cmd; in sdhci_send_tuning()
2726 if (cmd.opcode == MMC_SEND_TUNING_BLOCK_HS200 && in sdhci_send_tuning()
2727 mmc->ios.bus_width == MMC_BUS_WIDTH_8) in sdhci_send_tuning()
2740 if (!sdhci_send_command_retry(host, &cmd, flags)) { in sdhci_send_tuning()
2741 spin_unlock_irqrestore(&host->lock, flags); in sdhci_send_tuning()
2742 host->tuning_done = 0; in sdhci_send_tuning()
2746 host->cmd = NULL; in sdhci_send_tuning()
2750 host->tuning_done = 0; in sdhci_send_tuning()
2752 spin_unlock_irqrestore(&host->lock, flags); in sdhci_send_tuning()
2755 wait_event_timeout(host->buf_ready_int, (host->tuning_done == 1), in sdhci_send_tuning()
2769 for (i = 0; i < host->tuning_loop_count; i++) { in __sdhci_execute_tuning()
2774 if (!host->tuning_done) { in __sdhci_execute_tuning()
2776 mmc_hostname(host->mmc)); in __sdhci_execute_tuning()
2778 return -ETIMEDOUT; in __sdhci_execute_tuning()
2782 if (host->tuning_delay > 0) in __sdhci_execute_tuning()
2783 mdelay(host->tuning_delay); in __sdhci_execute_tuning()
2795 mmc_hostname(host->mmc)); in __sdhci_execute_tuning()
2797 return -EAGAIN; in __sdhci_execute_tuning()
2807 hs400_tuning = host->flags & SDHCI_HS400_TUNING; in sdhci_execute_tuning()
2809 if (host->tuning_mode == SDHCI_TUNING_MODE_1) in sdhci_execute_tuning()
2810 tuning_count = host->tuning_count; in sdhci_execute_tuning()
2819 switch (host->timing) { in sdhci_execute_tuning()
2822 err = -EINVAL; in sdhci_execute_tuning()
2827 * Periodic re-tuning for HS400 is not expected to be needed, so in sdhci_execute_tuning()
2839 if (host->flags & SDHCI_SDR50_NEEDS_TUNING) in sdhci_execute_tuning()
2847 if (host->ops->platform_execute_tuning) { in sdhci_execute_tuning()
2848 err = host->ops->platform_execute_tuning(host, opcode); in sdhci_execute_tuning()
2852 host->mmc->retune_period = tuning_count; in sdhci_execute_tuning()
2854 if (host->tuning_delay < 0) in sdhci_execute_tuning()
2855 host->tuning_delay = opcode == MMC_SEND_TUNING_BLOCK; in sdhci_execute_tuning()
2859 host->tuning_err = __sdhci_execute_tuning(host, opcode); in sdhci_execute_tuning()
2863 host->flags &= ~SDHCI_HS400_TUNING; in sdhci_execute_tuning()
2872 if (host->version < SDHCI_SPEC_300) in sdhci_enable_preset_value()
2879 if (host->preset_enabled != enable) { in sdhci_enable_preset_value()
2890 host->flags |= SDHCI_PV_ENABLED; in sdhci_enable_preset_value()
2892 host->flags &= ~SDHCI_PV_ENABLED; in sdhci_enable_preset_value()
2894 host->preset_enabled = enable; in sdhci_enable_preset_value()
2902 struct mmc_data *data = mrq->data; in sdhci_post_req()
2904 if (data->host_cookie != COOKIE_UNMAPPED) in sdhci_post_req()
2905 dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len, in sdhci_post_req()
2908 data->host_cookie = COOKIE_UNMAPPED; in sdhci_post_req()
2915 mrq->data->host_cookie = COOKIE_UNMAPPED; in sdhci_pre_req()
2918 * No pre-mapping in the pre hook if we're using the bounce buffer, in sdhci_pre_req()
2922 if (host->flags & SDHCI_REQ_USE_DMA && !host->bounce_buffer) in sdhci_pre_req()
2923 sdhci_pre_dma_transfer(host, mrq->data, COOKIE_PRE_MAPPED); in sdhci_pre_req()
2928 if (host->data_cmd) { in sdhci_error_out_mrqs()
2929 host->data_cmd->error = err; in sdhci_error_out_mrqs()
2930 sdhci_finish_mrq(host, host->data_cmd->mrq); in sdhci_error_out_mrqs()
2933 if (host->cmd) { in sdhci_error_out_mrqs()
2934 host->cmd->error = err; in sdhci_error_out_mrqs()
2935 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_error_out_mrqs()
2946 if (host->ops->card_event) in sdhci_card_event()
2947 host->ops->card_event(host); in sdhci_card_event()
2949 present = mmc->ops->get_cd(mmc); in sdhci_card_event()
2951 spin_lock_irqsave(&host->lock, flags); in sdhci_card_event()
2956 mmc_hostname(host->mmc)); in sdhci_card_event()
2958 mmc_hostname(host->mmc)); in sdhci_card_event()
2963 sdhci_error_out_mrqs(host, -ENOMEDIUM); in sdhci_card_event()
2966 spin_unlock_irqrestore(&host->lock, flags); in sdhci_card_event()
2998 spin_lock_irqsave(&host->lock, flags); in sdhci_request_done()
3001 mrq = host->mrqs_done[i]; in sdhci_request_done()
3007 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
3019 * also be in mrqs_done, otherwise host->cmd and host->data_cmd in sdhci_request_done()
3022 if (host->cmd || host->data_cmd) { in sdhci_request_done()
3023 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
3028 if (host->quirks & SDHCI_QUIRK_CLOCK_BEFORE_RESET) in sdhci_request_done()
3030 host->ops->set_clock(host, host->clock); in sdhci_request_done()
3039 host->pending_reset = false; in sdhci_request_done()
3047 if (host->flags & SDHCI_REQ_USE_DMA) { in sdhci_request_done()
3048 struct mmc_data *data = mrq->data; in sdhci_request_done()
3050 if (host->use_external_dma && data && in sdhci_request_done()
3051 (mrq->cmd->error || data->error)) { in sdhci_request_done()
3054 host->mrqs_done[i] = NULL; in sdhci_request_done()
3055 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
3057 spin_lock_irqsave(&host->lock, flags); in sdhci_request_done()
3061 if (data && data->host_cookie == COOKIE_MAPPED) { in sdhci_request_done()
3062 if (host->bounce_buffer) { in sdhci_request_done()
3068 unsigned int length = data->bytes_xfered; in sdhci_request_done()
3070 if (length > host->bounce_buffer_size) { in sdhci_request_done()
3072 mmc_hostname(host->mmc), in sdhci_request_done()
3073 host->bounce_buffer_size, in sdhci_request_done()
3074 data->bytes_xfered); in sdhci_request_done()
3076 length = host->bounce_buffer_size; in sdhci_request_done()
3079 host->mmc->parent, in sdhci_request_done()
3080 host->bounce_addr, in sdhci_request_done()
3081 host->bounce_buffer_size, in sdhci_request_done()
3083 sg_copy_from_buffer(data->sg, in sdhci_request_done()
3084 data->sg_len, in sdhci_request_done()
3085 host->bounce_buffer, in sdhci_request_done()
3090 host->mmc->parent, in sdhci_request_done()
3091 host->bounce_addr, in sdhci_request_done()
3092 host->bounce_buffer_size, in sdhci_request_done()
3097 dma_unmap_sg(mmc_dev(host->mmc), data->sg, in sdhci_request_done()
3098 data->sg_len, in sdhci_request_done()
3101 data->host_cookie = COOKIE_UNMAPPED; in sdhci_request_done()
3105 host->mrqs_done[i] = NULL; in sdhci_request_done()
3107 spin_unlock_irqrestore(&host->lock, flags); in sdhci_request_done()
3109 if (host->ops->request_done) in sdhci_request_done()
3110 host->ops->request_done(host, mrq); in sdhci_request_done()
3112 mmc_request_done(host->mmc, mrq); in sdhci_request_done()
3133 spin_lock_irqsave(&host->lock, flags); in sdhci_timeout_timer()
3135 if (host->cmd && !sdhci_data_line_cmd(host->cmd)) { in sdhci_timeout_timer()
3136 pr_err("%s: Timeout waiting for hardware cmd interrupt.\n", in sdhci_timeout_timer()
3137 mmc_hostname(host->mmc)); in sdhci_timeout_timer()
3140 host->cmd->error = -ETIMEDOUT; in sdhci_timeout_timer()
3141 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_timeout_timer()
3144 spin_unlock_irqrestore(&host->lock, flags); in sdhci_timeout_timer()
3154 spin_lock_irqsave(&host->lock, flags); in sdhci_timeout_data_timer()
3156 if (host->data || host->data_cmd || in sdhci_timeout_data_timer()
3157 (host->cmd && sdhci_data_line_cmd(host->cmd))) { in sdhci_timeout_data_timer()
3159 mmc_hostname(host->mmc)); in sdhci_timeout_data_timer()
3162 if (host->data) { in sdhci_timeout_data_timer()
3163 host->data->error = -ETIMEDOUT; in sdhci_timeout_data_timer()
3165 queue_work(host->complete_wq, &host->complete_work); in sdhci_timeout_data_timer()
3166 } else if (host->data_cmd) { in sdhci_timeout_data_timer()
3167 host->data_cmd->error = -ETIMEDOUT; in sdhci_timeout_data_timer()
3168 sdhci_finish_mrq(host, host->data_cmd->mrq); in sdhci_timeout_data_timer()
3170 host->cmd->error = -ETIMEDOUT; in sdhci_timeout_data_timer()
3171 sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_timeout_data_timer()
3175 spin_unlock_irqrestore(&host->lock, flags); in sdhci_timeout_data_timer()
3186 /* Handle auto-CMD12 error */ in sdhci_cmd_irq()
3187 if (intmask & SDHCI_INT_AUTO_CMD_ERR && host->data_cmd) { in sdhci_cmd_irq()
3188 struct mmc_request *mrq = host->data_cmd->mrq; in sdhci_cmd_irq()
3194 /* Treat auto-CMD12 error the same as data error */ in sdhci_cmd_irq()
3195 if (!mrq->sbc && (host->flags & SDHCI_AUTO_CMD12)) { in sdhci_cmd_irq()
3201 if (!host->cmd) { in sdhci_cmd_irq()
3203 * SDHCI recovers from errors by resetting the cmd and data in sdhci_cmd_irq()
3207 if (host->pending_reset) in sdhci_cmd_irq()
3210 mmc_hostname(host->mmc), (unsigned)intmask); in sdhci_cmd_irq()
3218 host->cmd->error = -ETIMEDOUT; in sdhci_cmd_irq()
3220 host->cmd->error = -EILSEQ; in sdhci_cmd_irq()
3223 if (host->cmd->data && in sdhci_cmd_irq()
3226 host->cmd = NULL; in sdhci_cmd_irq()
3231 __sdhci_finish_mrq(host, host->cmd->mrq); in sdhci_cmd_irq()
3235 /* Handle auto-CMD23 error */ in sdhci_cmd_irq()
3237 struct mmc_request *mrq = host->cmd->mrq; in sdhci_cmd_irq()
3240 -ETIMEDOUT : in sdhci_cmd_irq()
3241 -EILSEQ; in sdhci_cmd_irq()
3243 if (mrq->sbc && (host->flags & SDHCI_AUTO_CMD23)) { in sdhci_cmd_irq()
3244 mrq->sbc->error = err; in sdhci_cmd_irq()
3256 void *desc = host->adma_table; in sdhci_adma_show_error()
3257 dma_addr_t dma = host->adma_addr; in sdhci_adma_show_error()
3264 if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_adma_show_error()
3267 le32_to_cpu(dma_desc->addr_hi), in sdhci_adma_show_error()
3268 le32_to_cpu(dma_desc->addr_lo), in sdhci_adma_show_error()
3269 le16_to_cpu(dma_desc->len), in sdhci_adma_show_error()
3270 le16_to_cpu(dma_desc->cmd)); in sdhci_adma_show_error()
3274 le32_to_cpu(dma_desc->addr_lo), in sdhci_adma_show_error()
3275 le16_to_cpu(dma_desc->len), in sdhci_adma_show_error()
3276 le16_to_cpu(dma_desc->cmd)); in sdhci_adma_show_error()
3278 desc += host->desc_sz; in sdhci_adma_show_error()
3279 dma += host->desc_sz; in sdhci_adma_show_error()
3281 if (dma_desc->cmd & cpu_to_le16(ADMA2_END)) in sdhci_adma_show_error()
3295 host->tuning_done = 1; in sdhci_data_irq()
3296 wake_up(&host->buf_ready_int); in sdhci_data_irq()
3301 if (!host->data) { in sdhci_data_irq()
3302 struct mmc_command *data_cmd = host->data_cmd; in sdhci_data_irq()
3309 if (data_cmd && (data_cmd->flags & MMC_RSP_BUSY)) { in sdhci_data_irq()
3311 host->data_cmd = NULL; in sdhci_data_irq()
3312 data_cmd->error = -ETIMEDOUT; in sdhci_data_irq()
3313 __sdhci_finish_mrq(host, data_cmd->mrq); in sdhci_data_irq()
3317 host->data_cmd = NULL; in sdhci_data_irq()
3319 * Some cards handle busy-end interrupt in sdhci_data_irq()
3323 if (host->cmd == data_cmd) in sdhci_data_irq()
3326 __sdhci_finish_mrq(host, data_cmd->mrq); in sdhci_data_irq()
3332 * SDHCI recovers from errors by resetting the cmd and data in sdhci_data_irq()
3336 if (host->pending_reset) in sdhci_data_irq()
3340 mmc_hostname(host->mmc), (unsigned)intmask); in sdhci_data_irq()
3347 host->data->error = -ETIMEDOUT; in sdhci_data_irq()
3349 host->data->error = -EILSEQ; in sdhci_data_irq()
3353 host->data->error = -EILSEQ; in sdhci_data_irq()
3355 pr_err("%s: ADMA error: 0x%08x\n", mmc_hostname(host->mmc), in sdhci_data_irq()
3358 host->data->error = -EIO; in sdhci_data_irq()
3359 if (host->ops->adma_workaround) in sdhci_data_irq()
3360 host->ops->adma_workaround(host, intmask); in sdhci_data_irq()
3363 if (host->data->error) in sdhci_data_irq()
3382 dmanow = dmastart + host->data->bytes_xfered; in sdhci_data_irq()
3387 ~((dma_addr_t)SDHCI_DEFAULT_BOUNDARY_SIZE - 1)) + in sdhci_data_irq()
3389 host->data->bytes_xfered = dmanow - dmastart; in sdhci_data_irq()
3391 &dmastart, host->data->bytes_xfered, &dmanow); in sdhci_data_irq()
3396 if (host->cmd == host->data_cmd) { in sdhci_data_irq()
3402 host->data_early = 1; in sdhci_data_irq()
3413 struct mmc_data *data = mrq->data; in sdhci_defer_done()
3415 return host->pending_reset || host->always_defer_done || in sdhci_defer_done()
3416 ((host->flags & SDHCI_REQ_USE_DMA) && data && in sdhci_defer_done()
3417 data->host_cookie == COOKIE_MAPPED); in sdhci_defer_done()
3429 spin_lock(&host->lock); in sdhci_irq()
3431 if (host->runtime_suspended) { in sdhci_irq()
3432 spin_unlock(&host->lock); in sdhci_irq()
3445 if (host->ops->irq) { in sdhci_irq()
3446 intmask = host->ops->irq(host, intmask); in sdhci_irq()
3471 host->ier &= ~(SDHCI_INT_CARD_INSERT | in sdhci_irq()
3473 host->ier |= present ? SDHCI_INT_CARD_REMOVE : in sdhci_irq()
3475 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_irq()
3476 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_irq()
3481 host->thread_isr |= intmask & (SDHCI_INT_CARD_INSERT | in sdhci_irq()
3494 mmc_hostname(host->mmc)); in sdhci_irq()
3497 mmc_retune_needed(host->mmc); in sdhci_irq()
3500 (host->ier & SDHCI_INT_CARD_INT)) { in sdhci_irq()
3502 sdio_signal_irq(host->mmc); in sdhci_irq()
3519 } while (intmask && --max_loops); in sdhci_irq()
3523 struct mmc_request *mrq = host->mrqs_done[i]; in sdhci_irq()
3532 host->mrqs_done[i] = NULL; in sdhci_irq()
3536 if (host->deferred_cmd) in sdhci_irq()
3539 spin_unlock(&host->lock); in sdhci_irq()
3546 if (host->ops->request_done) in sdhci_irq()
3547 host->ops->request_done(host, mrqs_done[i]); in sdhci_irq()
3549 mmc_request_done(host->mmc, mrqs_done[i]); in sdhci_irq()
3554 mmc_hostname(host->mmc), unexpected); in sdhci_irq()
3564 struct mmc_command *cmd; in sdhci_thread_irq() local
3571 spin_lock_irqsave(&host->lock, flags); in sdhci_thread_irq()
3573 isr = host->thread_isr; in sdhci_thread_irq()
3574 host->thread_isr = 0; in sdhci_thread_irq()
3576 cmd = host->deferred_cmd; in sdhci_thread_irq()
3577 if (cmd && !sdhci_send_command_retry(host, cmd, flags)) in sdhci_thread_irq()
3578 sdhci_finish_mrq(host, cmd->mrq); in sdhci_thread_irq()
3580 spin_unlock_irqrestore(&host->lock, flags); in sdhci_thread_irq()
3583 struct mmc_host *mmc = host->mmc; in sdhci_thread_irq()
3585 mmc->ops->card_event(mmc); in sdhci_thread_irq()
3602 return mmc_card_is_removable(host->mmc) && in sdhci_cd_irq_can_wakeup()
3603 !(host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) && in sdhci_cd_irq_can_wakeup()
3604 !mmc_can_gpio_cd(host->mmc); in sdhci_cd_irq_can_wakeup()
3609 * the Interrupt Status Enable register too. See 'Table 1-6: Wakeup Signal
3628 if (mmc_card_wake_sdio_irq(host->mmc)) { in sdhci_enable_irq_wakeups()
3643 host->irq_wake_enabled = !enable_irq_wake(host->irq); in sdhci_enable_irq_wakeups()
3645 return host->irq_wake_enabled; in sdhci_enable_irq_wakeups()
3658 disable_irq_wake(host->irq); in sdhci_disable_irq_wakeups()
3660 host->irq_wake_enabled = false; in sdhci_disable_irq_wakeups()
3667 mmc_retune_timer_stop(host->mmc); in sdhci_suspend_host()
3669 if (!device_may_wakeup(mmc_dev(host->mmc)) || in sdhci_suspend_host()
3671 host->ier = 0; in sdhci_suspend_host()
3674 free_irq(host->irq, host); in sdhci_suspend_host()
3684 struct mmc_host *mmc = host->mmc; in sdhci_resume_host()
3687 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_resume_host()
3688 if (host->ops->enable_dma) in sdhci_resume_host()
3689 host->ops->enable_dma(host); in sdhci_resume_host()
3692 if ((host->mmc->pm_flags & MMC_PM_KEEP_POWER) && in sdhci_resume_host()
3693 (host->quirks2 & SDHCI_QUIRK2_HOST_OFF_CARD_ON)) { in sdhci_resume_host()
3696 host->pwr = 0; in sdhci_resume_host()
3697 host->clock = 0; in sdhci_resume_host()
3698 mmc->ops->set_ios(mmc, &mmc->ios); in sdhci_resume_host()
3700 sdhci_init(host, (host->mmc->pm_flags & MMC_PM_KEEP_POWER)); in sdhci_resume_host()
3703 if (host->irq_wake_enabled) { in sdhci_resume_host()
3706 ret = request_threaded_irq(host->irq, sdhci_irq, in sdhci_resume_host()
3708 mmc_hostname(host->mmc), host); in sdhci_resume_host()
3724 mmc_retune_timer_stop(host->mmc); in sdhci_runtime_suspend_host()
3726 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_suspend_host()
3727 host->ier &= SDHCI_INT_CARD_INT; in sdhci_runtime_suspend_host()
3728 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_runtime_suspend_host()
3729 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_runtime_suspend_host()
3730 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_suspend_host()
3732 synchronize_hardirq(host->irq); in sdhci_runtime_suspend_host()
3734 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_suspend_host()
3735 host->runtime_suspended = true; in sdhci_runtime_suspend_host()
3736 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_suspend_host()
3744 struct mmc_host *mmc = host->mmc; in sdhci_runtime_resume_host()
3746 int host_flags = host->flags; in sdhci_runtime_resume_host()
3749 if (host->ops->enable_dma) in sdhci_runtime_resume_host()
3750 host->ops->enable_dma(host); in sdhci_runtime_resume_host()
3755 if (mmc->ios.power_mode != MMC_POWER_UNDEFINED && in sdhci_runtime_resume_host()
3756 mmc->ios.power_mode != MMC_POWER_OFF) { in sdhci_runtime_resume_host()
3757 /* Force clock and power re-program */ in sdhci_runtime_resume_host()
3758 host->pwr = 0; in sdhci_runtime_resume_host()
3759 host->clock = 0; in sdhci_runtime_resume_host()
3760 mmc->ops->start_signal_voltage_switch(mmc, &mmc->ios); in sdhci_runtime_resume_host()
3761 mmc->ops->set_ios(mmc, &mmc->ios); in sdhci_runtime_resume_host()
3764 !(host->quirks2 & SDHCI_QUIRK2_PRESET_VALUE_BROKEN)) { in sdhci_runtime_resume_host()
3765 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_resume_host()
3767 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_resume_host()
3770 if ((mmc->caps2 & MMC_CAP2_HS400_ES) && in sdhci_runtime_resume_host()
3771 mmc->ops->hs400_enhanced_strobe) in sdhci_runtime_resume_host()
3772 mmc->ops->hs400_enhanced_strobe(mmc, &mmc->ios); in sdhci_runtime_resume_host()
3775 spin_lock_irqsave(&host->lock, flags); in sdhci_runtime_resume_host()
3777 host->runtime_suspended = false; in sdhci_runtime_resume_host()
3786 spin_unlock_irqrestore(&host->lock, flags); in sdhci_runtime_resume_host()
3806 spin_lock_irqsave(&host->lock, flags); in sdhci_cqe_enable()
3813 * for cmd queuing to fetch both command and transfer descriptors. in sdhci_cqe_enable()
3815 if (host->v4_mode && (host->caps1 & SDHCI_CAN_DO_ADMA3)) in sdhci_cqe_enable()
3817 else if (host->flags & SDHCI_USE_64_BIT_DMA) in sdhci_cqe_enable()
3823 sdhci_writew(host, SDHCI_MAKE_BLKSZ(host->sdma_boundary, 512), in sdhci_cqe_enable()
3829 host->ier = host->cqe_ier; in sdhci_cqe_enable()
3831 sdhci_writel(host, host->ier, SDHCI_INT_ENABLE); in sdhci_cqe_enable()
3832 sdhci_writel(host, host->ier, SDHCI_SIGNAL_ENABLE); in sdhci_cqe_enable()
3834 host->cqe_on = true; in sdhci_cqe_enable()
3837 mmc_hostname(mmc), host->ier, in sdhci_cqe_enable()
3840 spin_unlock_irqrestore(&host->lock, flags); in sdhci_cqe_enable()
3849 spin_lock_irqsave(&host->lock, flags); in sdhci_cqe_disable()
3853 host->cqe_on = false; in sdhci_cqe_disable()
3861 mmc_hostname(mmc), host->ier, in sdhci_cqe_disable()
3864 spin_unlock_irqrestore(&host->lock, flags); in sdhci_cqe_disable()
3873 if (!host->cqe_on) in sdhci_cqe_irq()
3877 *cmd_error = -EILSEQ; in sdhci_cqe_irq()
3879 *cmd_error = -ETIMEDOUT; in sdhci_cqe_irq()
3884 *data_error = -EILSEQ; in sdhci_cqe_irq()
3886 *data_error = -ETIMEDOUT; in sdhci_cqe_irq()
3888 *data_error = -EIO; in sdhci_cqe_irq()
3893 mask = intmask & host->cqe_ier; in sdhci_cqe_irq()
3898 mmc_hostname(host->mmc)); in sdhci_cqe_irq()
3900 intmask &= ~(host->cqe_ier | SDHCI_INT_ERROR); in sdhci_cqe_irq()
3904 mmc_hostname(host->mmc), intmask); in sdhci_cqe_irq()
3928 return ERR_PTR(-ENOMEM); in sdhci_alloc_host()
3931 host->mmc = mmc; in sdhci_alloc_host()
3932 host->mmc_host_ops = sdhci_ops; in sdhci_alloc_host()
3933 mmc->ops = &host->mmc_host_ops; in sdhci_alloc_host()
3935 host->flags = SDHCI_SIGNALING_330; in sdhci_alloc_host()
3937 host->cqe_ier = SDHCI_CQE_INT_MASK; in sdhci_alloc_host()
3938 host->cqe_err_ier = SDHCI_CQE_INT_ERR_MASK; in sdhci_alloc_host()
3940 host->tuning_delay = -1; in sdhci_alloc_host()
3941 host->tuning_loop_count = MAX_TUNING_LOOP; in sdhci_alloc_host()
3943 host->sdma_boundary = SDHCI_DEFAULT_BOUNDARY_ARG; in sdhci_alloc_host()
3950 host->adma_table_cnt = SDHCI_MAX_SEGS * 2 + 1; in sdhci_alloc_host()
3959 struct mmc_host *mmc = host->mmc; in sdhci_set_dma_mask()
3961 int ret = -EINVAL; in sdhci_set_dma_mask()
3963 if (host->quirks2 & SDHCI_QUIRK2_BROKEN_64_BIT_DMA) in sdhci_set_dma_mask()
3964 host->flags &= ~SDHCI_USE_64_BIT_DMA; in sdhci_set_dma_mask()
3966 /* Try 64-bit mask if hardware is capable of it */ in sdhci_set_dma_mask()
3967 if (host->flags & SDHCI_USE_64_BIT_DMA) { in sdhci_set_dma_mask()
3970 pr_warn("%s: Failed to set 64-bit DMA mask.\n", in sdhci_set_dma_mask()
3972 host->flags &= ~SDHCI_USE_64_BIT_DMA; in sdhci_set_dma_mask()
3976 /* 32-bit mask as default & fallback */ in sdhci_set_dma_mask()
3980 pr_warn("%s: Failed to set 32-bit DMA mask.\n", in sdhci_set_dma_mask()
3994 if (host->read_caps) in __sdhci_read_caps()
3997 host->read_caps = true; in __sdhci_read_caps()
4000 host->quirks = debug_quirks; in __sdhci_read_caps()
4003 host->quirks2 = debug_quirks2; in __sdhci_read_caps()
4007 if (host->v4_mode) in __sdhci_read_caps()
4010 device_property_read_u64_array(mmc_dev(host->mmc), in __sdhci_read_caps()
4011 "sdhci-caps-mask", &dt_caps_mask, 1); in __sdhci_read_caps()
4012 device_property_read_u64_array(mmc_dev(host->mmc), in __sdhci_read_caps()
4013 "sdhci-caps", &dt_caps, 1); in __sdhci_read_caps()
4016 host->version = (v & SDHCI_SPEC_VER_MASK) >> SDHCI_SPEC_VER_SHIFT; in __sdhci_read_caps()
4018 if (host->quirks & SDHCI_QUIRK_MISSING_CAPS) in __sdhci_read_caps()
4022 host->caps = *caps; in __sdhci_read_caps()
4024 host->caps = sdhci_readl(host, SDHCI_CAPABILITIES); in __sdhci_read_caps()
4025 host->caps &= ~lower_32_bits(dt_caps_mask); in __sdhci_read_caps()
4026 host->caps |= lower_32_bits(dt_caps); in __sdhci_read_caps()
4029 if (host->version < SDHCI_SPEC_300) in __sdhci_read_caps()
4033 host->caps1 = *caps1; in __sdhci_read_caps()
4035 host->caps1 = sdhci_readl(host, SDHCI_CAPABILITIES_1); in __sdhci_read_caps()
4036 host->caps1 &= ~upper_32_bits(dt_caps_mask); in __sdhci_read_caps()
4037 host->caps1 |= upper_32_bits(dt_caps); in __sdhci_read_caps()
4044 struct mmc_host *mmc = host->mmc; in sdhci_allocate_bounce_buffer()
4060 if (mmc->max_req_size < bounce_size) in sdhci_allocate_bounce_buffer()
4061 bounce_size = mmc->max_req_size; in sdhci_allocate_bounce_buffer()
4069 host->bounce_buffer = devm_kmalloc(mmc->parent, in sdhci_allocate_bounce_buffer()
4072 if (!host->bounce_buffer) { in sdhci_allocate_bounce_buffer()
4078 * mmc->max_segs == 1. in sdhci_allocate_bounce_buffer()
4083 host->bounce_addr = dma_map_single(mmc->parent, in sdhci_allocate_bounce_buffer()
4084 host->bounce_buffer, in sdhci_allocate_bounce_buffer()
4087 ret = dma_mapping_error(mmc->parent, host->bounce_addr); in sdhci_allocate_bounce_buffer()
4091 host->bounce_buffer_size = bounce_size; in sdhci_allocate_bounce_buffer()
4094 mmc->max_segs = max_blocks; in sdhci_allocate_bounce_buffer()
4095 mmc->max_seg_size = bounce_size; in sdhci_allocate_bounce_buffer()
4096 mmc->max_req_size = bounce_size; in sdhci_allocate_bounce_buffer()
4106 * version 4.10 in Capabilities Register is used as 64-bit System in sdhci_can_64bit_dma()
4109 if (host->version >= SDHCI_SPEC_410 && host->v4_mode) in sdhci_can_64bit_dma()
4110 return host->caps & SDHCI_CAN_64BIT_V4; in sdhci_can_64bit_dma()
4112 return host->caps & SDHCI_CAN_64BIT; in sdhci_can_64bit_dma()
4127 return -EINVAL; in sdhci_setup_host()
4129 mmc = host->mmc; in sdhci_setup_host()
4137 if (!mmc->supply.vqmmc) { in sdhci_setup_host()
4153 override_timeout_clk = host->timeout_clk; in sdhci_setup_host()
4155 if (host->version > SDHCI_SPEC_420) { in sdhci_setup_host()
4157 mmc_hostname(mmc), host->version); in sdhci_setup_host()
4160 if (host->quirks & SDHCI_QUIRK_FORCE_DMA) in sdhci_setup_host()
4161 host->flags |= SDHCI_USE_SDMA; in sdhci_setup_host()
4162 else if (!(host->caps & SDHCI_CAN_DO_SDMA)) in sdhci_setup_host()
4165 host->flags |= SDHCI_USE_SDMA; in sdhci_setup_host()
4167 if ((host->quirks & SDHCI_QUIRK_BROKEN_DMA) && in sdhci_setup_host()
4168 (host->flags & SDHCI_USE_SDMA)) { in sdhci_setup_host()
4170 host->flags &= ~SDHCI_USE_SDMA; in sdhci_setup_host()
4173 if ((host->version >= SDHCI_SPEC_200) && in sdhci_setup_host()
4174 (host->caps & SDHCI_CAN_DO_ADMA2)) in sdhci_setup_host()
4175 host->flags |= SDHCI_USE_ADMA; in sdhci_setup_host()
4177 if ((host->quirks & SDHCI_QUIRK_BROKEN_ADMA) && in sdhci_setup_host()
4178 (host->flags & SDHCI_USE_ADMA)) { in sdhci_setup_host()
4180 host->flags &= ~SDHCI_USE_ADMA; in sdhci_setup_host()
4184 host->flags |= SDHCI_USE_64_BIT_DMA; in sdhci_setup_host()
4186 if (host->use_external_dma) { in sdhci_setup_host()
4188 if (ret == -EPROBE_DEFER) in sdhci_setup_host()
4198 host->flags &= ~(SDHCI_USE_SDMA | SDHCI_USE_ADMA); in sdhci_setup_host()
4201 if (host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA)) { in sdhci_setup_host()
4202 if (host->ops->set_dma_mask) in sdhci_setup_host()
4203 ret = host->ops->set_dma_mask(host); in sdhci_setup_host()
4207 if (!ret && host->ops->enable_dma) in sdhci_setup_host()
4208 ret = host->ops->enable_dma(host); in sdhci_setup_host()
4211 pr_warn("%s: No suitable DMA available - falling back to PIO\n", in sdhci_setup_host()
4213 host->flags &= ~(SDHCI_USE_SDMA | SDHCI_USE_ADMA); in sdhci_setup_host()
4219 /* SDMA does not support 64-bit DMA if v4 mode not set */ in sdhci_setup_host()
4220 if ((host->flags & SDHCI_USE_64_BIT_DMA) && !host->v4_mode) in sdhci_setup_host()
4221 host->flags &= ~SDHCI_USE_SDMA; in sdhci_setup_host()
4223 if (host->flags & SDHCI_USE_ADMA) { in sdhci_setup_host()
4227 if (!(host->flags & SDHCI_USE_64_BIT_DMA)) in sdhci_setup_host()
4228 host->alloc_desc_sz = SDHCI_ADMA2_32_DESC_SZ; in sdhci_setup_host()
4229 else if (!host->alloc_desc_sz) in sdhci_setup_host()
4230 host->alloc_desc_sz = SDHCI_ADMA2_64_DESC_SZ(host); in sdhci_setup_host()
4232 host->desc_sz = host->alloc_desc_sz; in sdhci_setup_host()
4233 host->adma_table_sz = host->adma_table_cnt * host->desc_sz; in sdhci_setup_host()
4235 host->align_buffer_sz = SDHCI_MAX_SEGS * SDHCI_ADMA2_ALIGN; in sdhci_setup_host()
4237 * Use zalloc to zero the reserved high 32-bits of 128-bit in sdhci_setup_host()
4241 host->align_buffer_sz + host->adma_table_sz, in sdhci_setup_host()
4244 pr_warn("%s: Unable to allocate ADMA buffers - falling back to standard DMA\n", in sdhci_setup_host()
4246 host->flags &= ~SDHCI_USE_ADMA; in sdhci_setup_host()
4247 } else if ((dma + host->align_buffer_sz) & in sdhci_setup_host()
4248 (SDHCI_ADMA2_DESC_ALIGN - 1)) { in sdhci_setup_host()
4251 host->flags &= ~SDHCI_USE_ADMA; in sdhci_setup_host()
4252 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_setup_host()
4253 host->adma_table_sz, buf, dma); in sdhci_setup_host()
4255 host->align_buffer = buf; in sdhci_setup_host()
4256 host->align_addr = dma; in sdhci_setup_host()
4258 host->adma_table = buf + host->align_buffer_sz; in sdhci_setup_host()
4259 host->adma_addr = dma + host->align_buffer_sz; in sdhci_setup_host()
4268 if (!(host->flags & (SDHCI_USE_SDMA | SDHCI_USE_ADMA))) { in sdhci_setup_host()
4269 host->dma_mask = DMA_BIT_MASK(64); in sdhci_setup_host()
4270 mmc_dev(mmc)->dma_mask = &host->dma_mask; in sdhci_setup_host()
4273 if (host->version >= SDHCI_SPEC_300) in sdhci_setup_host()
4274 host->max_clk = FIELD_GET(SDHCI_CLOCK_V3_BASE_MASK, host->caps); in sdhci_setup_host()
4276 host->max_clk = FIELD_GET(SDHCI_CLOCK_BASE_MASK, host->caps); in sdhci_setup_host()
4278 host->max_clk *= 1000000; in sdhci_setup_host()
4279 if (host->max_clk == 0 || host->quirks & in sdhci_setup_host()
4281 if (!host->ops->get_max_clock) { in sdhci_setup_host()
4284 ret = -ENODEV; in sdhci_setup_host()
4287 host->max_clk = host->ops->get_max_clock(host); in sdhci_setup_host()
4294 host->clk_mul = FIELD_GET(SDHCI_CLOCK_MUL_MASK, host->caps1); in sdhci_setup_host()
4302 if (host->clk_mul) in sdhci_setup_host()
4303 host->clk_mul += 1; in sdhci_setup_host()
4308 max_clk = host->max_clk; in sdhci_setup_host()
4310 if (host->ops->get_min_clock) in sdhci_setup_host()
4311 mmc->f_min = host->ops->get_min_clock(host); in sdhci_setup_host()
4312 else if (host->version >= SDHCI_SPEC_300) { in sdhci_setup_host()
4313 if (host->clk_mul) in sdhci_setup_host()
4314 max_clk = host->max_clk * host->clk_mul; in sdhci_setup_host()
4319 mmc->f_min = host->max_clk / SDHCI_MAX_DIV_SPEC_300; in sdhci_setup_host()
4321 mmc->f_min = host->max_clk / SDHCI_MAX_DIV_SPEC_200; in sdhci_setup_host()
4323 if (!mmc->f_max || mmc->f_max > max_clk) in sdhci_setup_host()
4324 mmc->f_max = max_clk; in sdhci_setup_host()
4326 if (!(host->quirks & SDHCI_QUIRK_DATA_TIMEOUT_USES_SDCLK)) { in sdhci_setup_host()
4327 host->timeout_clk = FIELD_GET(SDHCI_TIMEOUT_CLK_MASK, host->caps); in sdhci_setup_host()
4329 if (host->caps & SDHCI_TIMEOUT_CLK_UNIT) in sdhci_setup_host()
4330 host->timeout_clk *= 1000; in sdhci_setup_host()
4332 if (host->timeout_clk == 0) { in sdhci_setup_host()
4333 if (!host->ops->get_timeout_clock) { in sdhci_setup_host()
4336 ret = -ENODEV; in sdhci_setup_host()
4340 host->timeout_clk = in sdhci_setup_host()
4341 DIV_ROUND_UP(host->ops->get_timeout_clock(host), in sdhci_setup_host()
4346 host->timeout_clk = override_timeout_clk; in sdhci_setup_host()
4348 mmc->max_busy_timeout = host->ops->get_max_timeout_count ? in sdhci_setup_host()
4349 host->ops->get_max_timeout_count(host) : 1 << 27; in sdhci_setup_host()
4350 mmc->max_busy_timeout /= host->timeout_clk; in sdhci_setup_host()
4353 if (host->quirks2 & SDHCI_QUIRK2_DISABLE_HW_TIMEOUT && in sdhci_setup_host()
4354 !host->ops->get_max_timeout_count) in sdhci_setup_host()
4355 mmc->max_busy_timeout = 0; in sdhci_setup_host()
4357 mmc->caps |= MMC_CAP_SDIO_IRQ | MMC_CAP_CMD23; in sdhci_setup_host()
4358 mmc->caps2 |= MMC_CAP2_SDIO_IRQ_NOTHREAD; in sdhci_setup_host()
4360 if (host->quirks & SDHCI_QUIRK_MULTIBLOCK_READ_ACMD12) in sdhci_setup_host()
4361 host->flags |= SDHCI_AUTO_CMD12; in sdhci_setup_host()
4364 * For v3 mode, Auto-CMD23 stuff only works in ADMA or PIO. in sdhci_setup_host()
4365 * For v4 mode, SDMA may use Auto-CMD23 as well. in sdhci_setup_host()
4367 if ((host->version >= SDHCI_SPEC_300) && in sdhci_setup_host()
4368 ((host->flags & SDHCI_USE_ADMA) || in sdhci_setup_host()
4369 !(host->flags & SDHCI_USE_SDMA) || host->v4_mode) && in sdhci_setup_host()
4370 !(host->quirks2 & SDHCI_QUIRK2_ACMD23_BROKEN)) { in sdhci_setup_host()
4371 host->flags |= SDHCI_AUTO_CMD23; in sdhci_setup_host()
4372 DBG("Auto-CMD23 available\n"); in sdhci_setup_host()
4374 DBG("Auto-CMD23 unavailable\n"); in sdhci_setup_host()
4378 * A controller may support 8-bit width, but the board itself in sdhci_setup_host()
4379 * might not have the pins brought out. Boards that support in sdhci_setup_host()
4380 * 8-bit width must set "mmc->caps |= MMC_CAP_8_BIT_DATA;" in in sdhci_setup_host()
4382 * won't assume 8-bit width for hosts without that CAP. in sdhci_setup_host()
4384 if (!(host->quirks & SDHCI_QUIRK_FORCE_1_BIT_DATA)) in sdhci_setup_host()
4385 mmc->caps |= MMC_CAP_4_BIT_DATA; in sdhci_setup_host()
4387 if (host->quirks2 & SDHCI_QUIRK2_HOST_NO_CMD23) in sdhci_setup_host()
4388 mmc->caps &= ~MMC_CAP_CMD23; in sdhci_setup_host()
4390 if (host->caps & SDHCI_CAN_DO_HISPD) in sdhci_setup_host()
4391 mmc->caps |= MMC_CAP_SD_HIGHSPEED | MMC_CAP_MMC_HIGHSPEED; in sdhci_setup_host()
4393 if ((host->quirks & SDHCI_QUIRK_BROKEN_CARD_DETECTION) && in sdhci_setup_host()
4395 mmc_gpio_get_cd(host->mmc) < 0) in sdhci_setup_host()
4396 mmc->caps |= MMC_CAP_NEEDS_POLL; in sdhci_setup_host()
4398 if (!IS_ERR(mmc->supply.vqmmc)) { in sdhci_setup_host()
4400 ret = regulator_enable(mmc->supply.vqmmc); in sdhci_setup_host()
4401 host->sdhci_core_to_disable_vqmmc = !ret; in sdhci_setup_host()
4405 if (!regulator_is_supported_voltage(mmc->supply.vqmmc, 1700000, in sdhci_setup_host()
4407 host->caps1 &= ~(SDHCI_SUPPORT_SDR104 | in sdhci_setup_host()
4412 if (!regulator_is_supported_voltage(mmc->supply.vqmmc, 2700000, in sdhci_setup_host()
4414 host->flags &= ~SDHCI_SIGNALING_330; in sdhci_setup_host()
4419 mmc->supply.vqmmc = ERR_PTR(-EINVAL); in sdhci_setup_host()
4424 if (host->quirks2 & SDHCI_QUIRK2_NO_1_8_V) { in sdhci_setup_host()
4425 host->caps1 &= ~(SDHCI_SUPPORT_SDR104 | SDHCI_SUPPORT_SDR50 | in sdhci_setup_host()
4429 * (indicated using mmc-hs200-1_8v/mmc-hs400-1_8v dt property), in sdhci_setup_host()
4435 mmc->caps2 &= ~(MMC_CAP2_HSX00_1_8V | MMC_CAP2_HS400_ES); in sdhci_setup_host()
4436 mmc->caps &= ~(MMC_CAP_1_8V_DDR | MMC_CAP_UHS); in sdhci_setup_host()
4439 /* Any UHS-I mode in caps implies SDR12 and SDR25 support. */ in sdhci_setup_host()
4440 if (host->caps1 & (SDHCI_SUPPORT_SDR104 | SDHCI_SUPPORT_SDR50 | in sdhci_setup_host()
4442 mmc->caps |= MMC_CAP_UHS_SDR12 | MMC_CAP_UHS_SDR25; in sdhci_setup_host()
4445 if (host->caps1 & SDHCI_SUPPORT_SDR104) { in sdhci_setup_host()
4446 mmc->caps |= MMC_CAP_UHS_SDR104 | MMC_CAP_UHS_SDR50; in sdhci_setup_host()
4450 if (!(host->quirks2 & SDHCI_QUIRK2_BROKEN_HS200)) in sdhci_setup_host()
4451 mmc->caps2 |= MMC_CAP2_HS200; in sdhci_setup_host()
4452 } else if (host->caps1 & SDHCI_SUPPORT_SDR50) { in sdhci_setup_host()
4453 mmc->caps |= MMC_CAP_UHS_SDR50; in sdhci_setup_host()
4456 if (host->quirks2 & SDHCI_QUIRK2_CAPS_BIT63_FOR_HS400 && in sdhci_setup_host()
4457 (host->caps1 & SDHCI_SUPPORT_HS400)) in sdhci_setup_host()
4458 mmc->caps2 |= MMC_CAP2_HS400; in sdhci_setup_host()
4460 if ((mmc->caps2 & MMC_CAP2_HSX00_1_2V) && in sdhci_setup_host()
4461 (IS_ERR(mmc->supply.vqmmc) || in sdhci_setup_host()
4462 !regulator_is_supported_voltage(mmc->supply.vqmmc, 1100000, in sdhci_setup_host()
4464 mmc->caps2 &= ~MMC_CAP2_HSX00_1_2V; in sdhci_setup_host()
4466 if ((host->caps1 & SDHCI_SUPPORT_DDR50) && in sdhci_setup_host()
4467 !(host->quirks2 & SDHCI_QUIRK2_BROKEN_DDR50)) in sdhci_setup_host()
4468 mmc->caps |= MMC_CAP_UHS_DDR50; in sdhci_setup_host()
4471 if (host->caps1 & SDHCI_USE_SDR50_TUNING) in sdhci_setup_host()
4472 host->flags |= SDHCI_SDR50_NEEDS_TUNING; in sdhci_setup_host()
4475 if (host->caps1 & SDHCI_DRIVER_TYPE_A) in sdhci_setup_host()
4476 mmc->caps |= MMC_CAP_DRIVER_TYPE_A; in sdhci_setup_host()
4477 if (host->caps1 & SDHCI_DRIVER_TYPE_C) in sdhci_setup_host()
4478 mmc->caps |= MMC_CAP_DRIVER_TYPE_C; in sdhci_setup_host()
4479 if (host->caps1 & SDHCI_DRIVER_TYPE_D) in sdhci_setup_host()
4480 mmc->caps |= MMC_CAP_DRIVER_TYPE_D; in sdhci_setup_host()
4482 /* Initial value for re-tuning timer count */ in sdhci_setup_host()
4483 host->tuning_count = FIELD_GET(SDHCI_RETUNING_TIMER_COUNT_MASK, in sdhci_setup_host()
4484 host->caps1); in sdhci_setup_host()
4487 * In case Re-tuning Timer is not disabled, the actual value of in sdhci_setup_host()
4488 * re-tuning timer will be 2 ^ (n - 1). in sdhci_setup_host()
4490 if (host->tuning_count) in sdhci_setup_host()
4491 host->tuning_count = 1 << (host->tuning_count - 1); in sdhci_setup_host()
4493 /* Re-tuning mode supported by the Host Controller */ in sdhci_setup_host()
4494 host->tuning_mode = FIELD_GET(SDHCI_RETUNING_MODE_MASK, host->caps1); in sdhci_setup_host()
4506 if (!max_current_caps && !IS_ERR(mmc->supply.vmmc)) { in sdhci_setup_host()
4507 int curr = regulator_get_current_limit(mmc->supply.vmmc); in sdhci_setup_host()
4522 if (host->caps & SDHCI_CAN_VDD_330) { in sdhci_setup_host()
4525 mmc->max_current_330 = FIELD_GET(SDHCI_MAX_CURRENT_330_MASK, in sdhci_setup_host()
4529 if (host->caps & SDHCI_CAN_VDD_300) { in sdhci_setup_host()
4532 mmc->max_current_300 = FIELD_GET(SDHCI_MAX_CURRENT_300_MASK, in sdhci_setup_host()
4536 if (host->caps & SDHCI_CAN_VDD_180) { in sdhci_setup_host()
4539 mmc->max_current_180 = FIELD_GET(SDHCI_MAX_CURRENT_180_MASK, in sdhci_setup_host()
4545 if (host->ocr_mask) in sdhci_setup_host()
4546 ocr_avail = host->ocr_mask; in sdhci_setup_host()
4549 if (mmc->ocr_avail) in sdhci_setup_host()
4550 ocr_avail = mmc->ocr_avail; in sdhci_setup_host()
4552 mmc->ocr_avail = ocr_avail; in sdhci_setup_host()
4553 mmc->ocr_avail_sdio = ocr_avail; in sdhci_setup_host()
4554 if (host->ocr_avail_sdio) in sdhci_setup_host()
4555 mmc->ocr_avail_sdio &= host->ocr_avail_sdio; in sdhci_setup_host()
4556 mmc->ocr_avail_sd = ocr_avail; in sdhci_setup_host()
4557 if (host->ocr_avail_sd) in sdhci_setup_host()
4558 mmc->ocr_avail_sd &= host->ocr_avail_sd; in sdhci_setup_host()
4560 mmc->ocr_avail_sd &= ~MMC_VDD_165_195; in sdhci_setup_host()
4561 mmc->ocr_avail_mmc = ocr_avail; in sdhci_setup_host()
4562 if (host->ocr_avail_mmc) in sdhci_setup_host()
4563 mmc->ocr_avail_mmc &= host->ocr_avail_mmc; in sdhci_setup_host()
4565 if (mmc->ocr_avail == 0) { in sdhci_setup_host()
4568 ret = -ENODEV; in sdhci_setup_host()
4572 if ((mmc->caps & (MMC_CAP_UHS_SDR12 | MMC_CAP_UHS_SDR25 | in sdhci_setup_host()
4575 (mmc->caps2 & (MMC_CAP2_HS200_1_8V_SDR | MMC_CAP2_HS400_1_8V))) in sdhci_setup_host()
4576 host->flags |= SDHCI_SIGNALING_180; in sdhci_setup_host()
4578 if (mmc->caps2 & MMC_CAP2_HSX00_1_2V) in sdhci_setup_host()
4579 host->flags |= SDHCI_SIGNALING_120; in sdhci_setup_host()
4581 spin_lock_init(&host->lock); in sdhci_setup_host()
4588 mmc->max_req_size = 524288; in sdhci_setup_host()
4594 if (host->flags & SDHCI_USE_ADMA) { in sdhci_setup_host()
4595 mmc->max_segs = SDHCI_MAX_SEGS; in sdhci_setup_host()
4596 } else if (host->flags & SDHCI_USE_SDMA) { in sdhci_setup_host()
4597 mmc->max_segs = 1; in sdhci_setup_host()
4601 mmc->max_req_size = min(mmc->max_req_size, in sdhci_setup_host()
4605 mmc->max_segs = SDHCI_MAX_SEGS; in sdhci_setup_host()
4613 if (host->flags & SDHCI_USE_ADMA) { in sdhci_setup_host()
4614 if (host->quirks & SDHCI_QUIRK_BROKEN_ADMA_ZEROLEN_DESC) in sdhci_setup_host()
4615 mmc->max_seg_size = 65535; in sdhci_setup_host()
4617 mmc->max_seg_size = 65536; in sdhci_setup_host()
4619 mmc->max_seg_size = mmc->max_req_size; in sdhci_setup_host()
4626 if (host->quirks & SDHCI_QUIRK_FORCE_BLK_SZ_2048) { in sdhci_setup_host()
4627 mmc->max_blk_size = 2; in sdhci_setup_host()
4629 mmc->max_blk_size = (host->caps & SDHCI_MAX_BLOCK_MASK) >> in sdhci_setup_host()
4631 if (mmc->max_blk_size >= 3) { in sdhci_setup_host()
4634 mmc->max_blk_size = 0; in sdhci_setup_host()
4638 mmc->max_blk_size = 512 << mmc->max_blk_size; in sdhci_setup_host()
4643 mmc->max_blk_count = (host->quirks & SDHCI_QUIRK_NO_MULTIBLOCK) ? 1 : 65535; in sdhci_setup_host()
4645 if (mmc->max_segs == 1) in sdhci_setup_host()
4646 /* This may alter mmc->*_blk_* parameters */ in sdhci_setup_host()
4652 if (host->sdhci_core_to_disable_vqmmc) in sdhci_setup_host()
4653 regulator_disable(mmc->supply.vqmmc); in sdhci_setup_host()
4655 if (host->align_buffer) in sdhci_setup_host()
4656 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_setup_host()
4657 host->adma_table_sz, host->align_buffer, in sdhci_setup_host()
4658 host->align_addr); in sdhci_setup_host()
4659 host->adma_table = NULL; in sdhci_setup_host()
4660 host->align_buffer = NULL; in sdhci_setup_host()
4668 struct mmc_host *mmc = host->mmc; in sdhci_cleanup_host()
4670 if (host->sdhci_core_to_disable_vqmmc) in sdhci_cleanup_host()
4671 regulator_disable(mmc->supply.vqmmc); in sdhci_cleanup_host()
4673 if (host->align_buffer) in sdhci_cleanup_host()
4674 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_cleanup_host()
4675 host->adma_table_sz, host->align_buffer, in sdhci_cleanup_host()
4676 host->align_addr); in sdhci_cleanup_host()
4678 if (host->use_external_dma) in sdhci_cleanup_host()
4681 host->adma_table = NULL; in sdhci_cleanup_host()
4682 host->align_buffer = NULL; in sdhci_cleanup_host()
4689 struct mmc_host *mmc = host->mmc; in __sdhci_add_host()
4692 if ((mmc->caps2 & MMC_CAP2_CQE) && in __sdhci_add_host()
4693 (host->quirks & SDHCI_QUIRK_BROKEN_CQE)) { in __sdhci_add_host()
4694 mmc->caps2 &= ~MMC_CAP2_CQE; in __sdhci_add_host()
4695 mmc->cqe_ops = NULL; in __sdhci_add_host()
4698 host->complete_wq = alloc_workqueue("sdhci", flags, 0); in __sdhci_add_host()
4699 if (!host->complete_wq) in __sdhci_add_host()
4700 return -ENOMEM; in __sdhci_add_host()
4702 INIT_WORK(&host->complete_work, sdhci_complete_work); in __sdhci_add_host()
4704 timer_setup(&host->timer, sdhci_timeout_timer, 0); in __sdhci_add_host()
4705 timer_setup(&host->data_timer, sdhci_timeout_data_timer, 0); in __sdhci_add_host()
4707 init_waitqueue_head(&host->buf_ready_int); in __sdhci_add_host()
4711 ret = request_threaded_irq(host->irq, sdhci_irq, sdhci_thread_irq, in __sdhci_add_host()
4715 mmc_hostname(mmc), host->irq, ret); in __sdhci_add_host()
4731 mmc_hostname(mmc), host->hw_name, dev_name(mmc_dev(mmc)), in __sdhci_add_host()
4732 host->use_external_dma ? "External DMA" : in __sdhci_add_host()
4733 (host->flags & SDHCI_USE_ADMA) ? in __sdhci_add_host()
4734 (host->flags & SDHCI_USE_64_BIT_DMA) ? "ADMA 64-bit" : "ADMA" : in __sdhci_add_host()
4735 (host->flags & SDHCI_USE_SDMA) ? "DMA" : "PIO"); in __sdhci_add_host()
4747 free_irq(host->irq, host); in __sdhci_add_host()
4749 destroy_workqueue(host->complete_wq); in __sdhci_add_host()
4778 struct mmc_host *mmc = host->mmc; in sdhci_remove_host()
4782 spin_lock_irqsave(&host->lock, flags); in sdhci_remove_host()
4784 host->flags |= SDHCI_DEVICE_DEAD; in sdhci_remove_host()
4789 sdhci_error_out_mrqs(host, -ENOMEDIUM); in sdhci_remove_host()
4792 spin_unlock_irqrestore(&host->lock, flags); in sdhci_remove_host()
4806 free_irq(host->irq, host); in sdhci_remove_host()
4808 del_timer_sync(&host->timer); in sdhci_remove_host()
4809 del_timer_sync(&host->data_timer); in sdhci_remove_host()
4811 destroy_workqueue(host->complete_wq); in sdhci_remove_host()
4813 if (host->sdhci_core_to_disable_vqmmc) in sdhci_remove_host()
4814 regulator_disable(mmc->supply.vqmmc); in sdhci_remove_host()
4816 if (host->align_buffer) in sdhci_remove_host()
4817 dma_free_coherent(mmc_dev(mmc), host->align_buffer_sz + in sdhci_remove_host()
4818 host->adma_table_sz, host->align_buffer, in sdhci_remove_host()
4819 host->align_addr); in sdhci_remove_host()
4821 if (host->use_external_dma) in sdhci_remove_host()
4824 host->adma_table = NULL; in sdhci_remove_host()
4825 host->align_buffer = NULL; in sdhci_remove_host()
4832 mmc_free_host(host->mmc); in sdhci_free_host()