1 /* 2 * linux/drivers/mmc/core/host.c 3 * 4 * Copyright (C) 2003 Russell King, All Rights Reserved. 5 * Copyright (C) 2007-2008 Pierre Ossman 6 * Copyright (C) 2010 Linus Walleij 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License version 2 as 10 * published by the Free Software Foundation. 11 * 12 * MMC host class device management 13 */ 14 15 #include <linux/device.h> 16 #include <linux/err.h> 17 #include <linux/idr.h> 18 #include <linux/of.h> 19 #include <linux/of_gpio.h> 20 #include <linux/pagemap.h> 21 #include <linux/export.h> 22 #include <linux/leds.h> 23 #include <linux/slab.h> 24 #include <linux/suspend.h> 25 26 #include <linux/mmc/host.h> 27 #include <linux/mmc/card.h> 28 #include <linux/mmc/slot-gpio.h> 29 30 #include "core.h" 31 #include "host.h" 32 mmc_host_classdev_release(struct device * dev)33 static void mmc_host_classdev_release(struct device *dev) 34 { 35 struct mmc_host *host = cls_dev_to_mmc_host(dev); 36 mutex_destroy(&host->slot.lock); 37 kfree(host); 38 } 39 40 static struct class mmc_host_class = { 41 .name = "mmc_host", 42 .dev_release = mmc_host_classdev_release, 43 }; 44 mmc_register_host_class(void)45 int mmc_register_host_class(void) 46 { 47 return class_register(&mmc_host_class); 48 } 49 mmc_unregister_host_class(void)50 void mmc_unregister_host_class(void) 51 { 52 class_unregister(&mmc_host_class); 53 } 54 55 static DEFINE_IDR(mmc_host_idr); 56 static DEFINE_SPINLOCK(mmc_host_lock); 57 58 #ifdef CONFIG_MMC_CLKGATE clkgate_delay_show(struct device * dev,struct device_attribute * attr,char * buf)59 static ssize_t clkgate_delay_show(struct device *dev, 60 struct device_attribute *attr, char *buf) 61 { 62 struct mmc_host *host = cls_dev_to_mmc_host(dev); 63 return snprintf(buf, PAGE_SIZE, "%lu\n", host->clkgate_delay); 64 } 65 clkgate_delay_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)66 static ssize_t clkgate_delay_store(struct device *dev, 67 struct device_attribute *attr, const char *buf, size_t count) 68 { 69 struct mmc_host *host = cls_dev_to_mmc_host(dev); 70 unsigned long flags, value; 71 72 if (kstrtoul(buf, 0, &value)) 73 return -EINVAL; 74 75 spin_lock_irqsave(&host->clk_lock, flags); 76 host->clkgate_delay = value; 77 spin_unlock_irqrestore(&host->clk_lock, flags); 78 return count; 79 } 80 81 /* 82 * Enabling clock gating will make the core call out to the host 83 * once up and once down when it performs a request or card operation 84 * intermingled in any fashion. The driver will see this through 85 * set_ios() operations with ios.clock field set to 0 to gate (disable) 86 * the block clock, and to the old frequency to enable it again. 87 */ mmc_host_clk_gate_delayed(struct mmc_host * host)88 static void mmc_host_clk_gate_delayed(struct mmc_host *host) 89 { 90 unsigned long tick_ns; 91 unsigned long freq = host->ios.clock; 92 unsigned long flags; 93 94 if (!freq) { 95 pr_debug("%s: frequency set to 0 in disable function, " 96 "this means the clock is already disabled.\n", 97 mmc_hostname(host)); 98 return; 99 } 100 /* 101 * New requests may have appeared while we were scheduling, 102 * then there is no reason to delay the check before 103 * clk_disable(). 104 */ 105 spin_lock_irqsave(&host->clk_lock, flags); 106 107 /* 108 * Delay n bus cycles (at least 8 from MMC spec) before attempting 109 * to disable the MCI block clock. The reference count may have 110 * gone up again after this delay due to rescheduling! 111 */ 112 if (!host->clk_requests) { 113 spin_unlock_irqrestore(&host->clk_lock, flags); 114 tick_ns = DIV_ROUND_UP(1000000000, freq); 115 ndelay(host->clk_delay * tick_ns); 116 } else { 117 /* New users appeared while waiting for this work */ 118 spin_unlock_irqrestore(&host->clk_lock, flags); 119 return; 120 } 121 mutex_lock(&host->clk_gate_mutex); 122 spin_lock_irqsave(&host->clk_lock, flags); 123 if (!host->clk_requests) { 124 spin_unlock_irqrestore(&host->clk_lock, flags); 125 /* This will set host->ios.clock to 0 */ 126 mmc_gate_clock(host); 127 spin_lock_irqsave(&host->clk_lock, flags); 128 pr_debug("%s: gated MCI clock\n", mmc_hostname(host)); 129 } 130 spin_unlock_irqrestore(&host->clk_lock, flags); 131 mutex_unlock(&host->clk_gate_mutex); 132 } 133 134 /* 135 * Internal work. Work to disable the clock at some later point. 136 */ mmc_host_clk_gate_work(struct work_struct * work)137 static void mmc_host_clk_gate_work(struct work_struct *work) 138 { 139 struct mmc_host *host = container_of(work, struct mmc_host, 140 clk_gate_work.work); 141 142 mmc_host_clk_gate_delayed(host); 143 } 144 145 /** 146 * mmc_host_clk_hold - ungate hardware MCI clocks 147 * @host: host to ungate. 148 * 149 * Makes sure the host ios.clock is restored to a non-zero value 150 * past this call. Increase clock reference count and ungate clock 151 * if we're the first user. 152 */ mmc_host_clk_hold(struct mmc_host * host)153 void mmc_host_clk_hold(struct mmc_host *host) 154 { 155 unsigned long flags; 156 157 /* cancel any clock gating work scheduled by mmc_host_clk_release() */ 158 cancel_delayed_work_sync(&host->clk_gate_work); 159 mutex_lock(&host->clk_gate_mutex); 160 spin_lock_irqsave(&host->clk_lock, flags); 161 if (host->clk_gated) { 162 spin_unlock_irqrestore(&host->clk_lock, flags); 163 mmc_ungate_clock(host); 164 spin_lock_irqsave(&host->clk_lock, flags); 165 pr_debug("%s: ungated MCI clock\n", mmc_hostname(host)); 166 } 167 host->clk_requests++; 168 spin_unlock_irqrestore(&host->clk_lock, flags); 169 mutex_unlock(&host->clk_gate_mutex); 170 } 171 172 /** 173 * mmc_host_may_gate_card - check if this card may be gated 174 * @card: card to check. 175 */ mmc_host_may_gate_card(struct mmc_card * card)176 static bool mmc_host_may_gate_card(struct mmc_card *card) 177 { 178 /* If there is no card we may gate it */ 179 if (!card) 180 return true; 181 /* 182 * Don't gate SDIO cards! These need to be clocked at all times 183 * since they may be independent systems generating interrupts 184 * and other events. The clock requests counter from the core will 185 * go down to zero since the core does not need it, but we will not 186 * gate the clock, because there is somebody out there that may still 187 * be using it. 188 */ 189 return !(card->quirks & MMC_QUIRK_BROKEN_CLK_GATING); 190 } 191 192 /** 193 * mmc_host_clk_release - gate off hardware MCI clocks 194 * @host: host to gate. 195 * 196 * Calls the host driver with ios.clock set to zero as often as possible 197 * in order to gate off hardware MCI clocks. Decrease clock reference 198 * count and schedule disabling of clock. 199 */ mmc_host_clk_release(struct mmc_host * host)200 void mmc_host_clk_release(struct mmc_host *host) 201 { 202 unsigned long flags; 203 204 spin_lock_irqsave(&host->clk_lock, flags); 205 host->clk_requests--; 206 if (mmc_host_may_gate_card(host->card) && 207 !host->clk_requests) 208 schedule_delayed_work(&host->clk_gate_work, 209 msecs_to_jiffies(host->clkgate_delay)); 210 spin_unlock_irqrestore(&host->clk_lock, flags); 211 } 212 213 /** 214 * mmc_host_clk_rate - get current clock frequency setting 215 * @host: host to get the clock frequency for. 216 * 217 * Returns current clock frequency regardless of gating. 218 */ mmc_host_clk_rate(struct mmc_host * host)219 unsigned int mmc_host_clk_rate(struct mmc_host *host) 220 { 221 unsigned long freq; 222 unsigned long flags; 223 224 spin_lock_irqsave(&host->clk_lock, flags); 225 if (host->clk_gated) 226 freq = host->clk_old; 227 else 228 freq = host->ios.clock; 229 spin_unlock_irqrestore(&host->clk_lock, flags); 230 return freq; 231 } 232 233 /** 234 * mmc_host_clk_init - set up clock gating code 235 * @host: host with potential clock to control 236 */ mmc_host_clk_init(struct mmc_host * host)237 static inline void mmc_host_clk_init(struct mmc_host *host) 238 { 239 host->clk_requests = 0; 240 /* Hold MCI clock for 8 cycles by default */ 241 host->clk_delay = 8; 242 /* 243 * Default clock gating delay is 0ms to avoid wasting power. 244 * This value can be tuned by writing into sysfs entry. 245 */ 246 host->clkgate_delay = 0; 247 host->clk_gated = false; 248 INIT_DELAYED_WORK(&host->clk_gate_work, mmc_host_clk_gate_work); 249 spin_lock_init(&host->clk_lock); 250 mutex_init(&host->clk_gate_mutex); 251 } 252 253 /** 254 * mmc_host_clk_exit - shut down clock gating code 255 * @host: host with potential clock to control 256 */ mmc_host_clk_exit(struct mmc_host * host)257 static inline void mmc_host_clk_exit(struct mmc_host *host) 258 { 259 /* 260 * Wait for any outstanding gate and then make sure we're 261 * ungated before exiting. 262 */ 263 if (cancel_delayed_work_sync(&host->clk_gate_work)) 264 mmc_host_clk_gate_delayed(host); 265 if (host->clk_gated) 266 mmc_host_clk_hold(host); 267 /* There should be only one user now */ 268 WARN_ON(host->clk_requests > 1); 269 } 270 mmc_host_clk_sysfs_init(struct mmc_host * host)271 static inline void mmc_host_clk_sysfs_init(struct mmc_host *host) 272 { 273 host->clkgate_delay_attr.show = clkgate_delay_show; 274 host->clkgate_delay_attr.store = clkgate_delay_store; 275 sysfs_attr_init(&host->clkgate_delay_attr.attr); 276 host->clkgate_delay_attr.attr.name = "clkgate_delay"; 277 host->clkgate_delay_attr.attr.mode = S_IRUGO | S_IWUSR; 278 if (device_create_file(&host->class_dev, &host->clkgate_delay_attr)) 279 pr_err("%s: Failed to create clkgate_delay sysfs entry\n", 280 mmc_hostname(host)); 281 } 282 #else 283 mmc_host_clk_init(struct mmc_host * host)284 static inline void mmc_host_clk_init(struct mmc_host *host) 285 { 286 } 287 mmc_host_clk_exit(struct mmc_host * host)288 static inline void mmc_host_clk_exit(struct mmc_host *host) 289 { 290 } 291 mmc_host_clk_sysfs_init(struct mmc_host * host)292 static inline void mmc_host_clk_sysfs_init(struct mmc_host *host) 293 { 294 } 295 296 #endif 297 298 /** 299 * mmc_of_parse() - parse host's device-tree node 300 * @host: host whose node should be parsed. 301 * 302 * To keep the rest of the MMC subsystem unaware of whether DT has been 303 * used to to instantiate and configure this host instance or not, we 304 * parse the properties and set respective generic mmc-host flags and 305 * parameters. 306 */ mmc_of_parse(struct mmc_host * host)307 void mmc_of_parse(struct mmc_host *host) 308 { 309 struct device_node *np; 310 u32 bus_width; 311 bool explicit_inv_wp, gpio_inv_wp = false; 312 enum of_gpio_flags flags; 313 int len, ret, gpio; 314 315 if (!host->parent || !host->parent->of_node) 316 return; 317 318 np = host->parent->of_node; 319 320 /* "bus-width" is translated to MMC_CAP_*_BIT_DATA flags */ 321 if (of_property_read_u32(np, "bus-width", &bus_width) < 0) { 322 dev_dbg(host->parent, 323 "\"bus-width\" property is missing, assuming 1 bit.\n"); 324 bus_width = 1; 325 } 326 327 switch (bus_width) { 328 case 8: 329 host->caps |= MMC_CAP_8_BIT_DATA; 330 /* Hosts capable of 8-bit transfers can also do 4 bits */ 331 case 4: 332 host->caps |= MMC_CAP_4_BIT_DATA; 333 break; 334 case 1: 335 break; 336 default: 337 dev_err(host->parent, 338 "Invalid \"bus-width\" value %ud!\n", bus_width); 339 } 340 341 /* f_max is obtained from the optional "max-frequency" property */ 342 of_property_read_u32(np, "max-frequency", &host->f_max); 343 344 /* 345 * Configure CD and WP pins. They are both by default active low to 346 * match the SDHCI spec. If GPIOs are provided for CD and / or WP, the 347 * mmc-gpio helpers are used to attach, configure and use them. If 348 * polarity inversion is specified in DT, one of MMC_CAP2_CD_ACTIVE_HIGH 349 * and MMC_CAP2_RO_ACTIVE_HIGH capability-2 flags is set. If the 350 * "broken-cd" property is provided, the MMC_CAP_NEEDS_POLL capability 351 * is set. If the "non-removable" property is found, the 352 * MMC_CAP_NONREMOVABLE capability is set and no card-detection 353 * configuration is performed. 354 */ 355 356 /* Parse Card Detection */ 357 if (of_find_property(np, "non-removable", &len)) { 358 host->caps |= MMC_CAP_NONREMOVABLE; 359 } else { 360 bool explicit_inv_cd, gpio_inv_cd = false; 361 362 explicit_inv_cd = of_property_read_bool(np, "cd-inverted"); 363 364 if (of_find_property(np, "broken-cd", &len)) 365 host->caps |= MMC_CAP_NEEDS_POLL; 366 367 gpio = of_get_named_gpio_flags(np, "cd-gpios", 0, &flags); 368 if (gpio_is_valid(gpio)) { 369 if (!(flags & OF_GPIO_ACTIVE_LOW)) 370 gpio_inv_cd = true; 371 372 ret = mmc_gpio_request_cd(host, gpio); 373 if (ret < 0) 374 dev_err(host->parent, 375 "Failed to request CD GPIO #%d: %d!\n", 376 gpio, ret); 377 else 378 dev_info(host->parent, "Got CD GPIO #%d.\n", 379 gpio); 380 } 381 382 if (explicit_inv_cd ^ gpio_inv_cd) 383 host->caps2 |= MMC_CAP2_CD_ACTIVE_HIGH; 384 } 385 386 /* Parse Write Protection */ 387 explicit_inv_wp = of_property_read_bool(np, "wp-inverted"); 388 389 gpio = of_get_named_gpio_flags(np, "wp-gpios", 0, &flags); 390 if (gpio_is_valid(gpio)) { 391 if (!(flags & OF_GPIO_ACTIVE_LOW)) 392 gpio_inv_wp = true; 393 394 ret = mmc_gpio_request_ro(host, gpio); 395 if (ret < 0) 396 dev_err(host->parent, 397 "Failed to request WP GPIO: %d!\n", ret); 398 } 399 if (explicit_inv_wp ^ gpio_inv_wp) 400 host->caps2 |= MMC_CAP2_RO_ACTIVE_HIGH; 401 402 if (of_find_property(np, "cap-sd-highspeed", &len)) 403 host->caps |= MMC_CAP_SD_HIGHSPEED; 404 if (of_find_property(np, "cap-mmc-highspeed", &len)) 405 host->caps |= MMC_CAP_MMC_HIGHSPEED; 406 if (of_find_property(np, "cap-power-off-card", &len)) 407 host->caps |= MMC_CAP_POWER_OFF_CARD; 408 if (of_find_property(np, "cap-sdio-irq", &len)) 409 host->caps |= MMC_CAP_SDIO_IRQ; 410 if (of_find_property(np, "full-pwr-cycle", &len)) 411 host->caps2 |= MMC_CAP2_FULL_PWR_CYCLE; 412 if (of_find_property(np, "keep-power-in-suspend", &len)) 413 host->pm_caps |= MMC_PM_KEEP_POWER; 414 if (of_find_property(np, "enable-sdio-wakeup", &len)) 415 host->pm_caps |= MMC_PM_WAKE_SDIO_IRQ; 416 } 417 418 EXPORT_SYMBOL(mmc_of_parse); 419 420 /** 421 * mmc_alloc_host - initialise the per-host structure. 422 * @extra: sizeof private data structure 423 * @dev: pointer to host device model structure 424 * 425 * Initialise the per-host structure. 426 */ mmc_alloc_host(int extra,struct device * dev)427 struct mmc_host *mmc_alloc_host(int extra, struct device *dev) 428 { 429 int err; 430 struct mmc_host *host; 431 432 host = kzalloc(sizeof(struct mmc_host) + extra, GFP_KERNEL); 433 if (!host) 434 return NULL; 435 436 /* scanning will be enabled when we're ready */ 437 host->rescan_disable = 1; 438 idr_preload(GFP_KERNEL); 439 spin_lock(&mmc_host_lock); 440 err = idr_alloc(&mmc_host_idr, host, 0, 0, GFP_NOWAIT); 441 if (err >= 0) 442 host->index = err; 443 spin_unlock(&mmc_host_lock); 444 idr_preload_end(); 445 if (err < 0) 446 goto free; 447 448 dev_set_name(&host->class_dev, "mmc%d", host->index); 449 450 host->parent = dev; 451 host->class_dev.parent = dev; 452 host->class_dev.class = &mmc_host_class; 453 device_initialize(&host->class_dev); 454 455 mmc_host_clk_init(host); 456 457 mutex_init(&host->slot.lock); 458 host->slot.cd_irq = -EINVAL; 459 460 spin_lock_init(&host->lock); 461 init_waitqueue_head(&host->wq); 462 host->wlock_name = kasprintf(GFP_KERNEL, 463 "%s_detect", mmc_hostname(host)); 464 wake_lock_init(&host->detect_wake_lock, WAKE_LOCK_SUSPEND, 465 host->wlock_name); 466 INIT_DELAYED_WORK(&host->detect, mmc_rescan); 467 #ifdef CONFIG_PM 468 host->pm_notify.notifier_call = mmc_pm_notify; 469 #endif 470 471 /* 472 * By default, hosts do not support SGIO or large requests. 473 * They have to set these according to their abilities. 474 */ 475 host->max_segs = 1; 476 host->max_seg_size = PAGE_CACHE_SIZE; 477 478 host->max_req_size = PAGE_CACHE_SIZE; 479 host->max_blk_size = 512; 480 host->max_blk_count = PAGE_CACHE_SIZE / 512; 481 482 return host; 483 484 free: 485 kfree(host); 486 return NULL; 487 } 488 489 EXPORT_SYMBOL(mmc_alloc_host); 490 491 /** 492 * mmc_add_host - initialise host hardware 493 * @host: mmc host 494 * 495 * Register the host with the driver model. The host must be 496 * prepared to start servicing requests before this function 497 * completes. 498 */ mmc_add_host(struct mmc_host * host)499 int mmc_add_host(struct mmc_host *host) 500 { 501 int err; 502 503 WARN_ON((host->caps & MMC_CAP_SDIO_IRQ) && 504 !host->ops->enable_sdio_irq); 505 506 err = device_add(&host->class_dev); 507 if (err) 508 return err; 509 510 led_trigger_register_simple(dev_name(&host->class_dev), &host->led); 511 512 #ifdef CONFIG_DEBUG_FS 513 mmc_add_host_debugfs(host); 514 #endif 515 mmc_host_clk_sysfs_init(host); 516 517 mmc_latency_hist_sysfs_init(host); 518 519 mmc_start_host(host); 520 if (!(host->pm_flags & MMC_PM_IGNORE_PM_NOTIFY)) 521 register_pm_notifier(&host->pm_notify); 522 523 return 0; 524 } 525 526 EXPORT_SYMBOL(mmc_add_host); 527 528 /** 529 * mmc_remove_host - remove host hardware 530 * @host: mmc host 531 * 532 * Unregister and remove all cards associated with this host, 533 * and power down the MMC bus. No new requests will be issued 534 * after this function has returned. 535 */ mmc_remove_host(struct mmc_host * host)536 void mmc_remove_host(struct mmc_host *host) 537 { 538 if (!(host->pm_flags & MMC_PM_IGNORE_PM_NOTIFY)) 539 unregister_pm_notifier(&host->pm_notify); 540 541 mmc_stop_host(host); 542 543 #ifdef CONFIG_DEBUG_FS 544 mmc_remove_host_debugfs(host); 545 #endif 546 547 device_del(&host->class_dev); 548 549 led_trigger_unregister_simple(host->led); 550 551 mmc_host_clk_exit(host); 552 } 553 554 EXPORT_SYMBOL(mmc_remove_host); 555 556 /** 557 * mmc_free_host - free the host structure 558 * @host: mmc host 559 * 560 * Free the host once all references to it have been dropped. 561 */ mmc_free_host(struct mmc_host * host)562 void mmc_free_host(struct mmc_host *host) 563 { 564 spin_lock(&mmc_host_lock); 565 idr_remove(&mmc_host_idr, host->index); 566 spin_unlock(&mmc_host_lock); 567 wake_lock_destroy(&host->detect_wake_lock); 568 kfree(host->wlock_name); 569 mmc_latency_hist_sysfs_exit(host); 570 put_device(&host->class_dev); 571 } 572 573 EXPORT_SYMBOL(mmc_free_host); 574