1 /* 2 * drivers/base/power/generic_ops.c - Generic PM callbacks for subsystems 3 * 4 * Copyright (c) 2010 Rafael J. Wysocki <rjw@sisk.pl>, Novell Inc. 5 * 6 * This file is released under the GPLv2. 7 */ 8 9 #include <linux/pm.h> 10 #include <linux/pm_runtime.h> 11 #include <linux/export.h> 12 13 #ifdef CONFIG_PM 14 /** 15 * pm_generic_runtime_suspend - Generic runtime suspend callback for subsystems. 16 * @dev: Device to suspend. 17 * 18 * If PM operations are defined for the @dev's driver and they include 19 * ->runtime_suspend(), execute it and return its error code. Otherwise, 20 * return 0. 21 */ pm_generic_runtime_suspend(struct device * dev)22int pm_generic_runtime_suspend(struct device *dev) 23 { 24 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 25 int ret; 26 27 ret = pm && pm->runtime_suspend ? pm->runtime_suspend(dev) : 0; 28 29 return ret; 30 } 31 EXPORT_SYMBOL_GPL(pm_generic_runtime_suspend); 32 33 /** 34 * pm_generic_runtime_resume - Generic runtime resume callback for subsystems. 35 * @dev: Device to resume. 36 * 37 * If PM operations are defined for the @dev's driver and they include 38 * ->runtime_resume(), execute it and return its error code. Otherwise, 39 * return 0. 40 */ pm_generic_runtime_resume(struct device * dev)41int pm_generic_runtime_resume(struct device *dev) 42 { 43 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 44 int ret; 45 46 ret = pm && pm->runtime_resume ? pm->runtime_resume(dev) : 0; 47 48 return ret; 49 } 50 EXPORT_SYMBOL_GPL(pm_generic_runtime_resume); 51 #endif /* CONFIG_PM */ 52 53 #ifdef CONFIG_PM_SLEEP 54 /** 55 * pm_generic_prepare - Generic routine preparing a device for power transition. 56 * @dev: Device to prepare. 57 * 58 * Prepare a device for a system-wide power transition. 59 */ pm_generic_prepare(struct device * dev)60int pm_generic_prepare(struct device *dev) 61 { 62 struct device_driver *drv = dev->driver; 63 int ret = 0; 64 65 if (drv && drv->pm && drv->pm->prepare) 66 ret = drv->pm->prepare(dev); 67 68 return ret; 69 } 70 71 /** 72 * pm_generic_suspend_noirq - Generic suspend_noirq callback for subsystems. 73 * @dev: Device to suspend. 74 */ pm_generic_suspend_noirq(struct device * dev)75int pm_generic_suspend_noirq(struct device *dev) 76 { 77 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 78 79 return pm && pm->suspend_noirq ? pm->suspend_noirq(dev) : 0; 80 } 81 EXPORT_SYMBOL_GPL(pm_generic_suspend_noirq); 82 83 /** 84 * pm_generic_suspend_late - Generic suspend_late callback for subsystems. 85 * @dev: Device to suspend. 86 */ pm_generic_suspend_late(struct device * dev)87int pm_generic_suspend_late(struct device *dev) 88 { 89 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 90 91 return pm && pm->suspend_late ? pm->suspend_late(dev) : 0; 92 } 93 EXPORT_SYMBOL_GPL(pm_generic_suspend_late); 94 95 /** 96 * pm_generic_suspend - Generic suspend callback for subsystems. 97 * @dev: Device to suspend. 98 */ pm_generic_suspend(struct device * dev)99int pm_generic_suspend(struct device *dev) 100 { 101 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 102 103 return pm && pm->suspend ? pm->suspend(dev) : 0; 104 } 105 EXPORT_SYMBOL_GPL(pm_generic_suspend); 106 107 /** 108 * pm_generic_freeze_noirq - Generic freeze_noirq callback for subsystems. 109 * @dev: Device to freeze. 110 */ pm_generic_freeze_noirq(struct device * dev)111int pm_generic_freeze_noirq(struct device *dev) 112 { 113 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 114 115 return pm && pm->freeze_noirq ? pm->freeze_noirq(dev) : 0; 116 } 117 EXPORT_SYMBOL_GPL(pm_generic_freeze_noirq); 118 119 /** 120 * pm_generic_freeze_late - Generic freeze_late callback for subsystems. 121 * @dev: Device to freeze. 122 */ pm_generic_freeze_late(struct device * dev)123int pm_generic_freeze_late(struct device *dev) 124 { 125 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 126 127 return pm && pm->freeze_late ? pm->freeze_late(dev) : 0; 128 } 129 EXPORT_SYMBOL_GPL(pm_generic_freeze_late); 130 131 /** 132 * pm_generic_freeze - Generic freeze callback for subsystems. 133 * @dev: Device to freeze. 134 */ pm_generic_freeze(struct device * dev)135int pm_generic_freeze(struct device *dev) 136 { 137 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 138 139 return pm && pm->freeze ? pm->freeze(dev) : 0; 140 } 141 EXPORT_SYMBOL_GPL(pm_generic_freeze); 142 143 /** 144 * pm_generic_poweroff_noirq - Generic poweroff_noirq callback for subsystems. 145 * @dev: Device to handle. 146 */ pm_generic_poweroff_noirq(struct device * dev)147int pm_generic_poweroff_noirq(struct device *dev) 148 { 149 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 150 151 return pm && pm->poweroff_noirq ? pm->poweroff_noirq(dev) : 0; 152 } 153 EXPORT_SYMBOL_GPL(pm_generic_poweroff_noirq); 154 155 /** 156 * pm_generic_poweroff_late - Generic poweroff_late callback for subsystems. 157 * @dev: Device to handle. 158 */ pm_generic_poweroff_late(struct device * dev)159int pm_generic_poweroff_late(struct device *dev) 160 { 161 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 162 163 return pm && pm->poweroff_late ? pm->poweroff_late(dev) : 0; 164 } 165 EXPORT_SYMBOL_GPL(pm_generic_poweroff_late); 166 167 /** 168 * pm_generic_poweroff - Generic poweroff callback for subsystems. 169 * @dev: Device to handle. 170 */ pm_generic_poweroff(struct device * dev)171int pm_generic_poweroff(struct device *dev) 172 { 173 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 174 175 return pm && pm->poweroff ? pm->poweroff(dev) : 0; 176 } 177 EXPORT_SYMBOL_GPL(pm_generic_poweroff); 178 179 /** 180 * pm_generic_thaw_noirq - Generic thaw_noirq callback for subsystems. 181 * @dev: Device to thaw. 182 */ pm_generic_thaw_noirq(struct device * dev)183int pm_generic_thaw_noirq(struct device *dev) 184 { 185 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 186 187 return pm && pm->thaw_noirq ? pm->thaw_noirq(dev) : 0; 188 } 189 EXPORT_SYMBOL_GPL(pm_generic_thaw_noirq); 190 191 /** 192 * pm_generic_thaw_early - Generic thaw_early callback for subsystems. 193 * @dev: Device to thaw. 194 */ pm_generic_thaw_early(struct device * dev)195int pm_generic_thaw_early(struct device *dev) 196 { 197 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 198 199 return pm && pm->thaw_early ? pm->thaw_early(dev) : 0; 200 } 201 EXPORT_SYMBOL_GPL(pm_generic_thaw_early); 202 203 /** 204 * pm_generic_thaw - Generic thaw callback for subsystems. 205 * @dev: Device to thaw. 206 */ pm_generic_thaw(struct device * dev)207int pm_generic_thaw(struct device *dev) 208 { 209 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 210 211 return pm && pm->thaw ? pm->thaw(dev) : 0; 212 } 213 EXPORT_SYMBOL_GPL(pm_generic_thaw); 214 215 /** 216 * pm_generic_resume_noirq - Generic resume_noirq callback for subsystems. 217 * @dev: Device to resume. 218 */ pm_generic_resume_noirq(struct device * dev)219int pm_generic_resume_noirq(struct device *dev) 220 { 221 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 222 223 return pm && pm->resume_noirq ? pm->resume_noirq(dev) : 0; 224 } 225 EXPORT_SYMBOL_GPL(pm_generic_resume_noirq); 226 227 /** 228 * pm_generic_resume_early - Generic resume_early callback for subsystems. 229 * @dev: Device to resume. 230 */ pm_generic_resume_early(struct device * dev)231int pm_generic_resume_early(struct device *dev) 232 { 233 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 234 235 return pm && pm->resume_early ? pm->resume_early(dev) : 0; 236 } 237 EXPORT_SYMBOL_GPL(pm_generic_resume_early); 238 239 /** 240 * pm_generic_resume - Generic resume callback for subsystems. 241 * @dev: Device to resume. 242 */ pm_generic_resume(struct device * dev)243int pm_generic_resume(struct device *dev) 244 { 245 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 246 247 return pm && pm->resume ? pm->resume(dev) : 0; 248 } 249 EXPORT_SYMBOL_GPL(pm_generic_resume); 250 251 /** 252 * pm_generic_restore_noirq - Generic restore_noirq callback for subsystems. 253 * @dev: Device to restore. 254 */ pm_generic_restore_noirq(struct device * dev)255int pm_generic_restore_noirq(struct device *dev) 256 { 257 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 258 259 return pm && pm->restore_noirq ? pm->restore_noirq(dev) : 0; 260 } 261 EXPORT_SYMBOL_GPL(pm_generic_restore_noirq); 262 263 /** 264 * pm_generic_restore_early - Generic restore_early callback for subsystems. 265 * @dev: Device to resume. 266 */ pm_generic_restore_early(struct device * dev)267int pm_generic_restore_early(struct device *dev) 268 { 269 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 270 271 return pm && pm->restore_early ? pm->restore_early(dev) : 0; 272 } 273 EXPORT_SYMBOL_GPL(pm_generic_restore_early); 274 275 /** 276 * pm_generic_restore - Generic restore callback for subsystems. 277 * @dev: Device to restore. 278 */ pm_generic_restore(struct device * dev)279int pm_generic_restore(struct device *dev) 280 { 281 const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; 282 283 return pm && pm->restore ? pm->restore(dev) : 0; 284 } 285 EXPORT_SYMBOL_GPL(pm_generic_restore); 286 287 /** 288 * pm_generic_complete - Generic routine completing a device power transition. 289 * @dev: Device to handle. 290 * 291 * Complete a device power transition during a system-wide power transition. 292 */ pm_generic_complete(struct device * dev)293void pm_generic_complete(struct device *dev) 294 { 295 struct device_driver *drv = dev->driver; 296 297 if (drv && drv->pm && drv->pm->complete) 298 drv->pm->complete(dev); 299 300 /* 301 * Let runtime PM try to suspend devices that haven't been in use before 302 * going into the system-wide sleep state we're resuming from. 303 */ 304 pm_request_idle(dev); 305 } 306 #endif /* CONFIG_PM_SLEEP */ 307