/block/ |
D | sed-opal.c | 33 int (*fn)(struct opal_dev *dev, void *data); 36 typedef int (cont_fn)(struct opal_dev *dev); 214 static int end_opal_session_error(struct opal_dev *dev); 215 static int opal_discovery0_step(struct opal_dev *dev); 320 static int opal_send_cmd(struct opal_dev *dev) in opal_send_cmd() argument 322 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_send_cmd() 323 dev->cmd, IO_BUFFER_LENGTH, in opal_send_cmd() 327 static int opal_recv_cmd(struct opal_dev *dev) in opal_recv_cmd() argument 329 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_recv_cmd() 330 dev->resp, IO_BUFFER_LENGTH, in opal_recv_cmd() [all …]
|
D | blk-pm.c | 31 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument 33 q->dev = dev; in blk_pm_runtime_init() 35 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init() 36 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init() 65 if (!q->dev) in blk_pre_runtime_suspend() 99 pm_runtime_mark_last_busy(q->dev); in blk_pre_runtime_suspend() 124 if (!q->dev) in blk_post_runtime_suspend() 132 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_suspend() 154 if (!q->dev) in blk_pre_runtime_resume() 203 if (!q->dev) in blk_set_runtime_active() [all …]
|
D | genhd.c | 508 &disk->bdi->dev->kobj, "bdi"); in device_add_disk() 655 static ssize_t disk_badblocks_show(struct device *dev, in disk_badblocks_show() argument 659 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_show() 667 static ssize_t disk_badblocks_store(struct device *dev, in disk_badblocks_store() argument 671 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_store() 707 struct device *dev; in printk_all_partitions() local 710 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions() 711 struct gendisk *disk = dev_to_disk(dev); in printk_all_partitions() 740 else if (dev->parent && dev->parent->driver) in printk_all_partitions() 742 dev->parent->driver->name); in printk_all_partitions() [all …]
|
D | bsg-lib.c | 154 put_device(job->dev); /* release reference for the request */ in bsg_teardown_job() 225 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job() argument 242 job->dev = dev; in bsg_prepare_job() 244 get_device(job->dev); in bsg_prepare_job() 269 struct device *dev = q->queuedata; in bsg_queue_rq() local 278 if (!get_device(dev)) in bsg_queue_rq() 281 if (!bsg_prepare_job(dev, req)) in bsg_queue_rq() 289 put_device(dev); in bsg_queue_rq() 366 struct request_queue *bsg_setup_queue(struct device *dev, const char *name, in bsg_setup_queue() argument 397 q->queuedata = dev; in bsg_setup_queue() [all …]
|
D | blk-pm.h | 11 if (!q->dev || !blk_queue_pm_only(q)) in blk_pm_resume_queue() 15 pm_request_resume(q->dev); in blk_pm_resume_queue() 21 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy() 22 pm_runtime_mark_last_busy(rq->q->dev); in blk_pm_mark_last_busy()
|
D | bdev.c | 505 void bdev_add(struct block_device *bdev, dev_t dev) in bdev_add() argument 507 bdev->bd_dev = dev; in bdev_add() 508 bdev->bd_inode->i_rdev = dev; in bdev_add() 509 bdev->bd_inode->i_ino = dev; in bdev_add() 733 struct block_device *blkdev_get_no_open(dev_t dev) in blkdev_get_no_open() argument 738 inode = ilookup(blockdev_superblock, dev); in blkdev_get_no_open() 740 blk_request_module(dev); in blkdev_get_no_open() 741 inode = ilookup(blockdev_superblock, dev); in blkdev_get_no_open() 790 struct block_device *blkdev_get_by_dev(dev_t dev, fmode_t mode, void *holder) in blkdev_get_by_dev() argument 798 MAJOR(dev), MINOR(dev), in blkdev_get_by_dev() [all …]
|
D | blk-timeout.c | 39 ssize_t part_timeout_show(struct device *dev, struct device_attribute *attr, in part_timeout_show() argument 42 struct gendisk *disk = dev_to_disk(dev); in part_timeout_show() 48 ssize_t part_timeout_store(struct device *dev, struct device_attribute *attr, in part_timeout_store() argument 51 struct gendisk *disk = dev_to_disk(dev); in part_timeout_store()
|
D | disk-events.c | 357 static ssize_t disk_events_show(struct device *dev, in disk_events_show() argument 360 struct gendisk *disk = dev_to_disk(dev); in disk_events_show() 367 static ssize_t disk_events_async_show(struct device *dev, in disk_events_async_show() argument 373 static ssize_t disk_events_poll_msecs_show(struct device *dev, in disk_events_poll_msecs_show() argument 377 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_show() 384 static ssize_t disk_events_poll_msecs_store(struct device *dev, in disk_events_poll_msecs_store() argument 388 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_store()
|
D | badblocks.c | 544 static int __badblocks_init(struct device *dev, struct badblocks *bb, in __badblocks_init() argument 547 bb->dev = dev; in __badblocks_init() 553 if (dev) in __badblocks_init() 554 bb->page = devm_kzalloc(dev, PAGE_SIZE, GFP_KERNEL); in __badblocks_init() 581 int devm_init_badblocks(struct device *dev, struct badblocks *bb) in devm_init_badblocks() argument 585 return __badblocks_init(dev, bb, 1); in devm_init_badblocks() 597 if (bb->dev) in badblocks_exit() 598 devm_kfree(bb->dev, bb->page); in badblocks_exit()
|
D | keyslot-manager.c | 54 if (ksm->dev) in blk_ksm_hw_enter() 55 pm_runtime_get_sync(ksm->dev); in blk_ksm_hw_enter() 62 if (ksm->dev) in blk_ksm_hw_exit() 63 pm_runtime_put_sync(ksm->dev); in blk_ksm_hw_exit() 152 int devm_blk_ksm_init(struct device *dev, struct blk_keyslot_manager *ksm, in devm_blk_ksm_init() argument 160 return devm_add_action_or_reset(dev, blk_ksm_destroy_callback, ksm); in devm_blk_ksm_init()
|
D | blk-crypto-profile.c | 51 if (profile->dev) in blk_crypto_hw_enter() 52 pm_runtime_get_sync(profile->dev); in blk_crypto_hw_enter() 59 if (profile->dev) in blk_crypto_hw_exit() 60 pm_runtime_put_sync(profile->dev); in blk_crypto_hw_exit() 158 int devm_blk_crypto_profile_init(struct device *dev, in devm_blk_crypto_profile_init() argument 167 return devm_add_action_or_reset(dev, in devm_blk_crypto_profile_init()
|
D | blk-sysfs.c | 850 struct device *dev = disk_to_dev(disk); in blk_register_queue() local 853 ret = blk_trace_init_sysfs(dev); in blk_register_queue() 859 ret = kobject_add(&q->kobj, kobject_get(&dev->kobj), "%s", "queue"); in blk_register_queue() 861 blk_trace_remove_sysfs(dev); in blk_register_queue() 867 blk_trace_remove_sysfs(dev); in blk_register_queue() 869 kobject_put(&dev->kobj); in blk_register_queue() 879 __blk_mq_register_dev(dev, q); in blk_register_queue() 890 blk_trace_remove_sysfs(dev); in blk_register_queue() 891 kobject_put(&dev->kobj); in blk_register_queue() 935 blk_trace_remove_sysfs(dev); in blk_register_queue() [all …]
|
D | blk-mq-sysfs.c | 215 void blk_mq_unregister_dev(struct device *dev, struct request_queue *q) in blk_mq_unregister_dev() argument 227 kobject_put(&dev->kobj); in blk_mq_unregister_dev() 264 int __blk_mq_register_dev(struct device *dev, struct request_queue *q) in __blk_mq_register_dev() argument 272 ret = kobject_add(q->mq_kobj, kobject_get(&dev->kobj), "%s", "mq"); in __blk_mq_register_dev() 295 kobject_put(&dev->kobj); in __blk_mq_register_dev()
|
D | blk-mq-rdma.c | 25 struct ib_device *dev, int first_vec) in blk_mq_rdma_map_queues() argument 31 mask = ib_get_vector_affinity(dev, first_vec + queue); in blk_mq_rdma_map_queues()
|
D | bsg.c | 168 static void bsg_device_release(struct device *dev) in bsg_device_release() argument 170 struct bsg_device *bd = container_of(dev, struct bsg_device, device); in bsg_device_release() 235 static char *bsg_devnode(struct device *dev, umode_t *mode) in bsg_devnode() argument 237 return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev)); in bsg_devnode()
|
D | blk.h | 254 ssize_t part_size_show(struct device *dev, struct device_attribute *attr, 256 ssize_t part_stat_show(struct device *dev, struct device_attribute *attr, 258 ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr, 260 ssize_t part_fail_show(struct device *dev, struct device_attribute *attr, 262 ssize_t part_fail_store(struct device *dev, struct device_attribute *attr,
|
D | kyber-iosched.c | 154 dev_t dev; member 261 trace_kyber_latency(kqd->dev, kyber_domain_names[sched_domain], in calculate_percentile() 274 trace_kyber_adjust(kqd->dev, kyber_domain_names[sched_domain], in kyber_resize_domain() 370 kqd->dev = disk_devt(q->disk); in kyber_queue_data_alloc() 779 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain() 792 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain()
|
D | blk-settings.c | 901 struct device *dev) in blk_queue_can_use_dma_map_merging() argument 903 unsigned long boundary = dma_get_merge_boundary(dev); in blk_queue_can_use_dma_map_merging()
|
D | blk-mq.h | 129 extern int __blk_mq_register_dev(struct device *dev, struct request_queue *q);
|
D | blk-cgroup.c | 505 if (!blkg->q->disk || !blkg->q->disk->bdi->dev) in blkg_dev_name() 862 struct device *dev; in blkcg_fill_root_iostats() local 865 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats() 866 struct block_device *bdev = dev_to_bdev(dev); in blkcg_fill_root_iostats()
|
/block/partitions/ |
D | core.c | 183 static ssize_t part_partition_show(struct device *dev, in part_partition_show() argument 186 return sprintf(buf, "%d\n", dev_to_bdev(dev)->bd_partno); in part_partition_show() 189 static ssize_t part_start_show(struct device *dev, in part_start_show() argument 192 return sprintf(buf, "%llu\n", dev_to_bdev(dev)->bd_start_sect); in part_start_show() 195 static ssize_t part_ro_show(struct device *dev, in part_ro_show() argument 198 return sprintf(buf, "%d\n", bdev_read_only(dev_to_bdev(dev))); in part_ro_show() 201 static ssize_t part_alignment_offset_show(struct device *dev, in part_alignment_offset_show() argument 204 struct block_device *bdev = dev_to_bdev(dev); in part_alignment_offset_show() 211 static ssize_t part_discard_alignment_show(struct device *dev, in part_discard_alignment_show() argument 214 struct block_device *bdev = dev_to_bdev(dev); in part_discard_alignment_show() [all …]
|
D | mac.c | 16 extern void note_bootable_part(dev_t dev, int part, int goodness);
|