/block/ |
D | sed-opal.c | 33 int (*fn)(struct opal_dev *dev, void *data); 36 typedef int (cont_fn)(struct opal_dev *dev); 214 static int end_opal_session_error(struct opal_dev *dev); 215 static int opal_discovery0_step(struct opal_dev *dev); 320 static int opal_send_cmd(struct opal_dev *dev) in opal_send_cmd() argument 322 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_send_cmd() 323 dev->cmd, IO_BUFFER_LENGTH, in opal_send_cmd() 327 static int opal_recv_cmd(struct opal_dev *dev) in opal_recv_cmd() argument 329 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_recv_cmd() 330 dev->resp, IO_BUFFER_LENGTH, in opal_recv_cmd() [all …]
|
D | blk-pm.c | 31 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument 33 q->dev = dev; in blk_pm_runtime_init() 35 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init() 36 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init() 65 if (!q->dev) in blk_pre_runtime_suspend() 99 pm_runtime_mark_last_busy(q->dev); in blk_pre_runtime_suspend() 124 if (!q->dev) in blk_post_runtime_suspend() 132 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_suspend() 154 if (!q->dev) in blk_pre_runtime_resume() 203 if (!q->dev) in blk_set_runtime_active() [all …]
|
D | genhd.c | 750 if (disk->queue->backing_dev_info->dev) { in register_disk() 752 &disk->queue->backing_dev_info->dev->kobj, in register_disk() 815 struct device *dev = disk_to_dev(disk); in __device_add_disk() local 819 dev->devt = devt; in __device_add_disk() 822 bdi_set_owner(bdi, dev); in __device_add_disk() 958 static ssize_t disk_badblocks_show(struct device *dev, in disk_badblocks_show() argument 962 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_show() 970 static ssize_t disk_badblocks_store(struct device *dev, in disk_badblocks_store() argument 974 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_store() 1070 struct device *dev; in printk_all_partitions() local [all …]
|
D | blk-pm.h | 11 if (q->dev && (q->rpm_status == RPM_SUSPENDED || in blk_pm_request_resume() 13 pm_request_resume(q->dev); in blk_pm_request_resume() 18 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy() 19 pm_runtime_mark_last_busy(rq->q->dev); in blk_pm_mark_last_busy() 26 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_requeue_request() 35 if (q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_add_request() 43 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_put_request()
|
D | bsg-lib.c | 153 put_device(job->dev); /* release reference for the request */ in bsg_teardown_job() 224 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job() argument 241 job->dev = dev; in bsg_prepare_job() 243 get_device(job->dev); in bsg_prepare_job() 268 struct device *dev = q->queuedata; in bsg_queue_rq() local 277 if (!get_device(dev)) in bsg_queue_rq() 280 if (!bsg_prepare_job(dev, req)) in bsg_queue_rq() 288 put_device(dev); in bsg_queue_rq() 365 struct request_queue *bsg_setup_queue(struct device *dev, const char *name, in bsg_setup_queue() argument 396 q->queuedata = dev; in bsg_setup_queue() [all …]
|
D | blk-timeout.c | 39 ssize_t part_timeout_show(struct device *dev, struct device_attribute *attr, in part_timeout_show() argument 42 struct gendisk *disk = dev_to_disk(dev); in part_timeout_show() 48 ssize_t part_timeout_store(struct device *dev, struct device_attribute *attr, in part_timeout_store() argument 51 struct gendisk *disk = dev_to_disk(dev); in part_timeout_store()
|
D | badblocks.c | 544 static int __badblocks_init(struct device *dev, struct badblocks *bb, in __badblocks_init() argument 547 bb->dev = dev; in __badblocks_init() 553 if (dev) in __badblocks_init() 554 bb->page = devm_kzalloc(dev, PAGE_SIZE, GFP_KERNEL); in __badblocks_init() 581 int devm_init_badblocks(struct device *dev, struct badblocks *bb) in devm_init_badblocks() argument 585 return __badblocks_init(dev, bb, 1); in devm_init_badblocks() 597 if (bb->dev) in badblocks_exit() 598 devm_kfree(bb->dev, bb->page); in badblocks_exit()
|
D | keyslot-manager.c | 54 if (ksm->dev) in blk_ksm_hw_enter() 55 pm_runtime_get_sync(ksm->dev); in blk_ksm_hw_enter() 62 if (ksm->dev) in blk_ksm_hw_exit() 63 pm_runtime_put_sync(ksm->dev); in blk_ksm_hw_exit() 152 int devm_blk_ksm_init(struct device *dev, struct blk_keyslot_manager *ksm, in devm_blk_ksm_init() argument 160 return devm_add_action_or_reset(dev, blk_ksm_destroy_callback, ksm); in devm_blk_ksm_init()
|
D | blk-mq-rdma.c | 25 struct ib_device *dev, int first_vec) in blk_mq_rdma_map_queues() argument 31 mask = ib_get_vector_affinity(dev, first_vec + queue); in blk_mq_rdma_map_queues()
|
D | blk-sysfs.c | 843 struct device *dev = disk_to_dev(disk); in blk_register_queue() local 851 kobject_name(&dev->kobj)); in blk_register_queue() 869 ret = blk_trace_init_sysfs(dev); in blk_register_queue() 875 ret = kobject_add(&q->kobj, kobject_get(&dev->kobj), "%s", "queue"); in blk_register_queue() 877 blk_trace_remove_sysfs(dev); in blk_register_queue() 883 blk_trace_remove_sysfs(dev); in blk_register_queue() 885 kobject_put(&dev->kobj); in blk_register_queue() 895 __blk_mq_register_dev(dev, q); in blk_register_queue() 906 blk_trace_remove_sysfs(dev); in blk_register_queue() 907 kobject_put(&dev->kobj); in blk_register_queue()
|
D | blk-mq-sysfs.c | 270 void blk_mq_unregister_dev(struct device *dev, struct request_queue *q) in blk_mq_unregister_dev() argument 282 kobject_put(&dev->kobj); in blk_mq_unregister_dev() 319 int __blk_mq_register_dev(struct device *dev, struct request_queue *q) in __blk_mq_register_dev() argument 327 ret = kobject_add(q->mq_kobj, kobject_get(&dev->kobj), "%s", "mq"); in __blk_mq_register_dev() 350 kobject_put(&dev->kobj); in __blk_mq_register_dev()
|
D | blk.h | 222 ssize_t part_size_show(struct device *dev, struct device_attribute *attr, 224 ssize_t part_stat_show(struct device *dev, struct device_attribute *attr, 226 ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr, 228 ssize_t part_fail_show(struct device *dev, struct device_attribute *attr, 230 ssize_t part_fail_store(struct device *dev, struct device_attribute *attr,
|
D | bsg.c | 424 dev_t dev; in bsg_register_queue() local 451 dev = MKDEV(bsg_major, bcd->minor); in bsg_register_queue() 452 class_dev = device_create(bsg_class, parent, dev, NULL, "%s", name); in bsg_register_queue() 490 static char *bsg_devnode(struct device *dev, umode_t *mode) in bsg_devnode() argument 492 return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev)); in bsg_devnode()
|
D | blk-settings.c | 821 struct device *dev) in blk_queue_can_use_dma_map_merging() argument 823 unsigned long boundary = dma_get_merge_boundary(dev); in blk_queue_can_use_dma_map_merging()
|
D | blk-mq.h | 129 extern int __blk_mq_register_dev(struct device *dev, struct request_queue *q);
|
D | blk-cgroup.c | 480 if (blkg->q->backing_dev_info->dev) in blkg_dev_name() 820 struct device *dev; in blkcg_fill_root_iostats() local 823 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats() 824 struct gendisk *disk = dev_to_disk(dev); in blkcg_fill_root_iostats()
|
/block/partitions/ |
D | core.c | 175 static ssize_t part_partition_show(struct device *dev, in part_partition_show() argument 178 struct hd_struct *p = dev_to_part(dev); in part_partition_show() 183 static ssize_t part_start_show(struct device *dev, in part_start_show() argument 186 struct hd_struct *p = dev_to_part(dev); in part_start_show() 191 static ssize_t part_ro_show(struct device *dev, in part_ro_show() argument 194 struct hd_struct *p = dev_to_part(dev); in part_ro_show() 198 static ssize_t part_alignment_offset_show(struct device *dev, in part_alignment_offset_show() argument 201 struct hd_struct *p = dev_to_part(dev); in part_alignment_offset_show() 208 static ssize_t part_discard_alignment_show(struct device *dev, in part_discard_alignment_show() argument 211 struct hd_struct *p = dev_to_part(dev); in part_discard_alignment_show() [all …]
|
D | mac.c | 16 extern void note_bootable_part(dev_t dev, int part, int goodness);
|