Home
last modified time | relevance | path

Searched refs:dev (Results 1 – 18 of 18) sorted by relevance

/block/
Dsed-opal.c33 int (*fn)(struct opal_dev *dev, void *data);
36 typedef int (cont_fn)(struct opal_dev *dev);
214 static int end_opal_session_error(struct opal_dev *dev);
215 static int opal_discovery0_step(struct opal_dev *dev);
320 static int opal_send_cmd(struct opal_dev *dev) in opal_send_cmd() argument
322 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_send_cmd()
323 dev->cmd, IO_BUFFER_LENGTH, in opal_send_cmd()
327 static int opal_recv_cmd(struct opal_dev *dev) in opal_recv_cmd() argument
329 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_recv_cmd()
330 dev->resp, IO_BUFFER_LENGTH, in opal_recv_cmd()
[all …]
Dblk-pm.c31 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument
33 q->dev = dev; in blk_pm_runtime_init()
35 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init()
36 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init()
65 if (!q->dev) in blk_pre_runtime_suspend()
99 pm_runtime_mark_last_busy(q->dev); in blk_pre_runtime_suspend()
124 if (!q->dev) in blk_post_runtime_suspend()
132 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_suspend()
154 if (!q->dev) in blk_pre_runtime_resume()
203 if (!q->dev) in blk_set_runtime_active()
[all …]
Dgenhd.c750 if (disk->queue->backing_dev_info->dev) { in register_disk()
752 &disk->queue->backing_dev_info->dev->kobj, in register_disk()
815 struct device *dev = disk_to_dev(disk); in __device_add_disk() local
819 dev->devt = devt; in __device_add_disk()
822 bdi_set_owner(bdi, dev); in __device_add_disk()
958 static ssize_t disk_badblocks_show(struct device *dev, in disk_badblocks_show() argument
962 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_show()
970 static ssize_t disk_badblocks_store(struct device *dev, in disk_badblocks_store() argument
974 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_store()
1070 struct device *dev; in printk_all_partitions() local
[all …]
Dblk-pm.h11 if (q->dev && (q->rpm_status == RPM_SUSPENDED || in blk_pm_request_resume()
13 pm_request_resume(q->dev); in blk_pm_request_resume()
18 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy()
19 pm_runtime_mark_last_busy(rq->q->dev); in blk_pm_mark_last_busy()
26 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_requeue_request()
35 if (q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_add_request()
43 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_put_request()
Dbsg-lib.c153 put_device(job->dev); /* release reference for the request */ in bsg_teardown_job()
224 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job() argument
241 job->dev = dev; in bsg_prepare_job()
243 get_device(job->dev); in bsg_prepare_job()
268 struct device *dev = q->queuedata; in bsg_queue_rq() local
277 if (!get_device(dev)) in bsg_queue_rq()
280 if (!bsg_prepare_job(dev, req)) in bsg_queue_rq()
288 put_device(dev); in bsg_queue_rq()
365 struct request_queue *bsg_setup_queue(struct device *dev, const char *name, in bsg_setup_queue() argument
396 q->queuedata = dev; in bsg_setup_queue()
[all …]
Dblk-timeout.c39 ssize_t part_timeout_show(struct device *dev, struct device_attribute *attr, in part_timeout_show() argument
42 struct gendisk *disk = dev_to_disk(dev); in part_timeout_show()
48 ssize_t part_timeout_store(struct device *dev, struct device_attribute *attr, in part_timeout_store() argument
51 struct gendisk *disk = dev_to_disk(dev); in part_timeout_store()
Dbadblocks.c544 static int __badblocks_init(struct device *dev, struct badblocks *bb, in __badblocks_init() argument
547 bb->dev = dev; in __badblocks_init()
553 if (dev) in __badblocks_init()
554 bb->page = devm_kzalloc(dev, PAGE_SIZE, GFP_KERNEL); in __badblocks_init()
581 int devm_init_badblocks(struct device *dev, struct badblocks *bb) in devm_init_badblocks() argument
585 return __badblocks_init(dev, bb, 1); in devm_init_badblocks()
597 if (bb->dev) in badblocks_exit()
598 devm_kfree(bb->dev, bb->page); in badblocks_exit()
Dkeyslot-manager.c54 if (ksm->dev) in blk_ksm_hw_enter()
55 pm_runtime_get_sync(ksm->dev); in blk_ksm_hw_enter()
62 if (ksm->dev) in blk_ksm_hw_exit()
63 pm_runtime_put_sync(ksm->dev); in blk_ksm_hw_exit()
152 int devm_blk_ksm_init(struct device *dev, struct blk_keyslot_manager *ksm, in devm_blk_ksm_init() argument
160 return devm_add_action_or_reset(dev, blk_ksm_destroy_callback, ksm); in devm_blk_ksm_init()
Dblk-mq-rdma.c25 struct ib_device *dev, int first_vec) in blk_mq_rdma_map_queues() argument
31 mask = ib_get_vector_affinity(dev, first_vec + queue); in blk_mq_rdma_map_queues()
Dblk-sysfs.c843 struct device *dev = disk_to_dev(disk); in blk_register_queue() local
851 kobject_name(&dev->kobj)); in blk_register_queue()
869 ret = blk_trace_init_sysfs(dev); in blk_register_queue()
875 ret = kobject_add(&q->kobj, kobject_get(&dev->kobj), "%s", "queue"); in blk_register_queue()
877 blk_trace_remove_sysfs(dev); in blk_register_queue()
883 blk_trace_remove_sysfs(dev); in blk_register_queue()
885 kobject_put(&dev->kobj); in blk_register_queue()
895 __blk_mq_register_dev(dev, q); in blk_register_queue()
906 blk_trace_remove_sysfs(dev); in blk_register_queue()
907 kobject_put(&dev->kobj); in blk_register_queue()
Dblk-mq-sysfs.c270 void blk_mq_unregister_dev(struct device *dev, struct request_queue *q) in blk_mq_unregister_dev() argument
282 kobject_put(&dev->kobj); in blk_mq_unregister_dev()
319 int __blk_mq_register_dev(struct device *dev, struct request_queue *q) in __blk_mq_register_dev() argument
327 ret = kobject_add(q->mq_kobj, kobject_get(&dev->kobj), "%s", "mq"); in __blk_mq_register_dev()
350 kobject_put(&dev->kobj); in __blk_mq_register_dev()
Dblk.h222 ssize_t part_size_show(struct device *dev, struct device_attribute *attr,
224 ssize_t part_stat_show(struct device *dev, struct device_attribute *attr,
226 ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr,
228 ssize_t part_fail_show(struct device *dev, struct device_attribute *attr,
230 ssize_t part_fail_store(struct device *dev, struct device_attribute *attr,
Dbsg.c424 dev_t dev; in bsg_register_queue() local
451 dev = MKDEV(bsg_major, bcd->minor); in bsg_register_queue()
452 class_dev = device_create(bsg_class, parent, dev, NULL, "%s", name); in bsg_register_queue()
490 static char *bsg_devnode(struct device *dev, umode_t *mode) in bsg_devnode() argument
492 return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev)); in bsg_devnode()
Dblk-settings.c821 struct device *dev) in blk_queue_can_use_dma_map_merging() argument
823 unsigned long boundary = dma_get_merge_boundary(dev); in blk_queue_can_use_dma_map_merging()
Dblk-mq.h129 extern int __blk_mq_register_dev(struct device *dev, struct request_queue *q);
Dblk-cgroup.c480 if (blkg->q->backing_dev_info->dev) in blkg_dev_name()
820 struct device *dev; in blkcg_fill_root_iostats() local
823 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats()
824 struct gendisk *disk = dev_to_disk(dev); in blkcg_fill_root_iostats()
/block/partitions/
Dcore.c175 static ssize_t part_partition_show(struct device *dev, in part_partition_show() argument
178 struct hd_struct *p = dev_to_part(dev); in part_partition_show()
183 static ssize_t part_start_show(struct device *dev, in part_start_show() argument
186 struct hd_struct *p = dev_to_part(dev); in part_start_show()
191 static ssize_t part_ro_show(struct device *dev, in part_ro_show() argument
194 struct hd_struct *p = dev_to_part(dev); in part_ro_show()
198 static ssize_t part_alignment_offset_show(struct device *dev, in part_alignment_offset_show() argument
201 struct hd_struct *p = dev_to_part(dev); in part_alignment_offset_show()
208 static ssize_t part_discard_alignment_show(struct device *dev, in part_discard_alignment_show() argument
211 struct hd_struct *p = dev_to_part(dev); in part_discard_alignment_show()
[all …]
Dmac.c16 extern void note_bootable_part(dev_t dev, int part, int goodness);