Home
last modified time | relevance | path

Searched refs:dev (Results 1 – 22 of 22) sorted by relevance

/block/
Dsed-opal.c33 int (*fn)(struct opal_dev *dev, void *data);
36 typedef int (cont_fn)(struct opal_dev *dev);
214 static int end_opal_session_error(struct opal_dev *dev);
215 static int opal_discovery0_step(struct opal_dev *dev);
320 static int opal_send_cmd(struct opal_dev *dev) in opal_send_cmd() argument
322 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_send_cmd()
323 dev->cmd, IO_BUFFER_LENGTH, in opal_send_cmd()
327 static int opal_recv_cmd(struct opal_dev *dev) in opal_recv_cmd() argument
329 return dev->send_recv(dev->data, dev->comid, TCG_SECP_01, in opal_recv_cmd()
330 dev->resp, IO_BUFFER_LENGTH, in opal_recv_cmd()
[all …]
Dblk-pm.c31 void blk_pm_runtime_init(struct request_queue *q, struct device *dev) in blk_pm_runtime_init() argument
33 q->dev = dev; in blk_pm_runtime_init()
35 pm_runtime_set_autosuspend_delay(q->dev, -1); in blk_pm_runtime_init()
36 pm_runtime_use_autosuspend(q->dev); in blk_pm_runtime_init()
65 if (!q->dev) in blk_pre_runtime_suspend()
99 pm_runtime_mark_last_busy(q->dev); in blk_pre_runtime_suspend()
124 if (!q->dev) in blk_post_runtime_suspend()
132 pm_runtime_mark_last_busy(q->dev); in blk_post_runtime_suspend()
154 if (!q->dev) in blk_pre_runtime_resume()
203 if (!q->dev) in blk_set_runtime_active()
[all …]
Dgenhd.c508 &disk->bdi->dev->kobj, "bdi"); in device_add_disk()
655 static ssize_t disk_badblocks_show(struct device *dev, in disk_badblocks_show() argument
659 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_show()
667 static ssize_t disk_badblocks_store(struct device *dev, in disk_badblocks_store() argument
671 struct gendisk *disk = dev_to_disk(dev); in disk_badblocks_store()
707 struct device *dev; in printk_all_partitions() local
710 while ((dev = class_dev_iter_next(&iter))) { in printk_all_partitions()
711 struct gendisk *disk = dev_to_disk(dev); in printk_all_partitions()
740 else if (dev->parent && dev->parent->driver) in printk_all_partitions()
742 dev->parent->driver->name); in printk_all_partitions()
[all …]
Dbsg-lib.c154 put_device(job->dev); /* release reference for the request */ in bsg_teardown_job()
225 static bool bsg_prepare_job(struct device *dev, struct request *req) in bsg_prepare_job() argument
242 job->dev = dev; in bsg_prepare_job()
244 get_device(job->dev); in bsg_prepare_job()
269 struct device *dev = q->queuedata; in bsg_queue_rq() local
278 if (!get_device(dev)) in bsg_queue_rq()
281 if (!bsg_prepare_job(dev, req)) in bsg_queue_rq()
289 put_device(dev); in bsg_queue_rq()
366 struct request_queue *bsg_setup_queue(struct device *dev, const char *name, in bsg_setup_queue() argument
397 q->queuedata = dev; in bsg_setup_queue()
[all …]
Dblk-pm.h11 if (!q->dev || !blk_queue_pm_only(q)) in blk_pm_resume_queue()
15 pm_request_resume(q->dev); in blk_pm_resume_queue()
21 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy()
22 pm_runtime_mark_last_busy(rq->q->dev); in blk_pm_mark_last_busy()
Dbdev.c505 void bdev_add(struct block_device *bdev, dev_t dev) in bdev_add() argument
507 bdev->bd_dev = dev; in bdev_add()
508 bdev->bd_inode->i_rdev = dev; in bdev_add()
509 bdev->bd_inode->i_ino = dev; in bdev_add()
733 struct block_device *blkdev_get_no_open(dev_t dev) in blkdev_get_no_open() argument
738 inode = ilookup(blockdev_superblock, dev); in blkdev_get_no_open()
740 blk_request_module(dev); in blkdev_get_no_open()
741 inode = ilookup(blockdev_superblock, dev); in blkdev_get_no_open()
790 struct block_device *blkdev_get_by_dev(dev_t dev, fmode_t mode, void *holder) in blkdev_get_by_dev() argument
798 MAJOR(dev), MINOR(dev), in blkdev_get_by_dev()
[all …]
Dblk-timeout.c39 ssize_t part_timeout_show(struct device *dev, struct device_attribute *attr, in part_timeout_show() argument
42 struct gendisk *disk = dev_to_disk(dev); in part_timeout_show()
48 ssize_t part_timeout_store(struct device *dev, struct device_attribute *attr, in part_timeout_store() argument
51 struct gendisk *disk = dev_to_disk(dev); in part_timeout_store()
Ddisk-events.c357 static ssize_t disk_events_show(struct device *dev, in disk_events_show() argument
360 struct gendisk *disk = dev_to_disk(dev); in disk_events_show()
367 static ssize_t disk_events_async_show(struct device *dev, in disk_events_async_show() argument
373 static ssize_t disk_events_poll_msecs_show(struct device *dev, in disk_events_poll_msecs_show() argument
377 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_show()
384 static ssize_t disk_events_poll_msecs_store(struct device *dev, in disk_events_poll_msecs_store() argument
388 struct gendisk *disk = dev_to_disk(dev); in disk_events_poll_msecs_store()
Dbadblocks.c544 static int __badblocks_init(struct device *dev, struct badblocks *bb, in __badblocks_init() argument
547 bb->dev = dev; in __badblocks_init()
553 if (dev) in __badblocks_init()
554 bb->page = devm_kzalloc(dev, PAGE_SIZE, GFP_KERNEL); in __badblocks_init()
581 int devm_init_badblocks(struct device *dev, struct badblocks *bb) in devm_init_badblocks() argument
585 return __badblocks_init(dev, bb, 1); in devm_init_badblocks()
597 if (bb->dev) in badblocks_exit()
598 devm_kfree(bb->dev, bb->page); in badblocks_exit()
Dkeyslot-manager.c54 if (ksm->dev) in blk_ksm_hw_enter()
55 pm_runtime_get_sync(ksm->dev); in blk_ksm_hw_enter()
62 if (ksm->dev) in blk_ksm_hw_exit()
63 pm_runtime_put_sync(ksm->dev); in blk_ksm_hw_exit()
152 int devm_blk_ksm_init(struct device *dev, struct blk_keyslot_manager *ksm, in devm_blk_ksm_init() argument
160 return devm_add_action_or_reset(dev, blk_ksm_destroy_callback, ksm); in devm_blk_ksm_init()
Dblk-crypto-profile.c51 if (profile->dev) in blk_crypto_hw_enter()
52 pm_runtime_get_sync(profile->dev); in blk_crypto_hw_enter()
59 if (profile->dev) in blk_crypto_hw_exit()
60 pm_runtime_put_sync(profile->dev); in blk_crypto_hw_exit()
158 int devm_blk_crypto_profile_init(struct device *dev, in devm_blk_crypto_profile_init() argument
167 return devm_add_action_or_reset(dev, in devm_blk_crypto_profile_init()
Dblk-sysfs.c850 struct device *dev = disk_to_dev(disk); in blk_register_queue() local
853 ret = blk_trace_init_sysfs(dev); in blk_register_queue()
859 ret = kobject_add(&q->kobj, kobject_get(&dev->kobj), "%s", "queue"); in blk_register_queue()
861 blk_trace_remove_sysfs(dev); in blk_register_queue()
867 blk_trace_remove_sysfs(dev); in blk_register_queue()
869 kobject_put(&dev->kobj); in blk_register_queue()
879 __blk_mq_register_dev(dev, q); in blk_register_queue()
890 blk_trace_remove_sysfs(dev); in blk_register_queue()
891 kobject_put(&dev->kobj); in blk_register_queue()
935 blk_trace_remove_sysfs(dev); in blk_register_queue()
[all …]
Dblk-mq-sysfs.c215 void blk_mq_unregister_dev(struct device *dev, struct request_queue *q) in blk_mq_unregister_dev() argument
227 kobject_put(&dev->kobj); in blk_mq_unregister_dev()
264 int __blk_mq_register_dev(struct device *dev, struct request_queue *q) in __blk_mq_register_dev() argument
272 ret = kobject_add(q->mq_kobj, kobject_get(&dev->kobj), "%s", "mq"); in __blk_mq_register_dev()
295 kobject_put(&dev->kobj); in __blk_mq_register_dev()
Dblk-mq-rdma.c25 struct ib_device *dev, int first_vec) in blk_mq_rdma_map_queues() argument
31 mask = ib_get_vector_affinity(dev, first_vec + queue); in blk_mq_rdma_map_queues()
Dbsg.c168 static void bsg_device_release(struct device *dev) in bsg_device_release() argument
170 struct bsg_device *bd = container_of(dev, struct bsg_device, device); in bsg_device_release()
235 static char *bsg_devnode(struct device *dev, umode_t *mode) in bsg_devnode() argument
237 return kasprintf(GFP_KERNEL, "bsg/%s", dev_name(dev)); in bsg_devnode()
Dblk.h254 ssize_t part_size_show(struct device *dev, struct device_attribute *attr,
256 ssize_t part_stat_show(struct device *dev, struct device_attribute *attr,
258 ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr,
260 ssize_t part_fail_show(struct device *dev, struct device_attribute *attr,
262 ssize_t part_fail_store(struct device *dev, struct device_attribute *attr,
Dkyber-iosched.c154 dev_t dev; member
261 trace_kyber_latency(kqd->dev, kyber_domain_names[sched_domain], in calculate_percentile()
274 trace_kyber_adjust(kqd->dev, kyber_domain_names[sched_domain], in kyber_resize_domain()
370 kqd->dev = disk_devt(q->disk); in kyber_queue_data_alloc()
779 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain()
792 trace_kyber_throttled(kqd->dev, in kyber_dispatch_cur_domain()
Dblk-settings.c901 struct device *dev) in blk_queue_can_use_dma_map_merging() argument
903 unsigned long boundary = dma_get_merge_boundary(dev); in blk_queue_can_use_dma_map_merging()
Dblk-mq.h129 extern int __blk_mq_register_dev(struct device *dev, struct request_queue *q);
Dblk-cgroup.c505 if (!blkg->q->disk || !blkg->q->disk->bdi->dev) in blkg_dev_name()
862 struct device *dev; in blkcg_fill_root_iostats() local
865 while ((dev = class_dev_iter_next(&iter))) { in blkcg_fill_root_iostats()
866 struct block_device *bdev = dev_to_bdev(dev); in blkcg_fill_root_iostats()
/block/partitions/
Dcore.c183 static ssize_t part_partition_show(struct device *dev, in part_partition_show() argument
186 return sprintf(buf, "%d\n", dev_to_bdev(dev)->bd_partno); in part_partition_show()
189 static ssize_t part_start_show(struct device *dev, in part_start_show() argument
192 return sprintf(buf, "%llu\n", dev_to_bdev(dev)->bd_start_sect); in part_start_show()
195 static ssize_t part_ro_show(struct device *dev, in part_ro_show() argument
198 return sprintf(buf, "%d\n", bdev_read_only(dev_to_bdev(dev))); in part_ro_show()
201 static ssize_t part_alignment_offset_show(struct device *dev, in part_alignment_offset_show() argument
204 struct block_device *bdev = dev_to_bdev(dev); in part_alignment_offset_show()
211 static ssize_t part_discard_alignment_show(struct device *dev, in part_discard_alignment_show() argument
214 struct block_device *bdev = dev_to_bdev(dev); in part_discard_alignment_show()
[all …]
Dmac.c16 extern void note_bootable_part(dev_t dev, int part, int goodness);