/block/ |
D | blk-ioc.c | 47 if (icq->flags & ICQ_EXITED) in ioc_exit_icq() 53 icq->flags |= ICQ_EXITED; in ioc_exit_icq() 98 unsigned long flags; in ioc_release_fn() local 106 spin_lock_irqsave_nested(&ioc->lock, flags, 1); in ioc_release_fn() 117 spin_unlock_irqrestore(&ioc->lock, flags); in ioc_release_fn() 119 spin_lock_irqsave_nested(&ioc->lock, flags, 1); in ioc_release_fn() 123 spin_unlock_irqrestore(&ioc->lock, flags); in ioc_release_fn() 137 unsigned long flags; in put_io_context() local 150 spin_lock_irqsave(&ioc->lock, flags); in put_io_context() 156 spin_unlock_irqrestore(&ioc->lock, flags); in put_io_context() [all …]
|
D | blk-flush.c | 97 struct blk_flush_queue *fq, unsigned int flags); 212 unsigned long flags = 0; in flush_end_io() local 217 spin_lock_irqsave(&fq->mq_flush_lock, flags); in flush_end_io() 221 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in flush_end_io() 252 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in flush_end_io() 269 unsigned int flags) in blk_kick_flush() argument 319 flush_rq->cmd_flags |= (flags & REQ_DRV) | (flags & REQ_FAILFAST_MASK); in blk_kick_flush() 332 unsigned long flags; in mq_flush_data_end_io() local 344 spin_lock_irqsave(&fq->mq_flush_lock, flags); in mq_flush_data_end_io() 346 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in mq_flush_data_end_io() [all …]
|
D | blk-lib.c | 26 sector_t nr_sects, gfp_t gfp_mask, int flags, in __blkdev_issue_discard() argument 40 if (flags & BLKDEV_DISCARD_SECURE) { in __blkdev_issue_discard() 98 sector_t nr_sects, gfp_t gfp_mask, unsigned long flags) in blkdev_issue_discard() argument 105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags, in blkdev_issue_discard() 214 struct bio **biop, unsigned flags) in __blkdev_issue_write_zeroes() argument 237 if (flags & BLKDEV_ZERO_NOUNMAP) in __blkdev_issue_write_zeroes() 326 unsigned flags) in __blkdev_issue_zeroout() argument 336 biop, flags); in __blkdev_issue_zeroout() 337 if (ret != -EOPNOTSUPP || (flags & BLKDEV_ZERO_NOFALLBACK)) in __blkdev_issue_zeroout() 359 sector_t nr_sects, gfp_t gfp_mask, unsigned flags) in blkdev_issue_zeroout() argument [all …]
|
D | blk-softirq.c | 45 unsigned long flags; in trigger_softirq() local 48 local_irq_save(flags); in trigger_softirq() 55 local_irq_restore(flags); in trigger_softirq() 68 data->flags = 0; in raise_blk_irq() 102 unsigned long flags; in __blk_complete_request() local 107 local_irq_save(flags); in __blk_complete_request() 144 local_irq_restore(flags); in __blk_complete_request()
|
D | genhd.c | 160 unsigned int flags) in disk_part_iter_init() argument 170 if (flags & DISK_PITER_REVERSE) in disk_part_iter_init() 172 else if (flags & (DISK_PITER_INCL_PART0 | DISK_PITER_INCL_EMPTY_PART0)) in disk_part_iter_init() 177 piter->flags = flags; in disk_part_iter_init() 206 if (piter->flags & DISK_PITER_REVERSE) { in disk_part_iter_next() 208 if (piter->flags & (DISK_PITER_INCL_PART0 | in disk_part_iter_next() 226 !(piter->flags & DISK_PITER_INCL_EMPTY) && in disk_part_iter_next() 227 !(piter->flags & DISK_PITER_INCL_EMPTY_PART0 && in disk_part_iter_next() 637 if (disk->flags & GENHD_FL_HIDDEN) { in register_disk() 713 !(disk->flags & (GENHD_FL_EXT_DEVT | GENHD_FL_HIDDEN))); in __device_add_disk() [all …]
|
D | blk-integrity.c | 276 bi->flags |= BLK_INTEGRITY_VERIFY; in integrity_verify_store() 278 bi->flags &= ~BLK_INTEGRITY_VERIFY; in integrity_verify_store() 285 return sprintf(page, "%d\n", (bi->flags & BLK_INTEGRITY_VERIFY) != 0); in integrity_verify_show() 295 bi->flags |= BLK_INTEGRITY_GENERATE; in integrity_generate_store() 297 bi->flags &= ~BLK_INTEGRITY_GENERATE; in integrity_generate_store() 304 return sprintf(page, "%d\n", (bi->flags & BLK_INTEGRITY_GENERATE) != 0); in integrity_generate_show() 310 (bi->flags & BLK_INTEGRITY_DEVICE_CAPABLE) != 0); in integrity_device_show() 403 bi->flags = BLK_INTEGRITY_VERIFY | BLK_INTEGRITY_GENERATE | in blk_integrity_register() 404 template->flags; in blk_integrity_register()
|
D | blk-wbt.c | 549 enum wbt_flags flags = 0; in bio_to_wbt_flags() local 555 flags = WBT_READ; in bio_to_wbt_flags() 558 flags |= WBT_KSWAPD; in bio_to_wbt_flags() 560 flags |= WBT_DISCARD; in bio_to_wbt_flags() 561 flags |= WBT_TRACKED; in bio_to_wbt_flags() 563 return flags; in bio_to_wbt_flags() 569 enum wbt_flags flags = bio_to_wbt_flags(rwb, bio); in wbt_cleanup() local 570 __wbt_done(rqos, flags); in wbt_cleanup() 582 enum wbt_flags flags; in wbt_wait() local 584 flags = bio_to_wbt_flags(rwb, bio); in wbt_wait() [all …]
|
D | blk-mq.c | 243 if (hctx->flags & BLK_MQ_F_BLOCKING) in blk_mq_quiesce_queue() 301 if (data->flags & BLK_MQ_REQ_INTERNAL) { in blk_mq_rq_ctx_init() 305 if (data->hctx->flags & BLK_MQ_F_TAG_SHARED) { in blk_mq_rq_ctx_init() 320 if (data->flags & BLK_MQ_REQ_PREEMPT) in blk_mq_rq_ctx_init() 381 data->flags |= BLK_MQ_REQ_NOWAIT; in blk_mq_get_request() 384 data->flags |= BLK_MQ_REQ_INTERNAL; in blk_mq_get_request() 393 !(data->flags & BLK_MQ_REQ_RESERVED)) in blk_mq_get_request() 423 blk_mq_req_flags_t flags) in blk_mq_alloc_request() argument 425 struct blk_mq_alloc_data alloc_data = { .flags = flags, .cmd_flags = op }; in blk_mq_alloc_request() 429 ret = blk_queue_enter(q, flags); in blk_mq_alloc_request() [all …]
|
D | blk-mq.h | 103 unsigned int flags, in blk_mq_map_queue() argument 111 if (flags & REQ_HIPRI) in blk_mq_map_queue() 113 else if ((flags & REQ_OP_MASK) == REQ_OP_READ) in blk_mq_map_queue() 160 blk_mq_req_flags_t flags; member 171 if (data->flags & BLK_MQ_REQ_INTERNAL) in blk_mq_tags_from_data()
|
D | ioctl.c | 202 unsigned long arg, unsigned long flags) in blk_ioctl_discard() argument 231 GFP_KERNEL, flags); in blk_ioctl_discard() 328 if (reg.flags & ~PR_FL_IGNORE_KEY) in blkdev_pr_register() 330 return ops->pr_register(bdev, reg.old_key, reg.new_key, reg.flags); in blkdev_pr_register() 346 if (rsv.flags & ~PR_FL_IGNORE_KEY) in blkdev_pr_reserve() 348 return ops->pr_reserve(bdev, rsv.key, rsv.type, rsv.flags); in blkdev_pr_reserve() 364 if (rsv.flags) in blkdev_pr_release() 382 if (p.flags) in blkdev_pr_preempt() 400 if (c.flags) in blkdev_pr_clear()
|
D | keyslot-manager.c | 160 unsigned long flags; in remove_slot_from_lru_list() local 162 spin_lock_irqsave(&ksm->idle_slots_lock, flags); in remove_slot_from_lru_list() 164 spin_unlock_irqrestore(&ksm->idle_slots_lock, flags); in remove_slot_from_lru_list() 305 unsigned long flags; in keyslot_manager_put_slot() local 314 &ksm->idle_slots_lock, flags)) { in keyslot_manager_put_slot() 317 spin_unlock_irqrestore(&ksm->idle_slots_lock, flags); in keyslot_manager_put_slot()
|
D | mq-deadline.c | 204 unsigned long flags; in deadline_fifo_request() local 220 spin_lock_irqsave(&dd->zone_lock, flags); in deadline_fifo_request() 227 spin_unlock_irqrestore(&dd->zone_lock, flags); in deadline_fifo_request() 240 unsigned long flags; in deadline_next_request() local 256 spin_lock_irqsave(&dd->zone_lock, flags); in deadline_next_request() 262 spin_unlock_irqrestore(&dd->zone_lock, flags); in deadline_next_request() 568 unsigned long flags; in dd_finish_request() local 570 spin_lock_irqsave(&dd->zone_lock, flags); in dd_finish_request() 574 spin_unlock_irqrestore(&dd->zone_lock, flags); in dd_finish_request()
|
D | partition-generic.c | 310 sector_t start, sector_t len, int flags, in add_partition() argument 388 if (flags & ADDPART_FLAG_WHOLEDISK) { in add_partition() 396 if (flags & ADDPART_FLAG_WHOLEDISK) in add_partition() 431 !(disk->flags & GENHD_FL_NATIVE_CAPACITY)) { in disk_unlock_native_capacity() 434 disk->flags |= GENHD_FL_NATIVE_CAPACITY; in disk_unlock_native_capacity() 569 state->parts[p].flags, in rescan_partitions() 577 if (state->parts[p].flags & ADDPART_FLAG_RAID) in rescan_partitions()
|
D | blk-mq-tag.h | 58 if (!(hctx->flags & BLK_MQ_F_TAG_SHARED)) in blk_mq_tag_busy() 66 if (!(hctx->flags & BLK_MQ_F_TAG_SHARED)) in blk_mq_tag_idle()
|
D | bio.c | 297 unsigned long flags = bio->bi_flags & (~0UL << BIO_RESET_BITS); in bio_reset() local 302 bio->bi_flags = flags; in bio_reset() 531 unsigned long flags; in zero_fill_bio_iter() local 536 char *data = bvec_kmap_irq(&bv, &flags); in zero_fill_bio_iter() 539 bvec_kunmap_irq(data, &flags); in zero_fill_bio_iter() 1737 unsigned long flags; in bio_check_pages_dirty() local 1749 spin_lock_irqsave(&bio_dirty_lock, flags); in bio_check_pages_dirty() 1752 spin_unlock_irqrestore(&bio_dirty_lock, flags); in bio_check_pages_dirty() 2003 int flags) in bioset_init() argument 2020 if ((flags & BIOSET_NEED_BVECS) && in bioset_init() [all …]
|
D | blk-mq-tag.c | 76 if (!hctx || !(hctx->flags & BLK_MQ_F_TAG_SHARED)) in hctx_may_queue() 101 if (!(data->flags & BLK_MQ_REQ_INTERNAL) && in __blk_mq_get_tag() 119 if (data->flags & BLK_MQ_REQ_RESERVED) { in blk_mq_get_tag() 135 if (data->flags & BLK_MQ_REQ_NOWAIT) in blk_mq_get_tag() 172 if (data->flags & BLK_MQ_REQ_RESERVED) in blk_mq_get_tag()
|
D | cmdline-parser.c | 60 new_subpart->flags = 0; in parse_subpart() 63 new_subpart->flags |= PF_RDONLY; in parse_subpart() 68 new_subpart->flags |= PF_POWERUP_LOCK; in parse_subpart()
|
D | blk-iolatency.c | 526 unsigned long flags; in iolatency_check_latencies() local 553 spin_lock_irqsave(&lat_info->lock, flags); in iolatency_check_latencies() 585 spin_unlock_irqrestore(&lat_info->lock, flags); in iolatency_check_latencies() 668 unsigned long flags; in blkiolatency_timer_fn() local 688 spin_lock_irqsave(&lat_info->lock, flags); in blkiolatency_timer_fn() 710 spin_unlock_irqrestore(&lat_info->lock, flags); in blkiolatency_timer_fn() 1025 .flags = CFTYPE_NOT_ON_ROOT,
|
D | bfq-cgroup.c | 106 stats->flags |= (1 << BFQG_stats_##name); \ 110 stats->flags &= ~(1 << BFQG_stats_##name); \ 114 return (stats->flags & (1 << BFQG_stats_##name)) != 0; \ 844 unsigned long flags; in bfq_pd_offline() local 847 spin_lock_irqsave(&bfqd->lock, flags); in bfq_pd_offline() 886 spin_unlock_irqrestore(&bfqd->lock, flags); in bfq_pd_offline() 1215 .flags = CFTYPE_NOT_ON_ROOT, 1221 .flags = CFTYPE_NOT_ON_ROOT, 1341 .flags = CFTYPE_NOT_ON_ROOT,
|
D | blk-iocost.c | 198 unsigned long flags; \ 200 spin_lock_irqsave(&trace_iocg_path_lock, flags); \ 205 spin_unlock_irqrestore(&trace_iocg_path_lock, flags); \ 1117 int flags, void *key) in iocg_wake_fn() argument 1138 default_wake_function(wq_entry, mode, flags, key); in iocg_wake_fn() 1204 unsigned long flags; in iocg_waitq_timer_fn() local 1208 spin_lock_irqsave(&iocg->waitq.lock, flags); in iocg_waitq_timer_fn() 1210 spin_unlock_irqrestore(&iocg->waitq.lock, flags); in iocg_waitq_timer_fn() 1997 unsigned long flags; in ioc_pd_init() local 2023 spin_lock_irqsave(&ioc->lock, flags); in ioc_pd_init() [all …]
|
D | bio-integrity.c | 226 !(bi->flags & BLK_INTEGRITY_VERIFY)) in bio_integrity_prep() 230 !(bi->flags & BLK_INTEGRITY_GENERATE)) in bio_integrity_prep() 261 if (bi->flags & BLK_INTEGRITY_IP_CHECKSUM) in bio_integrity_prep()
|
D | blk-core.c | 404 int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) in blk_queue_enter() argument 406 const bool pm = flags & BLK_MQ_REQ_PREEMPT; in blk_queue_enter() 429 if (flags & BLK_MQ_REQ_NOWAIT) in blk_queue_enter() 582 blk_mq_req_flags_t flags) in blk_get_request() argument 587 WARN_ON_ONCE(flags & ~(BLK_MQ_REQ_NOWAIT | BLK_MQ_REQ_PREEMPT)); in blk_get_request() 589 req = blk_mq_alloc_request(q, op, flags); in blk_get_request() 1061 blk_mq_req_flags_t flags = bio->bi_opf & REQ_NOWAIT ? in generic_make_request() local 1064 if (likely(blk_queue_enter(q, flags) == 0)) { in generic_make_request()
|
D | compat_ioctl.c | 179 compat_int_t flags; member 194 err |= get_user(n, &ua32->flags); in compat_blkpg_ioctl() 195 err |= put_user(n, &a->flags); in compat_blkpg_ioctl()
|
/block/partitions/ |
D | sun.c | 29 __be16 flags; in sun_partition() member 110 state->parts[slot].flags = 0; in sun_partition() 113 state->parts[slot].flags |= ADDPART_FLAG_RAID; in sun_partition() 115 state->parts[slot].flags |= ADDPART_FLAG_WHOLEDISK; in sun_partition()
|
D | check.h | 16 int flags; member
|