Home
last modified time | relevance | path

Searched refs:flags (Results 1 – 25 of 47) sorted by relevance

12

/block/
Dblk-ioc.c47 if (icq->flags & ICQ_EXITED) in ioc_exit_icq()
53 icq->flags |= ICQ_EXITED; in ioc_exit_icq()
98 unsigned long flags; in ioc_release_fn() local
106 spin_lock_irqsave_nested(&ioc->lock, flags, 1); in ioc_release_fn()
117 spin_unlock_irqrestore(&ioc->lock, flags); in ioc_release_fn()
119 spin_lock_irqsave_nested(&ioc->lock, flags, 1); in ioc_release_fn()
123 spin_unlock_irqrestore(&ioc->lock, flags); in ioc_release_fn()
137 unsigned long flags; in put_io_context() local
150 spin_lock_irqsave(&ioc->lock, flags); in put_io_context()
156 spin_unlock_irqrestore(&ioc->lock, flags); in put_io_context()
[all …]
Dblk-flush.c97 struct blk_flush_queue *fq, unsigned int flags);
212 unsigned long flags = 0; in flush_end_io() local
217 spin_lock_irqsave(&fq->mq_flush_lock, flags); in flush_end_io()
221 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in flush_end_io()
252 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in flush_end_io()
269 unsigned int flags) in blk_kick_flush() argument
319 flush_rq->cmd_flags |= (flags & REQ_DRV) | (flags & REQ_FAILFAST_MASK); in blk_kick_flush()
332 unsigned long flags; in mq_flush_data_end_io() local
344 spin_lock_irqsave(&fq->mq_flush_lock, flags); in mq_flush_data_end_io()
346 spin_unlock_irqrestore(&fq->mq_flush_lock, flags); in mq_flush_data_end_io()
[all …]
Dblk-lib.c26 sector_t nr_sects, gfp_t gfp_mask, int flags, in __blkdev_issue_discard() argument
40 if (flags & BLKDEV_DISCARD_SECURE) { in __blkdev_issue_discard()
98 sector_t nr_sects, gfp_t gfp_mask, unsigned long flags) in blkdev_issue_discard() argument
105 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags, in blkdev_issue_discard()
214 struct bio **biop, unsigned flags) in __blkdev_issue_write_zeroes() argument
237 if (flags & BLKDEV_ZERO_NOUNMAP) in __blkdev_issue_write_zeroes()
326 unsigned flags) in __blkdev_issue_zeroout() argument
336 biop, flags); in __blkdev_issue_zeroout()
337 if (ret != -EOPNOTSUPP || (flags & BLKDEV_ZERO_NOFALLBACK)) in __blkdev_issue_zeroout()
359 sector_t nr_sects, gfp_t gfp_mask, unsigned flags) in blkdev_issue_zeroout() argument
[all …]
Dblk-softirq.c45 unsigned long flags; in trigger_softirq() local
48 local_irq_save(flags); in trigger_softirq()
55 local_irq_restore(flags); in trigger_softirq()
68 data->flags = 0; in raise_blk_irq()
102 unsigned long flags; in __blk_complete_request() local
107 local_irq_save(flags); in __blk_complete_request()
144 local_irq_restore(flags); in __blk_complete_request()
Dgenhd.c160 unsigned int flags) in disk_part_iter_init() argument
170 if (flags & DISK_PITER_REVERSE) in disk_part_iter_init()
172 else if (flags & (DISK_PITER_INCL_PART0 | DISK_PITER_INCL_EMPTY_PART0)) in disk_part_iter_init()
177 piter->flags = flags; in disk_part_iter_init()
206 if (piter->flags & DISK_PITER_REVERSE) { in disk_part_iter_next()
208 if (piter->flags & (DISK_PITER_INCL_PART0 | in disk_part_iter_next()
226 !(piter->flags & DISK_PITER_INCL_EMPTY) && in disk_part_iter_next()
227 !(piter->flags & DISK_PITER_INCL_EMPTY_PART0 && in disk_part_iter_next()
637 if (disk->flags & GENHD_FL_HIDDEN) { in register_disk()
713 !(disk->flags & (GENHD_FL_EXT_DEVT | GENHD_FL_HIDDEN))); in __device_add_disk()
[all …]
Dblk-integrity.c276 bi->flags |= BLK_INTEGRITY_VERIFY; in integrity_verify_store()
278 bi->flags &= ~BLK_INTEGRITY_VERIFY; in integrity_verify_store()
285 return sprintf(page, "%d\n", (bi->flags & BLK_INTEGRITY_VERIFY) != 0); in integrity_verify_show()
295 bi->flags |= BLK_INTEGRITY_GENERATE; in integrity_generate_store()
297 bi->flags &= ~BLK_INTEGRITY_GENERATE; in integrity_generate_store()
304 return sprintf(page, "%d\n", (bi->flags & BLK_INTEGRITY_GENERATE) != 0); in integrity_generate_show()
310 (bi->flags & BLK_INTEGRITY_DEVICE_CAPABLE) != 0); in integrity_device_show()
403 bi->flags = BLK_INTEGRITY_VERIFY | BLK_INTEGRITY_GENERATE | in blk_integrity_register()
404 template->flags; in blk_integrity_register()
Dblk-wbt.c549 enum wbt_flags flags = 0; in bio_to_wbt_flags() local
555 flags = WBT_READ; in bio_to_wbt_flags()
558 flags |= WBT_KSWAPD; in bio_to_wbt_flags()
560 flags |= WBT_DISCARD; in bio_to_wbt_flags()
561 flags |= WBT_TRACKED; in bio_to_wbt_flags()
563 return flags; in bio_to_wbt_flags()
569 enum wbt_flags flags = bio_to_wbt_flags(rwb, bio); in wbt_cleanup() local
570 __wbt_done(rqos, flags); in wbt_cleanup()
582 enum wbt_flags flags; in wbt_wait() local
584 flags = bio_to_wbt_flags(rwb, bio); in wbt_wait()
[all …]
Dblk-mq.c243 if (hctx->flags & BLK_MQ_F_BLOCKING) in blk_mq_quiesce_queue()
301 if (data->flags & BLK_MQ_REQ_INTERNAL) { in blk_mq_rq_ctx_init()
305 if (data->hctx->flags & BLK_MQ_F_TAG_SHARED) { in blk_mq_rq_ctx_init()
320 if (data->flags & BLK_MQ_REQ_PREEMPT) in blk_mq_rq_ctx_init()
381 data->flags |= BLK_MQ_REQ_NOWAIT; in blk_mq_get_request()
384 data->flags |= BLK_MQ_REQ_INTERNAL; in blk_mq_get_request()
393 !(data->flags & BLK_MQ_REQ_RESERVED)) in blk_mq_get_request()
423 blk_mq_req_flags_t flags) in blk_mq_alloc_request() argument
425 struct blk_mq_alloc_data alloc_data = { .flags = flags, .cmd_flags = op }; in blk_mq_alloc_request()
429 ret = blk_queue_enter(q, flags); in blk_mq_alloc_request()
[all …]
Dblk-mq.h103 unsigned int flags, in blk_mq_map_queue() argument
111 if (flags & REQ_HIPRI) in blk_mq_map_queue()
113 else if ((flags & REQ_OP_MASK) == REQ_OP_READ) in blk_mq_map_queue()
160 blk_mq_req_flags_t flags; member
171 if (data->flags & BLK_MQ_REQ_INTERNAL) in blk_mq_tags_from_data()
Dioctl.c202 unsigned long arg, unsigned long flags) in blk_ioctl_discard() argument
231 GFP_KERNEL, flags); in blk_ioctl_discard()
328 if (reg.flags & ~PR_FL_IGNORE_KEY) in blkdev_pr_register()
330 return ops->pr_register(bdev, reg.old_key, reg.new_key, reg.flags); in blkdev_pr_register()
346 if (rsv.flags & ~PR_FL_IGNORE_KEY) in blkdev_pr_reserve()
348 return ops->pr_reserve(bdev, rsv.key, rsv.type, rsv.flags); in blkdev_pr_reserve()
364 if (rsv.flags) in blkdev_pr_release()
382 if (p.flags) in blkdev_pr_preempt()
400 if (c.flags) in blkdev_pr_clear()
Dkeyslot-manager.c160 unsigned long flags; in remove_slot_from_lru_list() local
162 spin_lock_irqsave(&ksm->idle_slots_lock, flags); in remove_slot_from_lru_list()
164 spin_unlock_irqrestore(&ksm->idle_slots_lock, flags); in remove_slot_from_lru_list()
305 unsigned long flags; in keyslot_manager_put_slot() local
314 &ksm->idle_slots_lock, flags)) { in keyslot_manager_put_slot()
317 spin_unlock_irqrestore(&ksm->idle_slots_lock, flags); in keyslot_manager_put_slot()
Dmq-deadline.c204 unsigned long flags; in deadline_fifo_request() local
220 spin_lock_irqsave(&dd->zone_lock, flags); in deadline_fifo_request()
227 spin_unlock_irqrestore(&dd->zone_lock, flags); in deadline_fifo_request()
240 unsigned long flags; in deadline_next_request() local
256 spin_lock_irqsave(&dd->zone_lock, flags); in deadline_next_request()
262 spin_unlock_irqrestore(&dd->zone_lock, flags); in deadline_next_request()
568 unsigned long flags; in dd_finish_request() local
570 spin_lock_irqsave(&dd->zone_lock, flags); in dd_finish_request()
574 spin_unlock_irqrestore(&dd->zone_lock, flags); in dd_finish_request()
Dpartition-generic.c310 sector_t start, sector_t len, int flags, in add_partition() argument
388 if (flags & ADDPART_FLAG_WHOLEDISK) { in add_partition()
396 if (flags & ADDPART_FLAG_WHOLEDISK) in add_partition()
431 !(disk->flags & GENHD_FL_NATIVE_CAPACITY)) { in disk_unlock_native_capacity()
434 disk->flags |= GENHD_FL_NATIVE_CAPACITY; in disk_unlock_native_capacity()
569 state->parts[p].flags, in rescan_partitions()
577 if (state->parts[p].flags & ADDPART_FLAG_RAID) in rescan_partitions()
Dblk-mq-tag.h58 if (!(hctx->flags & BLK_MQ_F_TAG_SHARED)) in blk_mq_tag_busy()
66 if (!(hctx->flags & BLK_MQ_F_TAG_SHARED)) in blk_mq_tag_idle()
Dbio.c297 unsigned long flags = bio->bi_flags & (~0UL << BIO_RESET_BITS); in bio_reset() local
302 bio->bi_flags = flags; in bio_reset()
531 unsigned long flags; in zero_fill_bio_iter() local
536 char *data = bvec_kmap_irq(&bv, &flags); in zero_fill_bio_iter()
539 bvec_kunmap_irq(data, &flags); in zero_fill_bio_iter()
1737 unsigned long flags; in bio_check_pages_dirty() local
1749 spin_lock_irqsave(&bio_dirty_lock, flags); in bio_check_pages_dirty()
1752 spin_unlock_irqrestore(&bio_dirty_lock, flags); in bio_check_pages_dirty()
2003 int flags) in bioset_init() argument
2020 if ((flags & BIOSET_NEED_BVECS) && in bioset_init()
[all …]
Dblk-mq-tag.c76 if (!hctx || !(hctx->flags & BLK_MQ_F_TAG_SHARED)) in hctx_may_queue()
101 if (!(data->flags & BLK_MQ_REQ_INTERNAL) && in __blk_mq_get_tag()
119 if (data->flags & BLK_MQ_REQ_RESERVED) { in blk_mq_get_tag()
135 if (data->flags & BLK_MQ_REQ_NOWAIT) in blk_mq_get_tag()
172 if (data->flags & BLK_MQ_REQ_RESERVED) in blk_mq_get_tag()
Dcmdline-parser.c60 new_subpart->flags = 0; in parse_subpart()
63 new_subpart->flags |= PF_RDONLY; in parse_subpart()
68 new_subpart->flags |= PF_POWERUP_LOCK; in parse_subpart()
Dblk-iolatency.c526 unsigned long flags; in iolatency_check_latencies() local
553 spin_lock_irqsave(&lat_info->lock, flags); in iolatency_check_latencies()
585 spin_unlock_irqrestore(&lat_info->lock, flags); in iolatency_check_latencies()
668 unsigned long flags; in blkiolatency_timer_fn() local
688 spin_lock_irqsave(&lat_info->lock, flags); in blkiolatency_timer_fn()
710 spin_unlock_irqrestore(&lat_info->lock, flags); in blkiolatency_timer_fn()
1025 .flags = CFTYPE_NOT_ON_ROOT,
Dbfq-cgroup.c106 stats->flags |= (1 << BFQG_stats_##name); \
110 stats->flags &= ~(1 << BFQG_stats_##name); \
114 return (stats->flags & (1 << BFQG_stats_##name)) != 0; \
844 unsigned long flags; in bfq_pd_offline() local
847 spin_lock_irqsave(&bfqd->lock, flags); in bfq_pd_offline()
886 spin_unlock_irqrestore(&bfqd->lock, flags); in bfq_pd_offline()
1215 .flags = CFTYPE_NOT_ON_ROOT,
1221 .flags = CFTYPE_NOT_ON_ROOT,
1341 .flags = CFTYPE_NOT_ON_ROOT,
Dblk-iocost.c198 unsigned long flags; \
200 spin_lock_irqsave(&trace_iocg_path_lock, flags); \
205 spin_unlock_irqrestore(&trace_iocg_path_lock, flags); \
1117 int flags, void *key) in iocg_wake_fn() argument
1138 default_wake_function(wq_entry, mode, flags, key); in iocg_wake_fn()
1204 unsigned long flags; in iocg_waitq_timer_fn() local
1208 spin_lock_irqsave(&iocg->waitq.lock, flags); in iocg_waitq_timer_fn()
1210 spin_unlock_irqrestore(&iocg->waitq.lock, flags); in iocg_waitq_timer_fn()
1997 unsigned long flags; in ioc_pd_init() local
2023 spin_lock_irqsave(&ioc->lock, flags); in ioc_pd_init()
[all …]
Dbio-integrity.c226 !(bi->flags & BLK_INTEGRITY_VERIFY)) in bio_integrity_prep()
230 !(bi->flags & BLK_INTEGRITY_GENERATE)) in bio_integrity_prep()
261 if (bi->flags & BLK_INTEGRITY_IP_CHECKSUM) in bio_integrity_prep()
Dblk-core.c404 int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) in blk_queue_enter() argument
406 const bool pm = flags & BLK_MQ_REQ_PREEMPT; in blk_queue_enter()
429 if (flags & BLK_MQ_REQ_NOWAIT) in blk_queue_enter()
582 blk_mq_req_flags_t flags) in blk_get_request() argument
587 WARN_ON_ONCE(flags & ~(BLK_MQ_REQ_NOWAIT | BLK_MQ_REQ_PREEMPT)); in blk_get_request()
589 req = blk_mq_alloc_request(q, op, flags); in blk_get_request()
1061 blk_mq_req_flags_t flags = bio->bi_opf & REQ_NOWAIT ? in generic_make_request() local
1064 if (likely(blk_queue_enter(q, flags) == 0)) { in generic_make_request()
Dcompat_ioctl.c179 compat_int_t flags; member
194 err |= get_user(n, &ua32->flags); in compat_blkpg_ioctl()
195 err |= put_user(n, &a->flags); in compat_blkpg_ioctl()
/block/partitions/
Dsun.c29 __be16 flags; in sun_partition() member
110 state->parts[slot].flags = 0; in sun_partition()
113 state->parts[slot].flags |= ADDPART_FLAG_RAID; in sun_partition()
115 state->parts[slot].flags |= ADDPART_FLAG_WHOLEDISK; in sun_partition()
Dcheck.h16 int flags; member

12