/block/ |
D | blk-mq-sysfs.c | 36 struct blk_mq_ctx *ctx; in blk_mq_sysfs_show() local 41 ctx = container_of(kobj, struct blk_mq_ctx, kobj); in blk_mq_sysfs_show() 42 q = ctx->queue; in blk_mq_sysfs_show() 50 res = entry->show(ctx, page); in blk_mq_sysfs_show() 59 struct blk_mq_ctx *ctx; in blk_mq_sysfs_store() local 64 ctx = container_of(kobj, struct blk_mq_ctx, kobj); in blk_mq_sysfs_store() 65 q = ctx->queue; in blk_mq_sysfs_store() 73 res = entry->store(ctx, page, length); in blk_mq_sysfs_store() 125 static ssize_t blk_mq_sysfs_dispatched_show(struct blk_mq_ctx *ctx, char *page) in blk_mq_sysfs_dispatched_show() argument 127 return sprintf(page, "%lu %lu\n", ctx->rq_dispatched[1], in blk_mq_sysfs_dispatched_show() [all …]
|
D | blk-mq.c | 53 struct blk_mq_ctx *ctx) in get_bm() argument 55 return &hctx->ctx_map.map[ctx->index_hw / hctx->ctx_map.bits_per_word]; in get_bm() 58 #define CTX_TO_BIT(hctx, ctx) \ argument 59 ((ctx)->index_hw & ((hctx)->ctx_map.bits_per_word - 1)) 65 struct blk_mq_ctx *ctx) in blk_mq_hctx_mark_pending() argument 67 struct blk_align_bitmap *bm = get_bm(hctx, ctx); in blk_mq_hctx_mark_pending() 69 if (!test_bit(CTX_TO_BIT(hctx, ctx), &bm->word)) in blk_mq_hctx_mark_pending() 70 set_bit(CTX_TO_BIT(hctx, ctx), &bm->word); in blk_mq_hctx_mark_pending() 74 struct blk_mq_ctx *ctx) in blk_mq_hctx_clear_pending() argument 76 struct blk_align_bitmap *bm = get_bm(hctx, ctx); in blk_mq_hctx_clear_pending() [all …]
|
D | blk-mq.h | 91 static inline void blk_mq_put_ctx(struct blk_mq_ctx *ctx) in blk_mq_put_ctx() argument 103 struct blk_mq_ctx *ctx; member 109 struct blk_mq_ctx *ctx, in blk_mq_set_alloc_data() argument 115 data->ctx = ctx; in blk_mq_set_alloc_data()
|
D | blk-throttle.c | 1200 struct blkg_conf_ctx ctx; in tg_set_conf() local 1205 ret = blkg_conf_prep(blkcg, &blkcg_policy_throtl, buf, &ctx); in tg_set_conf() 1210 if (sscanf(ctx.body, "%llu", &v) != 1) in tg_set_conf() 1215 tg = blkg_to_tg(ctx.blkg); in tg_set_conf() 1225 blkg_conf_finish(&ctx); in tg_set_conf() 1317 struct blkg_conf_ctx ctx; in tg_set_max() local 1322 ret = blkg_conf_prep(blkcg, &blkcg_policy_throtl, buf, &ctx); in tg_set_max() 1326 tg = blkg_to_tg(ctx.blkg); in tg_set_max() 1339 if (sscanf(ctx.body, "%26s%n", tok, &len) != 1) in tg_set_max() 1343 ctx.body += len; in tg_set_max() [all …]
|
D | blk-cgroup.c | 787 char *input, struct blkg_conf_ctx *ctx) in blkg_conf_prep() argument 843 ctx->disk = disk; in blkg_conf_prep() 844 ctx->blkg = blkg; in blkg_conf_prep() 845 ctx->body = body; in blkg_conf_prep() 857 void blkg_conf_finish(struct blkg_conf_ctx *ctx) in blkg_conf_finish() argument 858 __releases(ctx->disk->queue->queue_lock) __releases(rcu) in blkg_conf_finish() 862 spin_unlock_irq(ctx->disk->queue->queue_lock); in blkg_conf_finish() 864 owner = ctx->disk->fops->owner; in blkg_conf_finish() 865 put_disk(ctx->disk); in blkg_conf_finish()
|
D | blk-mq-tag.c | 299 blk_mq_put_ctx(data->ctx); in bt_get() 303 data->ctx = blk_mq_get_ctx(data->q); in bt_get() 305 data->ctx->cpu); in bt_get() 309 last_tag = &data->ctx->last_tag; in bt_get() 326 &data->ctx->last_tag, data->hctx->tags); in __blk_mq_get_tag()
|
D | blk-flush.c | 356 struct blk_mq_ctx *ctx = rq->mq_ctx; in mq_flush_data_end_io() local 358 struct blk_flush_queue *fq = blk_get_flush_queue(q, ctx); in mq_flush_data_end_io() 360 hctx = q->mq_ops->map_queue(q, ctx->cpu); in mq_flush_data_end_io()
|
D | blk.h | 40 struct request_queue *q, struct blk_mq_ctx *ctx) in blk_get_flush_queue() argument 47 hctx = q->mq_ops->map_queue(q, ctx->cpu); in blk_get_flush_queue()
|
D | cfq-iosched.c | 1764 struct blkg_conf_ctx ctx; in __cfqg_set_weight_device() local 1770 ret = blkg_conf_prep(blkcg, &blkcg_policy_cfq, buf, &ctx); in __cfqg_set_weight_device() 1774 if (sscanf(ctx.body, "%llu", &v) == 1) { in __cfqg_set_weight_device() 1779 } else if (!strcmp(strim(ctx.body), "default")) { in __cfqg_set_weight_device() 1786 cfqg = blkg_to_cfqg(ctx.blkg); in __cfqg_set_weight_device() 1801 blkg_conf_finish(&ctx); in __cfqg_set_weight_device()
|