/block/ |
D | blk-mq-sched.c | 97 if (e->type->ops.has_work && !e->type->ops.has_work(hctx)) in blk_mq_do_dispatch_sched() 103 rq = e->type->ops.dispatch_request(hctx); in blk_mq_do_dispatch_sched() 121 unsigned short idx = ctx->index_hw[hctx->type]; in blk_mq_next_ctx() 174 const bool has_sched_dispatch = e && e->type->ops.dispatch_request; in blk_mq_sched_dispatch_requests() 312 enum hctx_type type = hctx->type; in blk_mq_attempt_merge() local 316 if (blk_mq_bio_list_merge(q, &ctx->rq_lists[type], bio, nr_segs)) { in blk_mq_attempt_merge() 331 enum hctx_type type; in __blk_mq_sched_bio_merge() local 333 if (e && e->type->ops.bio_merge) in __blk_mq_sched_bio_merge() 334 return e->type->ops.bio_merge(hctx, bio, nr_segs); in __blk_mq_sched_bio_merge() 336 type = hctx->type; in __blk_mq_sched_bio_merge() [all …]
|
D | blk-mq-sched.h | 50 if (e && e->type->ops.allow_merge) in blk_mq_sched_allow_merge() 51 return e->type->ops.allow_merge(q, rq, bio); in blk_mq_sched_allow_merge() 60 if (e && e->type->ops.completed_request) in blk_mq_sched_completed_request() 61 e->type->ops.completed_request(rq, now); in blk_mq_sched_completed_request() 69 if (e && e->type->ops.requeue_request) in blk_mq_sched_requeue_request() 70 e->type->ops.requeue_request(rq); in blk_mq_sched_requeue_request() 77 if (e && e->type->ops.has_work) in blk_mq_sched_has_work() 78 return e->type->ops.has_work(hctx); in blk_mq_sched_has_work()
|
D | elevator.c | 65 if (e->type->ops.allow_merge) in elv_iosched_allow_bio_merge() 66 return e->type->ops.allow_merge(q, rq, bio); in elv_iosched_allow_bio_merge() 173 eq->type = e; in elevator_alloc() 187 elevator_put(e->type); in elevator_release() 194 if (e->type->ops.exit_sched) in __elevator_exit() 343 if (e->type->ops.request_merge) in elv_merge() 344 return e->type->ops.request_merge(q, req, bio); in elv_merge() 391 enum elv_merge type) in elv_merged_request() argument 395 if (e->type->ops.request_merged) in elv_merged_request() 396 e->type->ops.request_merged(q, rq, type); in elv_merged_request() [all …]
|
D | blk-mq.h | 90 enum hctx_type type, in blk_mq_map_queue_type() argument 93 return q->queue_hw_ctx[q->tag_set->map[type].mq_map[cpu]]; in blk_mq_map_queue_type() 106 enum hctx_type type = HCTX_TYPE_DEFAULT; in blk_mq_map_queue() local 112 type = HCTX_TYPE_POLL; in blk_mq_map_queue() 114 type = HCTX_TYPE_READ; in blk_mq_map_queue() 116 return ctx->hctxs[type]; in blk_mq_map_queue()
|
D | t10-pi.c | 30 csum_fn *fn, enum t10_dif_type type) in t10_pi_generate() argument 40 if (type == T10_PI_TYPE1_PROTECTION) in t10_pi_generate() 54 csum_fn *fn, enum t10_dif_type type) in t10_pi_verify() argument 58 BUG_ON(type == T10_PI_TYPE0_PROTECTION); in t10_pi_verify() 64 if (type == T10_PI_TYPE1_PROTECTION || in t10_pi_verify() 65 type == T10_PI_TYPE2_PROTECTION) { in t10_pi_verify() 77 } else if (type == T10_PI_TYPE3_PROTECTION) { in t10_pi_verify()
|
D | blk-mq.c | 80 const int bit = ctx->index_hw[hctx->type]; in blk_mq_hctx_mark_pending() 89 const int bit = ctx->index_hw[hctx->type]; in blk_mq_hctx_clear_pending() 392 e->type->ops.limit_depth && in blk_mq_get_request() 394 e->type->ops.limit_depth(data->cmd_flags, data); in blk_mq_get_request() 410 if (e && e->type->ops.prepare_request) { in blk_mq_get_request() 411 if (e->type->icq_cache) in blk_mq_get_request() 414 e->type->ops.prepare_request(rq, bio); in blk_mq_get_request() 517 if (e && e->type->ops.finish_request) in blk_mq_free_request() 518 e->type->ops.finish_request(rq); in blk_mq_free_request() 984 enum hctx_type type = hctx->type; in flush_busy_ctx() local [all …]
|
D | kyber-iosched.c | 211 unsigned int sched_domain, unsigned int type) in flush_latency_buckets() argument 213 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in flush_latency_buckets() 214 atomic_t *cpu_buckets = cpu_latency->buckets[sched_domain][type]; in flush_latency_buckets() 226 unsigned int sched_domain, unsigned int type, in calculate_percentile() argument 229 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in calculate_percentile() 256 memset(buckets, 0, sizeof(kqd->latency_buckets[sched_domain][type])); in calculate_percentile() 259 kyber_latency_type_names[type], percentile, in calculate_percentile() 570 struct kyber_ctx_queue *kcq = &khd->kcqs[ctx->index_hw[hctx->type]]; in kyber_bio_merge() 595 struct kyber_ctx_queue *kcq = &khd->kcqs[rq->mq_ctx->index_hw[hctx->type]]; in kyber_insert_requests() 604 rq->mq_ctx->index_hw[hctx->type]); in kyber_insert_requests() [all …]
|
D | sed-opal.c | 55 enum opal_response_token type; member 762 tok->type = OPAL_DTA_TOKENID_SINT; in response_parse_tiny() 764 tok->type = OPAL_DTA_TOKENID_UINT; in response_parse_tiny() 779 tok->type = OPAL_DTA_TOKENID_BYTESTRING; in response_parse_short() 781 tok->type = OPAL_DTA_TOKENID_SINT; in response_parse_short() 786 tok->type = OPAL_DTA_TOKENID_UINT; in response_parse_short() 809 tok->type = OPAL_DTA_TOKENID_BYTESTRING; in response_parse_medium() 811 tok->type = OPAL_DTA_TOKENID_SINT; in response_parse_medium() 813 tok->type = OPAL_DTA_TOKENID_UINT; in response_parse_medium() 826 tok->type = OPAL_DTA_TOKENID_BYTESTRING; in response_parse_long() [all …]
|
D | blk-ioc.c | 45 struct elevator_type *et = icq->q->elevator->type; in ioc_exit_icq() 64 struct elevator_type *et = q->elevator->type; in ioc_destroy_icq() 368 struct elevator_type *et = q->elevator->type; in ioc_create_icq()
|
D | blk-mq-debugfs.c | 431 seq_printf(m, "%s\n", hctx_types[hctx->type]); in hctx_type_show() 625 #define CTX_RQ_SEQ_OPS(name, type) \ argument 632 return seq_list_start(&ctx->rq_lists[type], *pos); \ 640 return seq_list_next(v, &ctx->rq_lists[type], pos); \ 919 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched() 975 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched_hctx()
|
D | blk-zoned.c | 396 switch (zone->type) { in blk_revalidate_zone_cb() 417 disk->disk_name, (int)zone->type, zone->start); in blk_revalidate_zone_cb()
|
D | ioctl.c | 348 return ops->pr_reserve(bdev, rsv.key, rsv.type, rsv.flags); in blkdev_pr_reserve() 366 return ops->pr_release(bdev, rsv.key, rsv.type); in blkdev_pr_release() 384 return ops->pr_preempt(bdev, p.old_key, p.new_key, p.type, abort); in blkdev_pr_preempt()
|
D | mq-deadline.c | 127 enum elv_merge type) in dd_request_merged() argument 134 if (type == ELEVATOR_FRONT_MERGE) { in dd_request_merged()
|
D | blk-iocost.c | 196 #define TRACE_IOCG_PATH(type, iocg, ...) \ argument 199 if (trace_iocost_##type##_enabled()) { \ 203 trace_iocost_##type(iocg, trace_iocg_path, \ 210 #define TRACE_IOCG_PATH(type, iocg, ...) do { } while (0) argument
|
D | partition-generic.c | 368 pdev->type = &part_type; in add_partition()
|
D | genhd.c | 1511 disk_to_dev(disk)->type = &disk_type; in __alloc_disk_node()
|
D | bfq-iosched.c | 2260 enum elv_merge type) in bfq_request_merged() argument 2262 if (type == ELEVATOR_FRONT_MERGE && in bfq_request_merged()
|
/block/partitions/ |
D | ibm.c | 70 char type[], in find_label() argument 114 strncpy(type, label->vol.vollbl, 4); in find_label() 117 strncpy(type, label->lnx.vollbl, 4); in find_label() 120 EBCASC(type, 4); in find_label() 229 if (!strcmp(info->type, "ECKD")) in find_lnx1_partitions() 297 char type[5] = {0,}; in ibm_partition() local 325 if (find_label(state, info, geo, blocksize, &labelsect, name, type, in ibm_partition() 327 if (!strncmp(type, "VOL1", 4)) { in ibm_partition() 330 } else if (!strncmp(type, "LNX1", 4)) { in ibm_partition() 334 } else if (!strncmp(type, "CMS1", 4)) { in ibm_partition()
|
D | mac.c | 89 if (!strncasecmp(part->type, "Linux_RAID", 10)) in mac_partition() 101 mac_fix_string(part->type, 32); in mac_partition() 107 if (strcasecmp(part->type, "Apple_UNIX_SVR2") == 0 in mac_partition() 108 || (strncasecmp(part->type, "Linux", 5) == 0 in mac_partition() 109 && strcasecmp(part->type, "Linux_swap") != 0)) { in mac_partition()
|
D | sgi.c | 25 __be32 type; /* Type of this partition */ member 75 if (be32_to_cpu(p->type) == LINUX_RAID_PARTITION) in sgi_partition()
|
D | ldm.h | 131 u8 type; member 174 u8 type; member
|
D | mac.h | 18 char type[32]; /* string type description */ member
|
D | ldm.c | 752 comp->type = buffer[0x18 + r_vstate]; in ldm_parse_cmp3() 1127 vb->type = buf[0x13]; in ldm_parse_vblk() 1131 switch (vb->type) { in ldm_parse_vblk() 1143 (unsigned long long) vb->obj_id, vb->type); in ldm_parse_vblk() 1146 (unsigned long long) vb->obj_id, vb->type); in ldm_parse_vblk() 1184 switch (vb->type) { in ldm_ldmdb_add()
|
D | acorn.c | 416 char type[8]; member
|