/block/partitions/ |
D | check.h | 31 void *read_part_sector(struct parsed_partitions *state, sector_t n, Sector *p); 38 put_partition(struct parsed_partitions *p, int n, sector_t from, sector_t size) in put_partition() argument 40 if (n < p->limit) { in put_partition() 43 p->parts[n].from = from; in put_partition() 44 p->parts[n].size = size; in put_partition() 45 snprintf(tmp, sizeof(tmp), " %s%d", p->name, n); in put_partition()
|
D | aix.c | 186 struct lvname *n = NULL; in aix_partition() local 225 n = alloc_lvn(state, vgda_sector + vgda_len - 33); in aix_partition() 226 if (n) { in aix_partition() 276 n[lv_ix].name); in aix_partition() 286 char tmp[sizeof(n[i].name) + 1]; // null char in aix_partition() 288 snprintf(tmp, sizeof(tmp), "%s", n[i].name); in aix_partition() 295 kfree(n); in aix_partition()
|
D | msdos.c | 674 sector_t n = 2; in msdos_partition() local 676 n = min(size, max(sector_size, n)); in msdos_partition() 677 put_partition(state, slot, start, n); in msdos_partition() 700 int n; in msdos_partition() local 705 for (n = 0; subtypes[n].parse && id != subtypes[n].id; n++) in msdos_partition() 708 if (!subtypes[n].parse) in msdos_partition() 710 subtypes[n].parse(state, start_sect(p) * sector_size, in msdos_partition()
|
D | core.c | 777 void *read_part_sector(struct parsed_partitions *state, sector_t n, Sector *p) in read_part_sector() argument 782 if (n >= get_capacity(state->bdev->bd_disk)) { in read_part_sector() 788 (pgoff_t)(n >> (PAGE_SHIFT - 9)), NULL); in read_part_sector() 796 ((n & ((1 << (PAGE_SHIFT - 9)) - 1)) << SECTOR_SHIFT); in read_part_sector()
|
D | efi.c | 242 sector_t n = lba * (bdev_logical_block_size(bdev) / 512); in read_lba() local 250 unsigned char *data = read_part_sector(state, n++, §); in read_lba()
|
/block/ |
D | genhd.c | 442 struct blk_major_name **n, *p; in register_blkdev() local 483 for (n = &major_names[index]; *n; n = &(*n)->next) { in register_blkdev() 484 if ((*n)->major == major) in register_blkdev() 487 if (!*n) in register_blkdev() 488 *n = p; in register_blkdev() 506 struct blk_major_name **n; in unregister_blkdev() local 511 for (n = &major_names[index]; *n; n = &(*n)->next) in unregister_blkdev() 512 if ((*n)->major == major) in unregister_blkdev() 514 if (!*n || strcmp((*n)->name, name)) { in unregister_blkdev() 517 p = *n; in unregister_blkdev() [all …]
|
D | elevator.c | 285 struct rb_node *n = root->rb_node; in elv_rb_find() local 288 while (n) { in elv_rb_find() 289 rq = rb_entry(n, struct request, rb_node); in elv_rb_find() 292 n = n->rb_left; in elv_rb_find() 294 n = n->rb_right; in elv_rb_find()
|
D | ioctl.c | 398 int ret, n; in blkdev_roset() local 406 if (get_user(n, (int __user *)arg)) in blkdev_roset() 408 set_device_ro(bdev, n); in blkdev_roset() 481 int ret, n; in blkdev_bszset() local 487 if (get_user(n, argp)) in blkdev_bszset() 491 return set_blocksize(bdev, n); in blkdev_bszset() 495 ret = set_blocksize(bdev, n); in blkdev_bszset()
|
D | blk-map.c | 279 unsigned int n = PAGE_SIZE - offs; in bio_map_user_iov() local 282 if (n > bytes) in bio_map_user_iov() 283 n = bytes; in bio_map_user_iov() 285 if (!bio_add_hw_page(rq->q, bio, page, n, offs, in bio_map_user_iov() 292 added += n; in bio_map_user_iov() 293 bytes -= n; in bio_map_user_iov()
|
D | blk-flush.c | 219 struct request *rq, *n; in flush_end_io() local 258 list_for_each_entry_safe(rq, n, running, flush.list) { in flush_end_io()
|
D | bfq-iosched.c | 569 struct rb_node **n; in bfq_rq_pos_tree_lookup() local 579 n = &(*p)->rb_right; in bfq_rq_pos_tree_lookup() 581 n = &(*p)->rb_left; in bfq_rq_pos_tree_lookup() 584 p = n; in bfq_rq_pos_tree_lookup() 1079 struct hlist_node *n; in bfq_reset_burst_list() local 1081 hlist_for_each_entry_safe(item, n, &bfqd->burst_list, burst_list_node) in bfq_reset_burst_list() 1106 struct hlist_node *n; in bfq_add_to_burst() local 1130 hlist_for_each_entry_safe(pos, n, &bfqd->burst_list, in bfq_add_to_burst() 4861 struct hlist_node *n; in bfq_put_queue() local 4924 hlist_for_each_entry_safe(item, n, &bfqq->woken_list, in bfq_put_queue() [all …]
|
D | blk-throttle.c | 643 struct rb_node *n; in throtl_rb_first() local 645 n = rb_first_cached(&parent_sq->pending_tree); in throtl_rb_first() 646 WARN_ON_ONCE(!n); in throtl_rb_first() 647 if (!n) in throtl_rb_first() 649 return rb_entry_tg(n); in throtl_rb_first() 652 static void throtl_rb_erase(struct rb_node *n, in throtl_rb_erase() argument 655 rb_erase_cached(n, &parent_sq->pending_tree); in throtl_rb_erase() 656 RB_CLEAR_NODE(n); in throtl_rb_erase()
|
D | sed-opal.c | 732 int n) in response_get_token() argument 741 if (n >= resp->num) { in response_get_token() 743 n, resp->num); in response_get_token() 747 tok = &resp->toks[n]; in response_get_token() 915 static size_t response_get_string(const struct parsed_resp *resp, int n, in response_get_string() argument 922 tok = response_get_token(resp, n); in response_get_string() 952 static u64 response_get_u64(const struct parsed_resp *resp, int n) in response_get_u64() argument 956 tok = response_get_token(resp, n); in response_get_u64()
|
D | bio.c | 39 #define BV(x, n) { .nr_vecs = x, .name = "biovec-"#n } argument
|
D | blk-cgroup.c | 426 struct blkcg_gq *blkg, *n; in blkg_destroy_all() local 429 list_for_each_entry_safe(blkg, n, &q->blkg_list, q_node) { in blkg_destroy_all()
|
D | blk-sysfs.c | 666 int n) in queue_attr_visible() argument
|