/block/partitions/ |
D | check.h | 32 sector_t n, Sector *p) in read_part_sector() argument 34 if (n >= get_capacity(state->bdev->bd_disk)) { in read_part_sector() 38 return read_dev_sector(state->bdev, n, p); in read_part_sector() 42 put_partition(struct parsed_partitions *p, int n, sector_t from, sector_t size) in put_partition() argument 44 if (n < p->limit) { in put_partition() 47 p->parts[n].from = from; in put_partition() 48 p->parts[n].size = size; in put_partition() 49 snprintf(tmp, sizeof(tmp), " %s%d", p->name, n); in put_partition()
|
D | aix.c | 187 struct lvname *n = NULL; in aix_partition() local 226 n = alloc_lvn(state, vgda_sector + vgda_len - 33); in aix_partition() 227 if (n) { in aix_partition() 277 n[lv_ix].name); in aix_partition() 287 char tmp[sizeof(n[i].name) + 1]; // null char in aix_partition() 289 snprintf(tmp, sizeof(tmp), "%s", n[i].name); in aix_partition() 296 kfree(n); in aix_partition()
|
D | msdos.c | 546 sector_t n = 2; in msdos_partition() local 548 n = min(size, max(sector_size, n)); in msdos_partition() 549 put_partition(state, slot, start, n); in msdos_partition() 572 int n; in msdos_partition() local 577 for (n = 0; subtypes[n].parse && id != subtypes[n].id; n++) in msdos_partition() 580 if (!subtypes[n].parse) in msdos_partition() 582 subtypes[n].parse(state, start_sect(p) * sector_size, in msdos_partition()
|
D | efi.c | 242 sector_t n = lba * (bdev_logical_block_size(bdev) / 512); in read_lba() local 250 unsigned char *data = read_part_sector(state, n++, §); in read_lba()
|
/block/ |
D | compat_ioctl.c | 189 compat_int_t n; in compat_blkpg_ioctl() local 192 err = get_user(n, &ua32->op); in compat_blkpg_ioctl() 193 err |= put_user(n, &a->op); in compat_blkpg_ioctl() 194 err |= get_user(n, &ua32->flags); in compat_blkpg_ioctl() 195 err |= put_user(n, &a->flags); in compat_blkpg_ioctl() 196 err |= get_user(n, &ua32->datalen); in compat_blkpg_ioctl() 197 err |= put_user(n, &a->datalen); in compat_blkpg_ioctl()
|
D | genhd.c | 356 struct blk_major_name **n, *p; in register_blkdev() local 397 for (n = &major_names[index]; *n; n = &(*n)->next) { in register_blkdev() 398 if ((*n)->major == major) in register_blkdev() 401 if (!*n) in register_blkdev() 402 *n = p; in register_blkdev() 420 struct blk_major_name **n; in unregister_blkdev() local 425 for (n = &major_names[index]; *n; n = &(*n)->next) in unregister_blkdev() 426 if ((*n)->major == major) in unregister_blkdev() 428 if (!*n || strcmp((*n)->name, name)) { in unregister_blkdev() 431 p = *n; in unregister_blkdev() [all …]
|
D | elevator.c | 286 struct rb_node *n = root->rb_node; in elv_rb_find() local 289 while (n) { in elv_rb_find() 290 rq = rb_entry(n, struct request, rb_node); in elv_rb_find() 293 n = n->rb_left; in elv_rb_find() 295 n = n->rb_right; in elv_rb_find()
|
D | ioctl.c | 445 int ret, n; in blkdev_roset() local 453 if (get_user(n, (int __user *)arg)) in blkdev_roset() 455 set_device_ro(bdev, n); in blkdev_roset() 489 int ret, n; in blkdev_bszset() local 495 if (get_user(n, argp)) in blkdev_bszset() 504 ret = set_blocksize(bdev, n); in blkdev_bszset()
|
D | bio.c | 37 #define BV(x, n) { .nr_vecs = x, .name = "biovec-"#n } argument 1426 unsigned int n = PAGE_SIZE - offs; in bio_map_user_iov() local 1429 if (n > bytes) in bio_map_user_iov() 1430 n = bytes; in bio_map_user_iov() 1432 if (!__bio_add_pc_page(q, bio, page, n, offs, in bio_map_user_iov() 1439 added += n; in bio_map_user_iov() 1440 bytes -= n; in bio_map_user_iov()
|
D | blk-flush.c | 211 struct request *rq, *n; in flush_end_io() local 244 list_for_each_entry_safe(rq, n, running, flush.list) { in flush_end_io()
|
D | partition-generic.c | 606 unsigned char *read_dev_sector(struct block_device *bdev, sector_t n, Sector *p) in read_dev_sector() argument 611 page = read_mapping_page(mapping, (pgoff_t)(n >> (PAGE_SHIFT-9)), NULL); in read_dev_sector() 616 return (unsigned char *)page_address(page) + ((n & ((1 << (PAGE_SHIFT - 9)) - 1)) << 9); in read_dev_sector()
|
D | blk-throttle.c | 620 struct rb_node *n; in throtl_rb_first() local 625 n = rb_first_cached(&parent_sq->pending_tree); in throtl_rb_first() 626 WARN_ON_ONCE(!n); in throtl_rb_first() 627 if (!n) in throtl_rb_first() 629 return rb_entry_tg(n); in throtl_rb_first() 632 static void throtl_rb_erase(struct rb_node *n, in throtl_rb_erase() argument 635 rb_erase_cached(n, &parent_sq->pending_tree); in throtl_rb_erase() 636 RB_CLEAR_NODE(n); in throtl_rb_erase()
|
D | sed-opal.c | 730 int n) in response_get_token() argument 739 if (n >= resp->num) { in response_get_token() 741 n, resp->num); in response_get_token() 745 tok = &resp->toks[n]; in response_get_token() 913 static size_t response_get_string(const struct parsed_resp *resp, int n, in response_get_string() argument 920 tok = response_get_token(resp, n); in response_get_string() 950 static u64 response_get_u64(const struct parsed_resp *resp, int n) in response_get_u64() argument 954 tok = response_get_token(resp, n); in response_get_u64()
|
D | bfq-iosched.c | 561 struct rb_node **n; in bfq_rq_pos_tree_lookup() local 571 n = &(*p)->rb_right; in bfq_rq_pos_tree_lookup() 573 n = &(*p)->rb_left; in bfq_rq_pos_tree_lookup() 576 p = n; in bfq_rq_pos_tree_lookup() 1067 struct hlist_node *n; in bfq_reset_burst_list() local 1069 hlist_for_each_entry_safe(item, n, &bfqd->burst_list, burst_list_node) in bfq_reset_burst_list() 1094 struct hlist_node *n; in bfq_add_to_burst() local 1118 hlist_for_each_entry_safe(pos, n, &bfqd->burst_list, in bfq_add_to_burst() 4812 struct hlist_node *n; in bfq_put_queue() local 4877 hlist_for_each_entry_safe(item, n, &bfqq->woken_list, in bfq_put_queue() [all …]
|
D | blk-cgroup.c | 447 struct blkcg_gq *blkg, *n; in blkg_destroy_all() local 450 list_for_each_entry_safe(blkg, n, &q->blkg_list, q_node) { in blkg_destroy_all()
|
D | blk-sysfs.c | 773 int n) in queue_attr_visible() argument
|