/block/partitions/ |
D | acorn.c | 55 __le32 start; member 95 le32_to_cpu(rr->part[part].start), in riscix_partition() 304 __le32 start; member 372 u32 start = le32_to_cpu(p->start); in adfspart_check_ICS() local 392 if (size > 1 && adfspart_check_ICSLinux(state, start)) { in adfspart_check_ICS() 393 start += 1; in adfspart_check_ICS() 399 put_partition(state, slot++, start, size); in adfspart_check_ICS() 412 __le32 start; member 466 u32 start = le32_to_cpu(p->start); in adfspart_check_POWERTEC() local 470 put_partition(state, slot++, start, size); in adfspart_check_POWERTEC() [all …]
|
D | core.c | 219 static DEVICE_ATTR(start, 0444, part_start_show, NULL); 371 sector_t start, sector_t len, int flags, in add_partition() argument 421 p->start_sect = start; in add_partition() 504 static bool partition_overlaps(struct gendisk *disk, sector_t start, in partition_overlaps() argument 514 start >= part->start_sect + part->nr_sects || in partition_overlaps() 515 start + length <= part->start_sect) in partition_overlaps() 526 sector_t start, sector_t length) in bdev_add_partition() argument 531 if (partition_overlaps(bdev->bd_disk, start, length, -1)) { in bdev_add_partition() 536 part = add_partition(bdev->bd_disk, partno, start, length, in bdev_add_partition() 579 sector_t start, sector_t length) in bdev_resize_partition() argument [all …]
|
D | sgi.c | 41 unsigned int start, blocks; in sgi_partition() local 77 start = be32_to_cpu(p->first_block); in sgi_partition() 79 put_partition(state, slot, start, blocks); in sgi_partition()
|
D | msdos.c | 662 sector_t start = start_sect(p)*sector_size; in msdos_partition() local 677 put_partition(state, slot, start, n); in msdos_partition() 680 parse_extended(state, start, size, disksig); in msdos_partition() 684 put_partition(state, slot, start, size); in msdos_partition()
|
D | ldm.h | 143 u64 start; member
|
D | efi.c | 721 u64 start = le64_to_cpu(ptes[i].starting_lba); in efi_partition() local 728 put_partition(state, i+1, start * ssz, size * ssz); in efi_partition()
|
D | ldm.c | 598 part->start, part->size); in ldm_create_data_partitions() 981 part->start = get_unaligned_be64(buffer + 0x24 + r_name); in ldm_parse_prt3() 1204 (v->vblk.part.start > vb->vblk.part.start)) { in ldm_ldmdb_add()
|
D | ibm.c | 324 geo->start = get_start_sect(bdev); in ibm_partition()
|
/block/ |
D | ioctl.c | 20 long long start, length; in blkpg_do_ioctl() local 35 start = p.start >> SECTOR_SHIFT; in blkpg_do_ioctl() 40 long pstart = start, plength = length; in blkpg_do_ioctl() 42 if (pstart != start || plength != length || pstart < 0 || in blkpg_do_ioctl() 50 if (p.start & (bdev_logical_block_size(bdev) - 1)) in blkpg_do_ioctl() 52 return bdev_add_partition(bdev, p.pno, start, length); in blkpg_do_ioctl() 54 return bdev_resize_partition(bdev, p.pno, start, length); in blkpg_do_ioctl() 122 uint64_t start, len; in blk_ioctl_discard() local 135 start = range[0]; in blk_ioctl_discard() 138 if (start & 511) in blk_ioctl_discard() [all …]
|
D | blk-zoned.c | 321 loff_t start, end; in blkdev_truncate_zone_range() local 328 start = zrange->sector << SECTOR_SHIFT; in blkdev_truncate_zone_range() 331 return truncate_bdev_range(bdev, mode, start, end); in blkdev_truncate_zone_range() 440 if (zone->start == 0) { in blk_revalidate_zone_cb() 449 } else if (zone->start + args->zone_sectors < capacity) { in blk_revalidate_zone_cb() 464 if (zone->start != args->sector) { in blk_revalidate_zone_cb() 466 disk->disk_name, args->sector, zone->start); in blk_revalidate_zone_cb() 492 disk->disk_name, (int)zone->type, zone->start); in blk_revalidate_zone_cb()
|
D | bfq-wf2q.c | 117 !bfq_gt(new_entity->start, st->vtime) in bfq_update_next_in_service() 292 entity->finish = entity->start + in bfq_calc_finish() 301 entity->start, entity->finish, in bfq_calc_finish() 426 entity->min_start = entity->start; in bfq_update_active_node() 808 entity->start = new_st->vtime; in __bfq_entity_update_weight_prio() 945 entity->start += delta; in bfq_update_fin_time_enqueue() 984 entity->start = bfq_gt(min_vstart, entity->finish) ? in __bfq_activate_entity() 992 entity->start = min_vstart; in __bfq_activate_entity() 1065 entity->start = entity->finish; in __bfq_requeue_entity() 1359 if (!bfq_gt(entry->start, vtime)) in bfq_first_active_entity()
|
D | badblocks.c | 401 sector_t start = BB_OFFSET(p[lo]); in badblocks_clear() local 403 p[lo] = BB_MAKE(start, s - start, ack); in badblocks_clear() 446 sector_t start = BB_OFFSET(p[i]); in ack_all_badblocks() local 449 p[i] = BB_MAKE(start, len, 1); in ack_all_badblocks()
|
D | blk-map.c | 386 unsigned long start = kaddr >> PAGE_SHIFT; in bio_map_kern() local 387 const int nr_pages = end - start; in bio_map_kern() 468 unsigned long start = kaddr >> PAGE_SHIFT; in bio_copy_kern() local 476 if (end < start) in bio_copy_kern() 479 nr_pages = end - start; in bio_copy_kern()
|
D | bio-integrity.c | 214 unsigned long start, end; in bio_integrity_prep() local 254 start = ((unsigned long) buf) >> PAGE_SHIFT; in bio_integrity_prep() 255 nr_pages = end - start; in bio_integrity_prep()
|
D | blk-settings.c | 501 sector_t start) in blk_stack_limits() argument 532 alignment = queue_limit_alignment_offset(b, start); in blk_stack_limits() 611 alignment = queue_limit_discard_alignment(b, start); in blk_stack_limits()
|
D | blk.h | 366 sector_t start, sector_t length); 369 sector_t start, sector_t length);
|
D | blk-iolatency.c | 494 u64 start = bio_issue_time(issue); in iolatency_record_time() local 503 if (now <= start) in iolatency_record_time() 506 req_time = now - start; in iolatency_record_time()
|
D | blk-mq-debugfs.c | 69 .start = queue_requeue_list_start, 389 .start = hctx_dispatch_start, 658 .start = ctx_##name##_rq_list_start, \
|
D | blk-mq.h | 51 struct blk_mq_ctx *start);
|
D | blk-mq.c | 1083 struct blk_mq_ctx *start) in blk_mq_dequeue_from_ctx() argument 1085 unsigned off = start ? start->index_hw[hctx->type] : 0; in blk_mq_dequeue_from_ctx() 2393 unsigned long start = (unsigned long)page_address(page); in blk_mq_clear_rq_mapping() local 2394 unsigned long end = start + order_to_size(page->private); in blk_mq_clear_rq_mapping() 2401 if (rq_addr >= start && rq_addr < end) { in blk_mq_clear_rq_mapping()
|
D | bfq-iosched.h | 156 u64 start, finish; member
|
D | mq-deadline-main.c | 967 .start = deadline_##name##_fifo_start, \ 1083 .start = deadline_dispatch##prio##_start, \
|
D | sed-opal.c | 635 u8 *start; in add_token_bytestring() local 637 start = add_bytestring_header(err, cmd, len); in add_token_bytestring() 638 if (!start) in add_token_bytestring() 640 memcpy(start, bytestring, len); in add_token_bytestring()
|
D | bio.c | 534 void zero_fill_bio_iter(struct bio *bio, struct bvec_iter start) in zero_fill_bio_iter() argument 540 __bio_for_each_segment(bv, bio, iter, start) { in zero_fill_bio_iter()
|
D | blk-throttle.c | 772 bool rw, unsigned long start) in throtl_start_new_slice_with_credit() argument 785 if (time_after_eq(start, tg->slice_start[rw])) in throtl_start_new_slice_with_credit() 786 tg->slice_start[rw] = start; in throtl_start_new_slice_with_credit()
|