Home
last modified time | relevance | path

Searched refs:nr_sectors (Results 1 – 19 of 19) sorted by relevance

/drivers/md/bcache/
Dwriteback.h80 unsigned int nr_sectors) in bcache_dev_stripe_dirty() argument
91 if (nr_sectors <= dc->disk.stripe_size) in bcache_dev_stripe_dirty()
94 nr_sectors -= dc->disk.stripe_size; in bcache_dev_stripe_dirty()
149 uint64_t offset, int nr_sectors);
Dwriteback.c589 uint64_t offset, int nr_sectors) in bcache_dev_sectors_dirty_add() argument
603 atomic_long_add(nr_sectors, &c->flash_dev_dirty_sectors); in bcache_dev_sectors_dirty_add()
607 while (nr_sectors) { in bcache_dev_sectors_dirty_add()
608 int s = min_t(unsigned int, abs(nr_sectors), in bcache_dev_sectors_dirty_add()
611 if (nr_sectors < 0) in bcache_dev_sectors_dirty_add()
624 nr_sectors -= s; in bcache_dev_sectors_dirty_add()
/drivers/block/null_blk/
Dzoned.c238 unsigned int nr_sectors = len >> SECTOR_SHIFT; in null_zone_valid_read_len() local
242 sector + nr_sectors <= zone->wp) in null_zone_valid_read_len()
367 unsigned int nr_sectors, bool append) in null_zone_write() argument
379 return null_process_cmd(cmd, REQ_OP_WRITE, sector, nr_sectors); in null_zone_write()
407 if (zone->wp + nr_sectors > zone->start + zone->capacity) { in null_zone_write()
434 ret = null_process_cmd(cmd, REQ_OP_WRITE, sector, nr_sectors); in null_zone_write()
438 zone->wp += nr_sectors; in null_zone_write()
654 sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd() argument
662 return null_zone_write(cmd, sector, nr_sectors, false); in null_process_zoned_cmd()
664 return null_zone_write(cmd, sector, nr_sectors, true); in null_process_zoned_cmd()
[all …]
Dnull_blk.h123 sector_t nr_sectors);
126 unsigned int nr_sectors);
136 sector_t nr_sectors);
152 enum req_opf op, sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd() argument
Dmain.c1090 sector_t sector, sector_t nr_sectors) in null_handle_discard() argument
1093 size_t n = nr_sectors << SECTOR_SHIFT; in null_handle_discard()
1253 sector_t nr_sectors) in null_handle_badblocks() argument
1259 if (badblocks_check(bb, sector, nr_sectors, &first_bad, &bad_sectors)) in null_handle_badblocks()
1268 sector_t nr_sectors) in null_handle_memory_backed() argument
1274 return null_handle_discard(dev, sector, nr_sectors); in null_handle_memory_backed()
1338 unsigned int nr_sectors) in null_process_cmd() argument
1344 ret = null_handle_badblocks(cmd, sector, nr_sectors); in null_process_cmd()
1350 return null_handle_memory_backed(cmd, op, sector, nr_sectors); in null_process_cmd()
1356 sector_t nr_sectors, enum req_opf op) in null_handle_cmd() argument
[all …]
/drivers/block/xen-blkback/
Dcommon.h93 uint64_t nr_sectors; member
147 uint64_t nr_sectors; member
422 dst->u.discard.nr_sectors = src->u.discard.nr_sectors; in blkif_get_x86_32_req()
470 dst->u.discard.nr_sectors = src->u.discard.nr_sectors; in blkif_get_x86_64_req()
Dblkback.c979 preq.nr_sects = req->u.discard.nr_sectors; in dispatch_discard_io()
995 req->u.discard.nr_sectors, in dispatch_discard_io()
/drivers/md/
Ddm-zone.c430 unsigned int nr_sectors) in dm_zone_map_bio_end() argument
451 WRITE_ONCE(md->zwp_offset[zno], zwp_offset + nr_sectors); in dm_zone_map_bio_end()
458 if (nr_sectors != bio_sectors(orig_bio)) { in dm_zone_map_bio_end()
462 WRITE_ONCE(md->zwp_offset[zno], zwp_offset + nr_sectors); in dm_zone_map_bio_end()
Ddm-log-writes.c97 __le64 nr_sectors; member
126 sector_t nr_sectors; member
339 entry.nr_sectors = cpu_to_le64(block->nr_sectors); in log_one_block()
473 lc->next_sector += dev_to_bio_sectors(lc, block->nr_sectors); in log_writes_kthread()
726 block->nr_sectors = bio_to_dev_sectors(lc, bio_sectors(bio)); in log_writes_map()
941 block->nr_sectors = ALIGN(bytes, lc->sectorsize) >> lc->sectorshift; in log_dax()
Dmd.h560 static inline void md_sync_acct(struct block_device *bdev, unsigned long nr_sectors) in md_sync_acct() argument
562 atomic_add(nr_sectors, &bdev->bd_disk->sync_io); in md_sync_acct()
565 static inline void md_sync_acct_bio(struct bio *bio, unsigned long nr_sectors) in md_sync_acct_bio() argument
567 md_sync_acct(bio->bi_bdev, nr_sectors); in md_sync_acct_bio()
Ddm-zoned-target.c632 unsigned int nr_sectors = bio_sectors(bio); in dmz_map() local
641 bio_op(bio), (unsigned long long)sector, nr_sectors, in dmz_map()
646 if (!nr_sectors && bio_op(bio) != REQ_OP_WRITE) in dmz_map()
650 if ((nr_sectors & DMZ_BLOCK_SECTORS_MASK) || (sector & DMZ_BLOCK_SECTORS_MASK)) in dmz_map()
660 if (!nr_sectors && bio_op(bio) == REQ_OP_WRITE) { in dmz_map()
670 if (chunk_sector + nr_sectors > dmz_zone_nr_sectors(zmd)) in dmz_map()
Draid10.c3302 sector_t max_sector, nr_sectors; in raid10_sync_request() local
3809 nr_sectors = 0; in raid10_sync_request()
3828 nr_sectors += len>>9; in raid10_sync_request()
3831 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3836 if (conf->cluster_sync_high < sector_nr + nr_sectors) { in raid10_sync_request()
3857 if (conf->cluster_sync_high < sect_va1 + nr_sectors) { in raid10_sync_request()
3885 r10_bio->sectors = nr_sectors; in raid10_sync_request()
3888 md_sync_acct_bio(bio, nr_sectors); in raid10_sync_request()
3900 return sectors_skipped + nr_sectors; in raid10_sync_request()
4817 int nr_sectors; in reshape_request() local
[all …]
Draid1.c2652 sector_t max_sector, nr_sectors; in raid1_sync_request() local
2868 nr_sectors = 0; in raid1_sync_request()
2902 nr_sectors += len>>9; in raid1_sync_request()
2907 r1_bio->sectors = nr_sectors; in raid1_sync_request()
2910 conf->cluster_sync_high < sector_nr + nr_sectors) { in raid1_sync_request()
2928 md_sync_acct_bio(bio, nr_sectors); in raid1_sync_request()
2937 md_sync_acct_bio(bio, nr_sectors); in raid1_sync_request()
2942 return nr_sectors; in raid1_sync_request()
/drivers/block/drbd/
Ddrbd_actlog.c852 sector_t esector, nr_sectors; in __drbd_change_sync() local
868 nr_sectors = get_capacity(device->vdisk); in __drbd_change_sync()
871 if (!expect(sector < nr_sectors)) in __drbd_change_sync()
873 if (!expect(esector < nr_sectors)) in __drbd_change_sync()
874 esector = nr_sectors - 1; in __drbd_change_sync()
876 lbnr = BM_SECT_TO_BIT(nr_sectors-1); in __drbd_change_sync()
883 if (unlikely(esector == (nr_sectors-1))) in __drbd_change_sync()
Ddrbd_req.c889 sector_t esector, nr_sectors; in drbd_may_do_local_read() local
896 nr_sectors = get_capacity(device->vdisk); in drbd_may_do_local_read()
897 D_ASSERT(device, sector < nr_sectors); in drbd_may_do_local_read()
898 D_ASSERT(device, esector < nr_sectors); in drbd_may_do_local_read()
Ddrbd_receiver.c1513 …discard_or_zero_out(struct drbd_device *device, sector_t start, unsigned int nr_sectors, int flags) in drbd_issue_discard_or_zero_out() argument
1534 if (nr_sectors < granularity) in drbd_issue_discard_or_zero_out()
1539 if (nr_sectors < 2*granularity) in drbd_issue_discard_or_zero_out()
1549 nr_sectors -= nr; in drbd_issue_discard_or_zero_out()
1552 while (nr_sectors >= max_discard_sectors) { in drbd_issue_discard_or_zero_out()
1554 nr_sectors -= max_discard_sectors; in drbd_issue_discard_or_zero_out()
1557 if (nr_sectors) { in drbd_issue_discard_or_zero_out()
1562 nr = nr_sectors; in drbd_issue_discard_or_zero_out()
1566 nr_sectors -= nr; in drbd_issue_discard_or_zero_out()
1571 if (nr_sectors) { in drbd_issue_discard_or_zero_out()
[all …]
Ddrbd_int.h1545 sector_t start, unsigned int nr_sectors, int flags);
/drivers/block/
Dfloppy.c2259 unsigned int nr_sectors = current_count_sectors; in floppy_end_request() local
2264 nr_sectors = blk_rq_cur_sectors(req); in floppy_end_request()
2265 if (blk_update_request(req, error, nr_sectors << 9)) in floppy_end_request()
2321 int nr_sectors; in rw_interrupt() local
2345 nr_sectors = (((reply_buffer[R_TRACK] - raw_cmd->cmd[TRACK]) * heads + in rw_interrupt()
2349 if (nr_sectors / ssize > in rw_interrupt()
2352 nr_sectors, current_count_sectors); in rw_interrupt()
2365 nr_sectors -= in_sector_offset; in rw_interrupt()
2366 INFBOUND(nr_sectors, 0); in rw_interrupt()
2367 SUPBOUND(current_count_sectors, nr_sectors); in rw_interrupt()
[all …]
Dxen-blkfront.c563 ring_req->u.discard.nr_sectors = blk_rq_sectors(req); in blkif_queue_discard_req()