/drivers/s390/block/ |
D | dasd_fba.c | 771 unsigned int max_bytes, max_discard_sectors; in dasd_fba_setup_blk_queue() local 790 max_discard_sectors = max_bytes / logical_block_size; in dasd_fba_setup_blk_queue() 792 blk_queue_max_discard_sectors(q, max_discard_sectors); in dasd_fba_setup_blk_queue() 793 blk_queue_max_write_zeroes_sectors(q, max_discard_sectors); in dasd_fba_setup_blk_queue()
|
/drivers/block/rnbd/ |
D | rnbd-clt.h | 126 u32 max_discard_sectors; member
|
D | rnbd-proto.h | 140 __le32 max_discard_sectors; member
|
D | rnbd-clt.c | 86 dev->max_discard_sectors = le32_to_cpu(rsp->max_discard_sectors); in rnbd_clt_set_dev_attr() 1371 blk_queue_max_discard_sectors(dev->queue, dev->max_discard_sectors); in setup_request_queue() 1374 if (dev->max_discard_sectors) in setup_request_queue() 1611 dev->max_write_same_sectors, dev->max_discard_sectors, in rnbd_clt_map_device()
|
D | rnbd-srv.c | 564 rsp->max_discard_sectors = in rnbd_srv_fill_msg_open_rsp()
|
/drivers/block/ |
D | loop.c | 926 u32 granularity, max_discard_sectors; in loop_config_discard() local 938 max_discard_sectors = backingq->limits.max_write_zeroes_sectors; in loop_config_discard() 949 max_discard_sectors = 0; in loop_config_discard() 955 max_discard_sectors = UINT_MAX >> 9; in loop_config_discard() 959 max_discard_sectors = 0; in loop_config_discard() 962 if (max_discard_sectors) { in loop_config_discard() 964 blk_queue_max_discard_sectors(q, max_discard_sectors); in loop_config_discard() 965 blk_queue_max_write_zeroes_sectors(q, max_discard_sectors); in loop_config_discard()
|
D | virtio_blk.c | 914 max_discard_sectors, &v); in virtblk_probe()
|
/drivers/md/ |
D | dm-clone-target.c | 2044 else if (dest_limits->max_discard_sectors < clone->region_size) in disable_passdown_if_not_supported() 2062 limits->max_discard_sectors = round_down(UINT_MAX >> SECTOR_SHIFT, clone->region_size); in set_discard_limits() 2071 limits->max_discard_sectors = dest_limits->max_discard_sectors; in set_discard_limits()
|
D | dm-rq.c | 219 !clone->q->limits.max_discard_sectors) in dm_done()
|
D | dm-io.c | 315 special_cmd_max_sectors = q->limits.max_discard_sectors; in do_region()
|
D | dm-bow.c | 640 if (limits->max_discard_sectors == 0) { in dm_bow_io_hints() 643 limits->max_discard_sectors = 1 << 15; in dm_bow_io_hints()
|
D | dm-cache-target.c | 3369 else if (origin_limits->max_discard_sectors < cache->sectors_per_block) in disable_passdown_if_not_supported() 3386 limits->max_discard_sectors = min_t(sector_t, cache->discard_block_size * 1024, in set_discard_limits() 3396 limits->max_discard_sectors = origin_limits->max_discard_sectors; in set_discard_limits()
|
D | dm-log-writes.c | 899 limits->max_discard_sectors = (UINT_MAX >> SECTOR_SHIFT); in log_writes_io_hints()
|
D | dm-zoned-target.c | 1016 limits->max_discard_sectors = chunk_sectors; in dmz_io_hints()
|
D | dm.c | 877 limits->max_discard_sectors = 0; in disable_discard() 913 !q->limits.max_discard_sectors) in clone_endio()
|
D | dm-table.c | 2054 q->limits.max_discard_sectors = 0; in dm_table_set_restrictions()
|
D | dm-thin.c | 2847 else if (data_limits->max_discard_sectors < pool->sectors_per_block) in disable_passdown_if_not_supported() 4490 limits->max_discard_sectors = 2048 * 1024 * 16; /* 16G */ in thin_io_hints()
|
D | dm-snap.c | 2435 limits->max_discard_sectors = snap->store->chunk_size; in snapshot_io_hints()
|
D | raid5.c | 7791 mddev->queue->limits.max_discard_sectors >= (stripe >> 9) && in raid5_run()
|
/drivers/block/drbd/ |
D | drbd_receiver.c | 1518 unsigned int max_discard_sectors, granularity; in drbd_issue_discard_or_zero_out() local 1529 max_discard_sectors = min(q->limits.max_discard_sectors, (1U << 22)); in drbd_issue_discard_or_zero_out() 1530 max_discard_sectors -= max_discard_sectors % granularity; in drbd_issue_discard_or_zero_out() 1531 if (unlikely(!max_discard_sectors)) in drbd_issue_discard_or_zero_out() 1552 while (nr_sectors >= max_discard_sectors) { in drbd_issue_discard_or_zero_out() 1553 err |= blkdev_issue_discard(bdev, start, max_discard_sectors, GFP_NOIO, 0); in drbd_issue_discard_or_zero_out() 1554 nr_sectors -= max_discard_sectors; in drbd_issue_discard_or_zero_out() 1555 start += max_discard_sectors; in drbd_issue_discard_or_zero_out()
|
D | drbd_nl.c | 1230 q->limits.max_discard_sectors = drbd_max_discard_sectors(connection); in decide_on_discard_support() 1236 q->limits.max_discard_sectors = 0; in decide_on_discard_support() 1532 if (disk_conf->rs_discard_granularity > q->limits.max_discard_sectors << 9) in sanitize_disk_conf() 1533 disk_conf->rs_discard_granularity = q->limits.max_discard_sectors << 9; in sanitize_disk_conf()
|
/drivers/nvme/host/ |
D | nvme.h | 289 u32 max_discard_sectors; member
|
D | core.c | 1710 if (ctrl->max_discard_sectors == 0) { in nvme_config_discard() 1728 blk_queue_max_discard_sectors(queue, ctrl->max_discard_sectors); in nvme_config_discard() 2900 ctrl->max_discard_sectors = UINT_MAX; in nvme_init_non_mdts_limits() 2903 ctrl->max_discard_sectors = 0; in nvme_init_non_mdts_limits() 2937 ctrl->max_discard_sectors = le32_to_cpu(id->dmrsl); in nvme_init_non_mdts_limits()
|
/drivers/target/ |
D | target_core_device.c | 841 q->limits.max_discard_sectors >> (ilog2(block_size) - 9); in target_configure_unmap_from_queue()
|