Lines Matching refs:disk
114 struct gendisk *disk; member
142 #define nbd_name(nbd) ((nbd)->disk->disk_name)
162 return disk_to_dev(nbd->disk); in nbd_to_dev()
209 struct gendisk *disk = dev_to_disk(dev); in pid_show() local
210 struct nbd_device *nbd = (struct nbd_device *)disk->private_data; in pid_show()
222 struct gendisk *disk = nbd->disk; in nbd_dev_remove() local
225 if (disk) { in nbd_dev_remove()
226 q = disk->queue; in nbd_dev_remove()
227 del_gendisk(disk); in nbd_dev_remove()
230 disk->private_data = NULL; in nbd_dev_remove()
231 put_disk(disk); in nbd_dev_remove()
294 set_capacity(nbd->disk, 0); in nbd_size_clear()
302 struct block_device *bdev = bdget_disk(nbd->disk, 0); in nbd_size_update()
305 nbd->disk->queue->limits.discard_granularity = config->blksize; in nbd_size_update()
306 nbd->disk->queue->limits.discard_alignment = config->blksize; in nbd_size_update()
307 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_size_update()
309 blk_queue_logical_block_size(nbd->disk->queue, config->blksize); in nbd_size_update()
310 blk_queue_physical_block_size(nbd->disk->queue, config->blksize); in nbd_size_update()
311 set_capacity(nbd->disk, config->bytesize >> 9); in nbd_size_update()
362 dev_warn(disk_to_dev(nbd->disk), "shutting down sockets\n"); in sock_shutdown()
469 dev_err_ratelimited(disk_to_dev(nbd->disk), in sock_xmit()
539 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
597 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_send_cmd()
638 dev_err(disk_to_dev(nbd->disk), in nbd_send_cmd()
681 dev_err(disk_to_dev(nbd->disk), in nbd_read_stat()
687 dev_err(disk_to_dev(nbd->disk), "Wrong magic (0x%lx)\n", in nbd_read_stat()
699 dev_err(disk_to_dev(nbd->disk), "Unexpected reply (%d) %p\n", in nbd_read_stat()
708 dev_err(disk_to_dev(nbd->disk), "Double reply on req %p, cmd_cookie %u, handle cookie %u\n", in nbd_read_stat()
714 dev_err(disk_to_dev(nbd->disk), "Command already handled %p\n", in nbd_read_stat()
720 dev_err(disk_to_dev(nbd->disk), "Raced with timeout on req %p\n", in nbd_read_stat()
726 dev_err(disk_to_dev(nbd->disk), "Other side returned error (%d)\n", in nbd_read_stat()
741 dev_err(disk_to_dev(nbd->disk), "Receive data failed (result %d)\n", in nbd_read_stat()
810 blk_mq_quiesce_queue(nbd->disk->queue); in nbd_clear_que()
812 blk_mq_unquiesce_queue(nbd->disk->queue); in nbd_clear_que()
813 dev_dbg(disk_to_dev(nbd->disk), "queue cleared\n"); in nbd_clear_que()
827 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
850 dev_err_ratelimited(disk_to_dev(nbd->disk), in find_fallback()
880 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
888 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
938 dev_err_ratelimited(disk_to_dev(nbd->disk), in nbd_handle_cmd()
994 dev_err(disk_to_dev(nbd->disk), "Unsupported socket: shutdown callout must be supported.\n"); in nbd_get_socket()
1023 dev_err(disk_to_dev(nbd->disk), in nbd_add_socket()
1129 set_disk_ro(nbd->disk, true); in nbd_parse_flags()
1131 set_disk_ro(nbd->disk, false); in nbd_parse_flags()
1133 blk_queue_flag_set(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_parse_flags()
1136 blk_queue_write_cache(nbd->disk->queue, true, true); in nbd_parse_flags()
1138 blk_queue_write_cache(nbd->disk->queue, true, false); in nbd_parse_flags()
1141 blk_queue_write_cache(nbd->disk->queue, false, false); in nbd_parse_flags()
1162 dev_err(disk_to_dev(nbd->disk), in send_disconnects()
1172 dev_info(disk_to_dev(nbd->disk), "NBD_DISCONNECT\n"); in nbd_disconnect()
1195 device_remove_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_config_put()
1214 nbd->disk->queue->limits.discard_granularity = 0; in nbd_config_put()
1215 nbd->disk->queue->limits.discard_alignment = 0; in nbd_config_put()
1216 blk_queue_max_discard_sectors(nbd->disk->queue, UINT_MAX); in nbd_config_put()
1217 blk_queue_flag_clear(QUEUE_FLAG_DISCARD, nbd->disk->queue); in nbd_config_put()
1237 dev_err(disk_to_dev(nbd->disk), "server does not support multiple connections per device.\n"); in nbd_start_device()
1245 dev_err(disk_to_dev(nbd->disk), "Could not allocate knbd recv work queue.\n"); in nbd_start_device()
1254 error = device_create_file(disk_to_dev(nbd->disk), &pid_attr); in nbd_start_device()
1256 dev_err(disk_to_dev(nbd->disk), "device_create_file failed!\n"); in nbd_start_device()
1336 blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ); in nbd_set_cmd_timeout()
1480 static void nbd_release(struct gendisk *disk, fmode_t mode) in nbd_release() argument
1482 struct nbd_device *nbd = disk->private_data; in nbd_release()
1483 struct block_device *bdev = bdget_disk(disk, 0); in nbd_release()
1651 struct gendisk *disk; in nbd_dev_add() local
1659 disk = alloc_disk(1 << part_shift); in nbd_dev_add()
1660 if (!disk) in nbd_dev_add()
1677 nbd->disk = disk; in nbd_dev_add()
1697 disk->queue = q; in nbd_dev_add()
1702 blk_queue_flag_set(QUEUE_FLAG_NONROT, disk->queue); in nbd_dev_add()
1703 blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, disk->queue); in nbd_dev_add()
1704 disk->queue->limits.discard_granularity = 0; in nbd_dev_add()
1705 disk->queue->limits.discard_alignment = 0; in nbd_dev_add()
1706 blk_queue_max_discard_sectors(disk->queue, 0); in nbd_dev_add()
1707 blk_queue_max_segment_size(disk->queue, UINT_MAX); in nbd_dev_add()
1708 blk_queue_max_segments(disk->queue, USHRT_MAX); in nbd_dev_add()
1709 blk_queue_max_hw_sectors(disk->queue, 65536); in nbd_dev_add()
1710 disk->queue->limits.max_sectors = 256; in nbd_dev_add()
1716 disk->major = NBD_MAJOR; in nbd_dev_add()
1717 disk->first_minor = index << part_shift; in nbd_dev_add()
1718 disk->fops = &nbd_fops; in nbd_dev_add()
1719 disk->private_data = nbd; in nbd_dev_add()
1720 sprintf(disk->disk_name, "nbd%d", index); in nbd_dev_add()
1721 add_disk(disk); in nbd_dev_add()
1730 put_disk(disk); in nbd_dev_add()