• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * SCSI Zoned Block commands
3  *
4  * Copyright (C) 2014-2015 SUSE Linux GmbH
5  * Written by: Hannes Reinecke <hare@suse.de>
6  * Modified by: Damien Le Moal <damien.lemoal@hgst.com>
7  * Modified by: Shaun Tancheff <shaun.tancheff@seagate.com>
8  *
9  * This program is free software; you can redistribute it and/or
10  * modify it under the terms of the GNU General Public License version
11  * 2 as published by the Free Software Foundation.
12  *
13  * This program is distributed in the hope that it will be useful, but
14  * WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
16  * General Public License for more details.
17  *
18  * You should have received a copy of the GNU General Public License
19  * along with this program; see the file COPYING.  If not, write to
20  * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139,
21  * USA.
22  *
23  */
24 
25 #include <linux/blkdev.h>
26 
27 #include <asm/unaligned.h>
28 
29 #include <scsi/scsi.h>
30 #include <scsi/scsi_cmnd.h>
31 #include <scsi/scsi_dbg.h>
32 #include <scsi/scsi_device.h>
33 #include <scsi/scsi_driver.h>
34 #include <scsi/scsi_host.h>
35 #include <scsi/scsi_eh.h>
36 
37 #include "sd.h"
38 #include "scsi_priv.h"
39 
40 enum zbc_zone_type {
41 	ZBC_ZONE_TYPE_CONV = 0x1,
42 	ZBC_ZONE_TYPE_SEQWRITE_REQ,
43 	ZBC_ZONE_TYPE_SEQWRITE_PREF,
44 	ZBC_ZONE_TYPE_RESERVED,
45 };
46 
47 enum zbc_zone_cond {
48 	ZBC_ZONE_COND_NO_WP,
49 	ZBC_ZONE_COND_EMPTY,
50 	ZBC_ZONE_COND_IMP_OPEN,
51 	ZBC_ZONE_COND_EXP_OPEN,
52 	ZBC_ZONE_COND_CLOSED,
53 	ZBC_ZONE_COND_READONLY = 0xd,
54 	ZBC_ZONE_COND_FULL,
55 	ZBC_ZONE_COND_OFFLINE,
56 };
57 
58 /**
59  * Convert a zone descriptor to a zone struct.
60  */
sd_zbc_parse_report(struct scsi_disk * sdkp,u8 * buf,struct blk_zone * zone)61 static void sd_zbc_parse_report(struct scsi_disk *sdkp,
62 				u8 *buf,
63 				struct blk_zone *zone)
64 {
65 	struct scsi_device *sdp = sdkp->device;
66 
67 	memset(zone, 0, sizeof(struct blk_zone));
68 
69 	zone->type = buf[0] & 0x0f;
70 	zone->cond = (buf[1] >> 4) & 0xf;
71 	if (buf[1] & 0x01)
72 		zone->reset = 1;
73 	if (buf[1] & 0x02)
74 		zone->non_seq = 1;
75 
76 	zone->len = logical_to_sectors(sdp, get_unaligned_be64(&buf[8]));
77 	zone->start = logical_to_sectors(sdp, get_unaligned_be64(&buf[16]));
78 	zone->wp = logical_to_sectors(sdp, get_unaligned_be64(&buf[24]));
79 	if (zone->type != ZBC_ZONE_TYPE_CONV &&
80 	    zone->cond == ZBC_ZONE_COND_FULL)
81 		zone->wp = zone->start + zone->len;
82 }
83 
84 /**
85  * Issue a REPORT ZONES scsi command.
86  */
sd_zbc_report_zones(struct scsi_disk * sdkp,unsigned char * buf,unsigned int buflen,sector_t lba)87 static int sd_zbc_report_zones(struct scsi_disk *sdkp, unsigned char *buf,
88 			       unsigned int buflen, sector_t lba)
89 {
90 	struct scsi_device *sdp = sdkp->device;
91 	const int timeout = sdp->request_queue->rq_timeout;
92 	struct scsi_sense_hdr sshdr;
93 	unsigned char cmd[16];
94 	unsigned int rep_len;
95 	int result;
96 
97 	memset(cmd, 0, 16);
98 	cmd[0] = ZBC_IN;
99 	cmd[1] = ZI_REPORT_ZONES;
100 	put_unaligned_be64(lba, &cmd[2]);
101 	put_unaligned_be32(buflen, &cmd[10]);
102 	memset(buf, 0, buflen);
103 
104 	result = scsi_execute_req(sdp, cmd, DMA_FROM_DEVICE,
105 				  buf, buflen, &sshdr,
106 				  timeout, SD_MAX_RETRIES, NULL);
107 	if (result) {
108 		sd_printk(KERN_ERR, sdkp,
109 			  "REPORT ZONES lba %llu failed with %d/%d\n",
110 			  (unsigned long long)lba,
111 			  host_byte(result), driver_byte(result));
112 		return -EIO;
113 	}
114 
115 	rep_len = get_unaligned_be32(&buf[0]);
116 	if (rep_len < 64) {
117 		sd_printk(KERN_ERR, sdkp,
118 			  "REPORT ZONES report invalid length %u\n",
119 			  rep_len);
120 		return -EIO;
121 	}
122 
123 	return 0;
124 }
125 
sd_zbc_setup_report_cmnd(struct scsi_cmnd * cmd)126 int sd_zbc_setup_report_cmnd(struct scsi_cmnd *cmd)
127 {
128 	struct request *rq = cmd->request;
129 	struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
130 	sector_t lba, sector = blk_rq_pos(rq);
131 	unsigned int nr_bytes = blk_rq_bytes(rq);
132 	int ret;
133 
134 	WARN_ON(nr_bytes == 0);
135 
136 	if (!sd_is_zoned(sdkp))
137 		/* Not a zoned device */
138 		return BLKPREP_KILL;
139 
140 	ret = scsi_init_io(cmd);
141 	if (ret != BLKPREP_OK)
142 		return ret;
143 
144 	cmd->cmd_len = 16;
145 	memset(cmd->cmnd, 0, cmd->cmd_len);
146 	cmd->cmnd[0] = ZBC_IN;
147 	cmd->cmnd[1] = ZI_REPORT_ZONES;
148 	lba = sectors_to_logical(sdkp->device, sector);
149 	put_unaligned_be64(lba, &cmd->cmnd[2]);
150 	put_unaligned_be32(nr_bytes, &cmd->cmnd[10]);
151 	/* Do partial report for speeding things up */
152 	cmd->cmnd[14] = ZBC_REPORT_ZONE_PARTIAL;
153 
154 	cmd->sc_data_direction = DMA_FROM_DEVICE;
155 	cmd->sdb.length = nr_bytes;
156 	cmd->transfersize = sdkp->device->sector_size;
157 	cmd->allowed = 0;
158 
159 	/*
160 	 * Report may return less bytes than requested. Make sure
161 	 * to report completion on the entire initial request.
162 	 */
163 	rq->__data_len = nr_bytes;
164 
165 	return BLKPREP_OK;
166 }
167 
sd_zbc_report_zones_complete(struct scsi_cmnd * scmd,unsigned int good_bytes)168 static void sd_zbc_report_zones_complete(struct scsi_cmnd *scmd,
169 					 unsigned int good_bytes)
170 {
171 	struct request *rq = scmd->request;
172 	struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
173 	struct sg_mapping_iter miter;
174 	struct blk_zone_report_hdr hdr;
175 	struct blk_zone zone;
176 	unsigned int offset, bytes = 0;
177 	unsigned long flags;
178 	u8 *buf;
179 
180 	if (good_bytes < 64)
181 		return;
182 
183 	memset(&hdr, 0, sizeof(struct blk_zone_report_hdr));
184 
185 	sg_miter_start(&miter, scsi_sglist(scmd), scsi_sg_count(scmd),
186 		       SG_MITER_TO_SG | SG_MITER_ATOMIC);
187 
188 	local_irq_save(flags);
189 	while (sg_miter_next(&miter) && bytes < good_bytes) {
190 
191 		buf = miter.addr;
192 		offset = 0;
193 
194 		if (bytes == 0) {
195 			/* Set the report header */
196 			hdr.nr_zones = min_t(unsigned int,
197 					 (good_bytes - 64) / 64,
198 					 get_unaligned_be32(&buf[0]) / 64);
199 			memcpy(buf, &hdr, sizeof(struct blk_zone_report_hdr));
200 			offset += 64;
201 			bytes += 64;
202 		}
203 
204 		/* Parse zone descriptors */
205 		while (offset < miter.length && hdr.nr_zones) {
206 			WARN_ON(offset > miter.length);
207 			buf = miter.addr + offset;
208 			sd_zbc_parse_report(sdkp, buf, &zone);
209 			memcpy(buf, &zone, sizeof(struct blk_zone));
210 			offset += 64;
211 			bytes += 64;
212 			hdr.nr_zones--;
213 		}
214 
215 		if (!hdr.nr_zones)
216 			break;
217 
218 	}
219 	sg_miter_stop(&miter);
220 	local_irq_restore(flags);
221 }
222 
sd_zbc_zone_sectors(struct scsi_disk * sdkp)223 static inline sector_t sd_zbc_zone_sectors(struct scsi_disk *sdkp)
224 {
225 	return logical_to_sectors(sdkp->device, sdkp->zone_blocks);
226 }
227 
sd_zbc_zone_no(struct scsi_disk * sdkp,sector_t sector)228 static inline unsigned int sd_zbc_zone_no(struct scsi_disk *sdkp,
229 					  sector_t sector)
230 {
231 	return sectors_to_logical(sdkp->device, sector) >> sdkp->zone_shift;
232 }
233 
sd_zbc_setup_reset_cmnd(struct scsi_cmnd * cmd)234 int sd_zbc_setup_reset_cmnd(struct scsi_cmnd *cmd)
235 {
236 	struct request *rq = cmd->request;
237 	struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
238 	sector_t sector = blk_rq_pos(rq);
239 	sector_t block = sectors_to_logical(sdkp->device, sector);
240 
241 	if (!sd_is_zoned(sdkp))
242 		/* Not a zoned device */
243 		return BLKPREP_KILL;
244 
245 	if (sdkp->device->changed)
246 		return BLKPREP_KILL;
247 
248 	if (sector & (sd_zbc_zone_sectors(sdkp) - 1))
249 		/* Unaligned request */
250 		return BLKPREP_KILL;
251 
252 	cmd->cmd_len = 16;
253 	memset(cmd->cmnd, 0, cmd->cmd_len);
254 	cmd->cmnd[0] = ZBC_OUT;
255 	cmd->cmnd[1] = ZO_RESET_WRITE_POINTER;
256 	put_unaligned_be64(block, &cmd->cmnd[2]);
257 
258 	rq->timeout = SD_TIMEOUT;
259 	cmd->sc_data_direction = DMA_NONE;
260 	cmd->transfersize = 0;
261 	cmd->allowed = 0;
262 
263 	return BLKPREP_OK;
264 }
265 
sd_zbc_write_lock_zone(struct scsi_cmnd * cmd)266 int sd_zbc_write_lock_zone(struct scsi_cmnd *cmd)
267 {
268 	struct request *rq = cmd->request;
269 	struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
270 	sector_t sector = blk_rq_pos(rq);
271 	sector_t zone_sectors = sd_zbc_zone_sectors(sdkp);
272 	unsigned int zno = sd_zbc_zone_no(sdkp, sector);
273 
274 	/*
275 	 * Note: Checks of the alignment of the write command on
276 	 * logical blocks is done in sd.c
277 	 */
278 
279 	/* Do not allow zone boundaries crossing on host-managed drives */
280 	if (blk_queue_zoned_model(sdkp->disk->queue) == BLK_ZONED_HM &&
281 	    (sector & (zone_sectors - 1)) + blk_rq_sectors(rq) > zone_sectors)
282 		return BLKPREP_KILL;
283 
284 	/*
285 	 * Do not issue more than one write at a time per
286 	 * zone. This solves write ordering problems due to
287 	 * the unlocking of the request queue in the dispatch
288 	 * path in the non scsi-mq case. For scsi-mq, this
289 	 * also avoids potential write reordering when multiple
290 	 * threads running on different CPUs write to the same
291 	 * zone (with a synchronized sequential pattern).
292 	 */
293 	if (sdkp->zones_wlock &&
294 	    test_and_set_bit(zno, sdkp->zones_wlock))
295 		return BLKPREP_DEFER;
296 
297 	WARN_ON_ONCE(cmd->flags & SCMD_ZONE_WRITE_LOCK);
298 	cmd->flags |= SCMD_ZONE_WRITE_LOCK;
299 
300 	return BLKPREP_OK;
301 }
302 
sd_zbc_write_unlock_zone(struct scsi_cmnd * cmd)303 void sd_zbc_write_unlock_zone(struct scsi_cmnd *cmd)
304 {
305 	struct request *rq = cmd->request;
306 	struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
307 
308 	if (sdkp->zones_wlock && cmd->flags & SCMD_ZONE_WRITE_LOCK) {
309 		unsigned int zno = sd_zbc_zone_no(sdkp, blk_rq_pos(rq));
310 		WARN_ON_ONCE(!test_bit(zno, sdkp->zones_wlock));
311 		cmd->flags &= ~SCMD_ZONE_WRITE_LOCK;
312 		clear_bit_unlock(zno, sdkp->zones_wlock);
313 		smp_mb__after_atomic();
314 	}
315 }
316 
sd_zbc_complete(struct scsi_cmnd * cmd,unsigned int good_bytes,struct scsi_sense_hdr * sshdr)317 void sd_zbc_complete(struct scsi_cmnd *cmd,
318 		     unsigned int good_bytes,
319 		     struct scsi_sense_hdr *sshdr)
320 {
321 	int result = cmd->result;
322 	struct request *rq = cmd->request;
323 
324 	switch (req_op(rq)) {
325 	case REQ_OP_ZONE_RESET:
326 
327 		if (result &&
328 		    sshdr->sense_key == ILLEGAL_REQUEST &&
329 		    sshdr->asc == 0x24)
330 			/*
331 			 * INVALID FIELD IN CDB error: reset of a conventional
332 			 * zone was attempted. Nothing to worry about, so be
333 			 * quiet about the error.
334 			 */
335 			rq->rq_flags |= RQF_QUIET;
336 		break;
337 
338 	case REQ_OP_WRITE:
339 	case REQ_OP_WRITE_ZEROES:
340 	case REQ_OP_WRITE_SAME:
341 
342 		if (result &&
343 		    sshdr->sense_key == ILLEGAL_REQUEST &&
344 		    sshdr->asc == 0x21)
345 			/*
346 			 * INVALID ADDRESS FOR WRITE error: It is unlikely that
347 			 * retrying write requests failed with any kind of
348 			 * alignement error will result in success. So don't.
349 			 */
350 			cmd->allowed = 0;
351 		break;
352 
353 	case REQ_OP_ZONE_REPORT:
354 
355 		if (!result)
356 			sd_zbc_report_zones_complete(cmd, good_bytes);
357 		break;
358 
359 	}
360 }
361 
362 /**
363  * Read zoned block device characteristics (VPD page B6).
364  */
sd_zbc_read_zoned_characteristics(struct scsi_disk * sdkp,unsigned char * buf)365 static int sd_zbc_read_zoned_characteristics(struct scsi_disk *sdkp,
366 					     unsigned char *buf)
367 {
368 
369 	if (scsi_get_vpd_page(sdkp->device, 0xb6, buf, 64)) {
370 		sd_printk(KERN_NOTICE, sdkp,
371 			  "Unconstrained-read check failed\n");
372 		return -ENODEV;
373 	}
374 
375 	if (sdkp->device->type != TYPE_ZBC) {
376 		/* Host-aware */
377 		sdkp->urswrz = 1;
378 		sdkp->zones_optimal_open = get_unaligned_be32(&buf[8]);
379 		sdkp->zones_optimal_nonseq = get_unaligned_be32(&buf[12]);
380 		sdkp->zones_max_open = 0;
381 	} else {
382 		/* Host-managed */
383 		sdkp->urswrz = buf[4] & 1;
384 		sdkp->zones_optimal_open = 0;
385 		sdkp->zones_optimal_nonseq = 0;
386 		sdkp->zones_max_open = get_unaligned_be32(&buf[16]);
387 	}
388 
389 	return 0;
390 }
391 
392 /**
393  * Check reported capacity.
394  */
sd_zbc_check_capacity(struct scsi_disk * sdkp,unsigned char * buf)395 static int sd_zbc_check_capacity(struct scsi_disk *sdkp,
396 				 unsigned char *buf)
397 {
398 	sector_t lba;
399 	int ret;
400 
401 	if (sdkp->rc_basis != 0)
402 		return 0;
403 
404 	/* Do a report zone to get the maximum LBA to check capacity */
405 	ret = sd_zbc_report_zones(sdkp, buf, SD_BUF_SIZE, 0);
406 	if (ret)
407 		return ret;
408 
409 	/* The max_lba field is the capacity of this device */
410 	lba = get_unaligned_be64(&buf[8]);
411 	if (lba + 1 == sdkp->capacity)
412 		return 0;
413 
414 	if (sdkp->first_scan)
415 		sd_printk(KERN_WARNING, sdkp,
416 			  "Changing capacity from %llu to max LBA+1 %llu\n",
417 			  (unsigned long long)sdkp->capacity,
418 			  (unsigned long long)lba + 1);
419 	sdkp->capacity = lba + 1;
420 
421 	return 0;
422 }
423 
424 #define SD_ZBC_BUF_SIZE 131072
425 
426 /**
427  * sd_zbc_check_zone_size - Check the device zone sizes
428  * @sdkp: Target disk
429  *
430  * Check that all zones of the device are equal. The last zone can however
431  * be smaller. The zone size must also be a power of two number of LBAs.
432  *
433  * Returns the zone size in number of blocks upon success or an error code
434  * upon failure.
435  */
sd_zbc_check_zone_size(struct scsi_disk * sdkp)436 static s64 sd_zbc_check_zone_size(struct scsi_disk *sdkp)
437 {
438 	u64 zone_blocks = 0;
439 	sector_t block = 0;
440 	unsigned char *buf;
441 	unsigned char *rec;
442 	unsigned int buf_len;
443 	unsigned int list_length;
444 	s64 ret;
445 	u8 same;
446 
447 	/* Get a buffer */
448 	buf = kmalloc(SD_ZBC_BUF_SIZE, GFP_KERNEL);
449 	if (!buf)
450 		return -ENOMEM;
451 
452 	/* Do a report zone to get the same field */
453 	ret = sd_zbc_report_zones(sdkp, buf, SD_ZBC_BUF_SIZE, 0);
454 	if (ret)
455 		goto out_free;
456 
457 	same = buf[4] & 0x0f;
458 	if (same > 0) {
459 		rec = &buf[64];
460 		zone_blocks = get_unaligned_be64(&rec[8]);
461 		goto out;
462 	}
463 
464 	/*
465 	 * Check the size of all zones: all zones must be of
466 	 * equal size, except the last zone which can be smaller
467 	 * than other zones.
468 	 */
469 	do {
470 
471 		/* Parse REPORT ZONES header */
472 		list_length = get_unaligned_be32(&buf[0]) + 64;
473 		rec = buf + 64;
474 		if (list_length < SD_ZBC_BUF_SIZE)
475 			buf_len = list_length;
476 		else
477 			buf_len = SD_ZBC_BUF_SIZE;
478 
479 		/* Parse zone descriptors */
480 		while (rec < buf + buf_len) {
481 			u64 this_zone_blocks = get_unaligned_be64(&rec[8]);
482 
483 			if (zone_blocks == 0) {
484 				zone_blocks = this_zone_blocks;
485 			} else if (this_zone_blocks != zone_blocks &&
486 				   (block + this_zone_blocks < sdkp->capacity
487 				    || this_zone_blocks > zone_blocks)) {
488 				zone_blocks = 0;
489 				goto out;
490 			}
491 			block += this_zone_blocks;
492 			rec += 64;
493 		}
494 
495 		if (block < sdkp->capacity) {
496 			ret = sd_zbc_report_zones(sdkp, buf,
497 						  SD_ZBC_BUF_SIZE, block);
498 			if (ret)
499 				goto out_free;
500 		}
501 
502 	} while (block < sdkp->capacity);
503 
504 out:
505 	if (!zone_blocks) {
506 		if (sdkp->first_scan)
507 			sd_printk(KERN_NOTICE, sdkp,
508 				  "Devices with non constant zone "
509 				  "size are not supported\n");
510 		ret = -ENODEV;
511 	} else if (!is_power_of_2(zone_blocks)) {
512 		if (sdkp->first_scan)
513 			sd_printk(KERN_NOTICE, sdkp,
514 				  "Devices with non power of 2 zone "
515 				  "size are not supported\n");
516 		ret = -ENODEV;
517 	} else if (logical_to_sectors(sdkp->device, zone_blocks) > UINT_MAX) {
518 		if (sdkp->first_scan)
519 			sd_printk(KERN_NOTICE, sdkp,
520 				  "Zone size too large\n");
521 		ret = -ENODEV;
522 	} else {
523 		ret = zone_blocks;
524 	}
525 
526 out_free:
527 	kfree(buf);
528 
529 	return ret;
530 }
531 
sd_zbc_setup(struct scsi_disk * sdkp,u32 zone_blocks)532 static int sd_zbc_setup(struct scsi_disk *sdkp, u32 zone_blocks)
533 {
534 	struct request_queue *q = sdkp->disk->queue;
535 	u32 zone_shift = ilog2(zone_blocks);
536 	u32 nr_zones;
537 
538 	/* chunk_sectors indicates the zone size */
539 	blk_queue_chunk_sectors(q,
540 			logical_to_sectors(sdkp->device, zone_blocks));
541 	nr_zones = round_up(sdkp->capacity, zone_blocks) >> zone_shift;
542 
543 	/*
544 	 * Initialize the disk zone write lock bitmap if the number
545 	 * of zones changed.
546 	 */
547 	if (nr_zones != sdkp->nr_zones) {
548 		unsigned long *zones_wlock = NULL;
549 
550 		if (nr_zones) {
551 			zones_wlock = kcalloc(BITS_TO_LONGS(nr_zones),
552 					      sizeof(unsigned long),
553 					      GFP_KERNEL);
554 			if (!zones_wlock)
555 				return -ENOMEM;
556 		}
557 
558 		blk_mq_freeze_queue(q);
559 		sdkp->zone_blocks = zone_blocks;
560 		sdkp->zone_shift = zone_shift;
561 		sdkp->nr_zones = nr_zones;
562 		swap(sdkp->zones_wlock, zones_wlock);
563 		blk_mq_unfreeze_queue(q);
564 
565 		kfree(zones_wlock);
566 
567 		/* READ16/WRITE16 is mandatory for ZBC disks */
568 		sdkp->device->use_16_for_rw = 1;
569 		sdkp->device->use_10_for_rw = 0;
570 	}
571 
572 	return 0;
573 }
574 
sd_zbc_read_zones(struct scsi_disk * sdkp,unsigned char * buf)575 int sd_zbc_read_zones(struct scsi_disk *sdkp,
576 		      unsigned char *buf)
577 {
578 	int64_t zone_blocks;
579 	int ret;
580 
581 	if (!sd_is_zoned(sdkp))
582 		/*
583 		 * Device managed or normal SCSI disk,
584 		 * no special handling required
585 		 */
586 		return 0;
587 
588 
589 	/* Get zoned block device characteristics */
590 	ret = sd_zbc_read_zoned_characteristics(sdkp, buf);
591 	if (ret)
592 		goto err;
593 
594 	/*
595 	 * Check for unconstrained reads: host-managed devices with
596 	 * constrained reads (drives failing read after write pointer)
597 	 * are not supported.
598 	 */
599 	if (!sdkp->urswrz) {
600 		if (sdkp->first_scan)
601 			sd_printk(KERN_NOTICE, sdkp,
602 			  "constrained reads devices are not supported\n");
603 		ret = -ENODEV;
604 		goto err;
605 	}
606 
607 	/* Check capacity */
608 	ret = sd_zbc_check_capacity(sdkp, buf);
609 	if (ret)
610 		goto err;
611 
612 	/*
613 	 * Check zone size: only devices with a constant zone size (except
614 	 * an eventual last runt zone) that is a power of 2 are supported.
615 	 */
616 	zone_blocks = sd_zbc_check_zone_size(sdkp);
617 	ret = -EFBIG;
618 	if (zone_blocks != (u32)zone_blocks)
619 		goto err;
620 	ret = zone_blocks;
621 	if (ret < 0)
622 		goto err;
623 
624 	/* The drive satisfies the kernel restrictions: set it up */
625 	ret = sd_zbc_setup(sdkp, zone_blocks);
626 	if (ret)
627 		goto err;
628 
629 	return 0;
630 
631 err:
632 	sdkp->capacity = 0;
633 
634 	return ret;
635 }
636 
sd_zbc_remove(struct scsi_disk * sdkp)637 void sd_zbc_remove(struct scsi_disk *sdkp)
638 {
639 	kfree(sdkp->zones_wlock);
640 	sdkp->zones_wlock = NULL;
641 	sdkp->nr_zones = 0;
642 }
643 
sd_zbc_print_zones(struct scsi_disk * sdkp)644 void sd_zbc_print_zones(struct scsi_disk *sdkp)
645 {
646 	if (!sd_is_zoned(sdkp) || !sdkp->capacity)
647 		return;
648 
649 	if (sdkp->capacity & (sdkp->zone_blocks - 1))
650 		sd_printk(KERN_NOTICE, sdkp,
651 			  "%u zones of %u logical blocks + 1 runt zone\n",
652 			  sdkp->nr_zones - 1,
653 			  sdkp->zone_blocks);
654 	else
655 		sd_printk(KERN_NOTICE, sdkp,
656 			  "%u zones of %u logical blocks\n",
657 			  sdkp->nr_zones,
658 			  sdkp->zone_blocks);
659 }
660