GNU Linux-libre 4.14.290-gnu1
[releases.git] / drivers / scsi / sd_zbc.c
1 /*
2  * SCSI Zoned Block commands
3  *
4  * Copyright (C) 2014-2015 SUSE Linux GmbH
5  * Written by: Hannes Reinecke <hare@suse.de>
6  * Modified by: Damien Le Moal <damien.lemoal@hgst.com>
7  * Modified by: Shaun Tancheff <shaun.tancheff@seagate.com>
8  *
9  * This program is free software; you can redistribute it and/or
10  * modify it under the terms of the GNU General Public License version
11  * 2 as published by the Free Software Foundation.
12  *
13  * This program is distributed in the hope that it will be useful, but
14  * WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
16  * General Public License for more details.
17  *
18  * You should have received a copy of the GNU General Public License
19  * along with this program; see the file COPYING.  If not, write to
20  * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139,
21  * USA.
22  *
23  */
24
25 #include <linux/blkdev.h>
26
27 #include <asm/unaligned.h>
28
29 #include <scsi/scsi.h>
30 #include <scsi/scsi_cmnd.h>
31 #include <scsi/scsi_dbg.h>
32 #include <scsi/scsi_device.h>
33 #include <scsi/scsi_driver.h>
34 #include <scsi/scsi_host.h>
35 #include <scsi/scsi_eh.h>
36
37 #include "sd.h"
38 #include "scsi_priv.h"
39
40 enum zbc_zone_type {
41         ZBC_ZONE_TYPE_CONV = 0x1,
42         ZBC_ZONE_TYPE_SEQWRITE_REQ,
43         ZBC_ZONE_TYPE_SEQWRITE_PREF,
44         ZBC_ZONE_TYPE_RESERVED,
45 };
46
47 enum zbc_zone_cond {
48         ZBC_ZONE_COND_NO_WP,
49         ZBC_ZONE_COND_EMPTY,
50         ZBC_ZONE_COND_IMP_OPEN,
51         ZBC_ZONE_COND_EXP_OPEN,
52         ZBC_ZONE_COND_CLOSED,
53         ZBC_ZONE_COND_READONLY = 0xd,
54         ZBC_ZONE_COND_FULL,
55         ZBC_ZONE_COND_OFFLINE,
56 };
57
58 /**
59  * Convert a zone descriptor to a zone struct.
60  */
61 static void sd_zbc_parse_report(struct scsi_disk *sdkp,
62                                 u8 *buf,
63                                 struct blk_zone *zone)
64 {
65         struct scsi_device *sdp = sdkp->device;
66
67         memset(zone, 0, sizeof(struct blk_zone));
68
69         zone->type = buf[0] & 0x0f;
70         zone->cond = (buf[1] >> 4) & 0xf;
71         if (buf[1] & 0x01)
72                 zone->reset = 1;
73         if (buf[1] & 0x02)
74                 zone->non_seq = 1;
75
76         zone->len = logical_to_sectors(sdp, get_unaligned_be64(&buf[8]));
77         zone->start = logical_to_sectors(sdp, get_unaligned_be64(&buf[16]));
78         zone->wp = logical_to_sectors(sdp, get_unaligned_be64(&buf[24]));
79         if (zone->type != ZBC_ZONE_TYPE_CONV &&
80             zone->cond == ZBC_ZONE_COND_FULL)
81                 zone->wp = zone->start + zone->len;
82 }
83
84 /**
85  * Issue a REPORT ZONES scsi command.
86  */
87 static int sd_zbc_report_zones(struct scsi_disk *sdkp, unsigned char *buf,
88                                unsigned int buflen, sector_t lba)
89 {
90         struct scsi_device *sdp = sdkp->device;
91         const int timeout = sdp->request_queue->rq_timeout;
92         struct scsi_sense_hdr sshdr;
93         unsigned char cmd[16];
94         unsigned int rep_len;
95         int result;
96
97         memset(cmd, 0, 16);
98         cmd[0] = ZBC_IN;
99         cmd[1] = ZI_REPORT_ZONES;
100         put_unaligned_be64(lba, &cmd[2]);
101         put_unaligned_be32(buflen, &cmd[10]);
102         memset(buf, 0, buflen);
103
104         result = scsi_execute_req(sdp, cmd, DMA_FROM_DEVICE,
105                                   buf, buflen, &sshdr,
106                                   timeout, SD_MAX_RETRIES, NULL);
107         if (result) {
108                 sd_printk(KERN_ERR, sdkp,
109                           "REPORT ZONES lba %llu failed with %d/%d\n",
110                           (unsigned long long)lba,
111                           host_byte(result), driver_byte(result));
112                 return -EIO;
113         }
114
115         rep_len = get_unaligned_be32(&buf[0]);
116         if (rep_len < 64) {
117                 sd_printk(KERN_ERR, sdkp,
118                           "REPORT ZONES report invalid length %u\n",
119                           rep_len);
120                 return -EIO;
121         }
122
123         return 0;
124 }
125
126 int sd_zbc_setup_report_cmnd(struct scsi_cmnd *cmd)
127 {
128         struct request *rq = cmd->request;
129         struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
130         sector_t lba, sector = blk_rq_pos(rq);
131         unsigned int nr_bytes = blk_rq_bytes(rq);
132         int ret;
133
134         WARN_ON(nr_bytes == 0);
135
136         if (!sd_is_zoned(sdkp))
137                 /* Not a zoned device */
138                 return BLKPREP_KILL;
139
140         ret = scsi_init_io(cmd);
141         if (ret != BLKPREP_OK)
142                 return ret;
143
144         cmd->cmd_len = 16;
145         memset(cmd->cmnd, 0, cmd->cmd_len);
146         cmd->cmnd[0] = ZBC_IN;
147         cmd->cmnd[1] = ZI_REPORT_ZONES;
148         lba = sectors_to_logical(sdkp->device, sector);
149         put_unaligned_be64(lba, &cmd->cmnd[2]);
150         put_unaligned_be32(nr_bytes, &cmd->cmnd[10]);
151         /* Do partial report for speeding things up */
152         cmd->cmnd[14] = ZBC_REPORT_ZONE_PARTIAL;
153
154         cmd->sc_data_direction = DMA_FROM_DEVICE;
155         cmd->sdb.length = nr_bytes;
156         cmd->transfersize = sdkp->device->sector_size;
157         cmd->allowed = 0;
158
159         /*
160          * Report may return less bytes than requested. Make sure
161          * to report completion on the entire initial request.
162          */
163         rq->__data_len = nr_bytes;
164
165         return BLKPREP_OK;
166 }
167
168 static void sd_zbc_report_zones_complete(struct scsi_cmnd *scmd,
169                                          unsigned int good_bytes)
170 {
171         struct request *rq = scmd->request;
172         struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
173         struct sg_mapping_iter miter;
174         struct blk_zone_report_hdr hdr;
175         struct blk_zone zone;
176         unsigned int offset, bytes = 0;
177         unsigned long flags;
178         u8 *buf;
179
180         if (good_bytes < 64)
181                 return;
182
183         memset(&hdr, 0, sizeof(struct blk_zone_report_hdr));
184
185         sg_miter_start(&miter, scsi_sglist(scmd), scsi_sg_count(scmd),
186                        SG_MITER_TO_SG | SG_MITER_ATOMIC);
187
188         local_irq_save(flags);
189         while (sg_miter_next(&miter) && bytes < good_bytes) {
190
191                 buf = miter.addr;
192                 offset = 0;
193
194                 if (bytes == 0) {
195                         /* Set the report header */
196                         hdr.nr_zones = min_t(unsigned int,
197                                          (good_bytes - 64) / 64,
198                                          get_unaligned_be32(&buf[0]) / 64);
199                         memcpy(buf, &hdr, sizeof(struct blk_zone_report_hdr));
200                         offset += 64;
201                         bytes += 64;
202                 }
203
204                 /* Parse zone descriptors */
205                 while (offset < miter.length && hdr.nr_zones) {
206                         WARN_ON(offset > miter.length);
207                         buf = miter.addr + offset;
208                         sd_zbc_parse_report(sdkp, buf, &zone);
209                         memcpy(buf, &zone, sizeof(struct blk_zone));
210                         offset += 64;
211                         bytes += 64;
212                         hdr.nr_zones--;
213                 }
214
215                 if (!hdr.nr_zones)
216                         break;
217
218         }
219         sg_miter_stop(&miter);
220         local_irq_restore(flags);
221 }
222
223 static inline sector_t sd_zbc_zone_sectors(struct scsi_disk *sdkp)
224 {
225         return logical_to_sectors(sdkp->device, sdkp->zone_blocks);
226 }
227
228 static inline unsigned int sd_zbc_zone_no(struct scsi_disk *sdkp,
229                                           sector_t sector)
230 {
231         return sectors_to_logical(sdkp->device, sector) >> sdkp->zone_shift;
232 }
233
234 int sd_zbc_setup_reset_cmnd(struct scsi_cmnd *cmd)
235 {
236         struct request *rq = cmd->request;
237         struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
238         sector_t sector = blk_rq_pos(rq);
239         sector_t block = sectors_to_logical(sdkp->device, sector);
240
241         if (!sd_is_zoned(sdkp))
242                 /* Not a zoned device */
243                 return BLKPREP_KILL;
244
245         if (sdkp->device->changed)
246                 return BLKPREP_KILL;
247
248         if (sector & (sd_zbc_zone_sectors(sdkp) - 1))
249                 /* Unaligned request */
250                 return BLKPREP_KILL;
251
252         cmd->cmd_len = 16;
253         memset(cmd->cmnd, 0, cmd->cmd_len);
254         cmd->cmnd[0] = ZBC_OUT;
255         cmd->cmnd[1] = ZO_RESET_WRITE_POINTER;
256         put_unaligned_be64(block, &cmd->cmnd[2]);
257
258         rq->timeout = SD_TIMEOUT;
259         cmd->sc_data_direction = DMA_NONE;
260         cmd->transfersize = 0;
261         cmd->allowed = 0;
262
263         return BLKPREP_OK;
264 }
265
266 int sd_zbc_write_lock_zone(struct scsi_cmnd *cmd)
267 {
268         struct request *rq = cmd->request;
269         struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
270         sector_t sector = blk_rq_pos(rq);
271         sector_t zone_sectors = sd_zbc_zone_sectors(sdkp);
272         unsigned int zno = sd_zbc_zone_no(sdkp, sector);
273
274         /*
275          * Note: Checks of the alignment of the write command on
276          * logical blocks is done in sd.c
277          */
278
279         /* Do not allow zone boundaries crossing on host-managed drives */
280         if (blk_queue_zoned_model(sdkp->disk->queue) == BLK_ZONED_HM &&
281             (sector & (zone_sectors - 1)) + blk_rq_sectors(rq) > zone_sectors)
282                 return BLKPREP_KILL;
283
284         /*
285          * Do not issue more than one write at a time per
286          * zone. This solves write ordering problems due to
287          * the unlocking of the request queue in the dispatch
288          * path in the non scsi-mq case. For scsi-mq, this
289          * also avoids potential write reordering when multiple
290          * threads running on different CPUs write to the same
291          * zone (with a synchronized sequential pattern).
292          */
293         if (sdkp->zones_wlock &&
294             test_and_set_bit(zno, sdkp->zones_wlock))
295                 return BLKPREP_DEFER;
296
297         WARN_ON_ONCE(cmd->flags & SCMD_ZONE_WRITE_LOCK);
298         cmd->flags |= SCMD_ZONE_WRITE_LOCK;
299
300         return BLKPREP_OK;
301 }
302
303 void sd_zbc_write_unlock_zone(struct scsi_cmnd *cmd)
304 {
305         struct request *rq = cmd->request;
306         struct scsi_disk *sdkp = scsi_disk(rq->rq_disk);
307
308         if (sdkp->zones_wlock && cmd->flags & SCMD_ZONE_WRITE_LOCK) {
309                 unsigned int zno = sd_zbc_zone_no(sdkp, blk_rq_pos(rq));
310                 WARN_ON_ONCE(!test_bit(zno, sdkp->zones_wlock));
311                 cmd->flags &= ~SCMD_ZONE_WRITE_LOCK;
312                 clear_bit_unlock(zno, sdkp->zones_wlock);
313                 smp_mb__after_atomic();
314         }
315 }
316
317 void sd_zbc_complete(struct scsi_cmnd *cmd,
318                      unsigned int good_bytes,
319                      struct scsi_sense_hdr *sshdr)
320 {
321         int result = cmd->result;
322         struct request *rq = cmd->request;
323
324         switch (req_op(rq)) {
325         case REQ_OP_ZONE_RESET:
326
327                 if (result &&
328                     sshdr->sense_key == ILLEGAL_REQUEST &&
329                     sshdr->asc == 0x24)
330                         /*
331                          * INVALID FIELD IN CDB error: reset of a conventional
332                          * zone was attempted. Nothing to worry about, so be
333                          * quiet about the error.
334                          */
335                         rq->rq_flags |= RQF_QUIET;
336                 break;
337
338         case REQ_OP_WRITE:
339         case REQ_OP_WRITE_ZEROES:
340         case REQ_OP_WRITE_SAME:
341
342                 if (result &&
343                     sshdr->sense_key == ILLEGAL_REQUEST &&
344                     sshdr->asc == 0x21)
345                         /*
346                          * INVALID ADDRESS FOR WRITE error: It is unlikely that
347                          * retrying write requests failed with any kind of
348                          * alignement error will result in success. So don't.
349                          */
350                         cmd->allowed = 0;
351                 break;
352
353         case REQ_OP_ZONE_REPORT:
354
355                 if (!result)
356                         sd_zbc_report_zones_complete(cmd, good_bytes);
357                 break;
358
359         }
360 }
361
362 /**
363  * Read zoned block device characteristics (VPD page B6).
364  */
365 static int sd_zbc_read_zoned_characteristics(struct scsi_disk *sdkp,
366                                              unsigned char *buf)
367 {
368
369         if (scsi_get_vpd_page(sdkp->device, 0xb6, buf, 64)) {
370                 sd_printk(KERN_NOTICE, sdkp,
371                           "Unconstrained-read check failed\n");
372                 return -ENODEV;
373         }
374
375         if (sdkp->device->type != TYPE_ZBC) {
376                 /* Host-aware */
377                 sdkp->urswrz = 1;
378                 sdkp->zones_optimal_open = get_unaligned_be32(&buf[8]);
379                 sdkp->zones_optimal_nonseq = get_unaligned_be32(&buf[12]);
380                 sdkp->zones_max_open = 0;
381         } else {
382                 /* Host-managed */
383                 sdkp->urswrz = buf[4] & 1;
384                 sdkp->zones_optimal_open = 0;
385                 sdkp->zones_optimal_nonseq = 0;
386                 sdkp->zones_max_open = get_unaligned_be32(&buf[16]);
387         }
388
389         return 0;
390 }
391
392 /**
393  * Check reported capacity.
394  */
395 static int sd_zbc_check_capacity(struct scsi_disk *sdkp,
396                                  unsigned char *buf)
397 {
398         sector_t lba;
399         int ret;
400
401         if (sdkp->rc_basis != 0)
402                 return 0;
403
404         /* Do a report zone to get the maximum LBA to check capacity */
405         ret = sd_zbc_report_zones(sdkp, buf, SD_BUF_SIZE, 0);
406         if (ret)
407                 return ret;
408
409         /* The max_lba field is the capacity of this device */
410         lba = get_unaligned_be64(&buf[8]);
411         if (lba + 1 == sdkp->capacity)
412                 return 0;
413
414         if (sdkp->first_scan)
415                 sd_printk(KERN_WARNING, sdkp,
416                           "Changing capacity from %llu to max LBA+1 %llu\n",
417                           (unsigned long long)sdkp->capacity,
418                           (unsigned long long)lba + 1);
419         sdkp->capacity = lba + 1;
420
421         return 0;
422 }
423
424 #define SD_ZBC_BUF_SIZE 131072
425
426 /**
427  * sd_zbc_check_zone_size - Check the device zone sizes
428  * @sdkp: Target disk
429  *
430  * Check that all zones of the device are equal. The last zone can however
431  * be smaller. The zone size must also be a power of two number of LBAs.
432  *
433  * Returns the zone size in number of blocks upon success or an error code
434  * upon failure.
435  */
436 static s64 sd_zbc_check_zone_size(struct scsi_disk *sdkp)
437 {
438         u64 zone_blocks = 0;
439         sector_t block = 0;
440         unsigned char *buf;
441         unsigned char *rec;
442         unsigned int buf_len;
443         unsigned int list_length;
444         s64 ret;
445         u8 same;
446
447         /* Get a buffer */
448         buf = kmalloc(SD_ZBC_BUF_SIZE, GFP_KERNEL);
449         if (!buf)
450                 return -ENOMEM;
451
452         /* Do a report zone to get the same field */
453         ret = sd_zbc_report_zones(sdkp, buf, SD_ZBC_BUF_SIZE, 0);
454         if (ret)
455                 goto out_free;
456
457         same = buf[4] & 0x0f;
458         if (same > 0) {
459                 rec = &buf[64];
460                 zone_blocks = get_unaligned_be64(&rec[8]);
461                 goto out;
462         }
463
464         /*
465          * Check the size of all zones: all zones must be of
466          * equal size, except the last zone which can be smaller
467          * than other zones.
468          */
469         do {
470
471                 /* Parse REPORT ZONES header */
472                 list_length = get_unaligned_be32(&buf[0]) + 64;
473                 rec = buf + 64;
474                 if (list_length < SD_ZBC_BUF_SIZE)
475                         buf_len = list_length;
476                 else
477                         buf_len = SD_ZBC_BUF_SIZE;
478
479                 /* Parse zone descriptors */
480                 while (rec < buf + buf_len) {
481                         u64 this_zone_blocks = get_unaligned_be64(&rec[8]);
482
483                         if (zone_blocks == 0) {
484                                 zone_blocks = this_zone_blocks;
485                         } else if (this_zone_blocks != zone_blocks &&
486                                    (block + this_zone_blocks < sdkp->capacity
487                                     || this_zone_blocks > zone_blocks)) {
488                                 zone_blocks = 0;
489                                 goto out;
490                         }
491                         block += this_zone_blocks;
492                         rec += 64;
493                 }
494
495                 if (block < sdkp->capacity) {
496                         ret = sd_zbc_report_zones(sdkp, buf,
497                                                   SD_ZBC_BUF_SIZE, block);
498                         if (ret)
499                                 goto out_free;
500                 }
501
502         } while (block < sdkp->capacity);
503
504 out:
505         if (!zone_blocks) {
506                 if (sdkp->first_scan)
507                         sd_printk(KERN_NOTICE, sdkp,
508                                   "Devices with non constant zone "
509                                   "size are not supported\n");
510                 ret = -ENODEV;
511         } else if (!is_power_of_2(zone_blocks)) {
512                 if (sdkp->first_scan)
513                         sd_printk(KERN_NOTICE, sdkp,
514                                   "Devices with non power of 2 zone "
515                                   "size are not supported\n");
516                 ret = -ENODEV;
517         } else if (logical_to_sectors(sdkp->device, zone_blocks) > UINT_MAX) {
518                 if (sdkp->first_scan)
519                         sd_printk(KERN_NOTICE, sdkp,
520                                   "Zone size too large\n");
521                 ret = -ENODEV;
522         } else {
523                 ret = zone_blocks;
524         }
525
526 out_free:
527         kfree(buf);
528
529         return ret;
530 }
531
532 static int sd_zbc_setup(struct scsi_disk *sdkp, u32 zone_blocks)
533 {
534         struct request_queue *q = sdkp->disk->queue;
535         u32 zone_shift = ilog2(zone_blocks);
536         u32 nr_zones;
537
538         /* chunk_sectors indicates the zone size */
539         blk_queue_chunk_sectors(q,
540                         logical_to_sectors(sdkp->device, zone_blocks));
541         nr_zones = round_up(sdkp->capacity, zone_blocks) >> zone_shift;
542
543         /*
544          * Initialize the disk zone write lock bitmap if the number
545          * of zones changed.
546          */
547         if (nr_zones != sdkp->nr_zones) {
548                 unsigned long *zones_wlock = NULL;
549
550                 if (nr_zones) {
551                         zones_wlock = kcalloc(BITS_TO_LONGS(nr_zones),
552                                               sizeof(unsigned long),
553                                               GFP_KERNEL);
554                         if (!zones_wlock)
555                                 return -ENOMEM;
556                 }
557
558                 blk_mq_freeze_queue(q);
559                 sdkp->zone_blocks = zone_blocks;
560                 sdkp->zone_shift = zone_shift;
561                 sdkp->nr_zones = nr_zones;
562                 swap(sdkp->zones_wlock, zones_wlock);
563                 blk_mq_unfreeze_queue(q);
564
565                 kfree(zones_wlock);
566
567                 /* READ16/WRITE16 is mandatory for ZBC disks */
568                 sdkp->device->use_16_for_rw = 1;
569                 sdkp->device->use_10_for_rw = 0;
570         }
571
572         return 0;
573 }
574
575 int sd_zbc_read_zones(struct scsi_disk *sdkp,
576                       unsigned char *buf)
577 {
578         int64_t zone_blocks;
579         int ret;
580
581         if (!sd_is_zoned(sdkp))
582                 /*
583                  * Device managed or normal SCSI disk,
584                  * no special handling required
585                  */
586                 return 0;
587
588
589         /* Get zoned block device characteristics */
590         ret = sd_zbc_read_zoned_characteristics(sdkp, buf);
591         if (ret)
592                 goto err;
593
594         /*
595          * Check for unconstrained reads: host-managed devices with
596          * constrained reads (drives failing read after write pointer)
597          * are not supported.
598          */
599         if (!sdkp->urswrz) {
600                 if (sdkp->first_scan)
601                         sd_printk(KERN_NOTICE, sdkp,
602                           "constrained reads devices are not supported\n");
603                 ret = -ENODEV;
604                 goto err;
605         }
606
607         /* Check capacity */
608         ret = sd_zbc_check_capacity(sdkp, buf);
609         if (ret)
610                 goto err;
611
612         /*
613          * Check zone size: only devices with a constant zone size (except
614          * an eventual last runt zone) that is a power of 2 are supported.
615          */
616         zone_blocks = sd_zbc_check_zone_size(sdkp);
617         ret = -EFBIG;
618         if (zone_blocks != (u32)zone_blocks)
619                 goto err;
620         ret = zone_blocks;
621         if (ret < 0)
622                 goto err;
623
624         /* The drive satisfies the kernel restrictions: set it up */
625         ret = sd_zbc_setup(sdkp, zone_blocks);
626         if (ret)
627                 goto err;
628
629         return 0;
630
631 err:
632         sdkp->capacity = 0;
633
634         return ret;
635 }
636
637 void sd_zbc_remove(struct scsi_disk *sdkp)
638 {
639         kfree(sdkp->zones_wlock);
640         sdkp->zones_wlock = NULL;
641         sdkp->nr_zones = 0;
642 }
643
644 void sd_zbc_print_zones(struct scsi_disk *sdkp)
645 {
646         if (!sd_is_zoned(sdkp) || !sdkp->capacity)
647                 return;
648
649         if (sdkp->capacity & (sdkp->zone_blocks - 1))
650                 sd_printk(KERN_NOTICE, sdkp,
651                           "%u zones of %u logical blocks + 1 runt zone\n",
652                           sdkp->nr_zones - 1,
653                           sdkp->zone_blocks);
654         else
655                 sd_printk(KERN_NOTICE, sdkp,
656                           "%u zones of %u logical blocks\n",
657                           sdkp->nr_zones,
658                           sdkp->zone_blocks);
659 }