GNU Linux-libre 5.10.215-gnu1
[releases.git] / drivers / nvme / host / zns.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2020 Western Digital Corporation or its affiliates.
4  */
5
6 #include <linux/blkdev.h>
7 #include <linux/vmalloc.h>
8 #include "nvme.h"
9
10 int nvme_revalidate_zones(struct nvme_ns *ns)
11 {
12         struct request_queue *q = ns->queue;
13         int ret;
14
15         ret = blk_revalidate_disk_zones(ns->disk, NULL);
16         if (!ret)
17                 blk_queue_max_zone_append_sectors(q, ns->ctrl->max_zone_append);
18         return ret;
19 }
20
21 static int nvme_set_max_append(struct nvme_ctrl *ctrl)
22 {
23         struct nvme_command c = { };
24         struct nvme_id_ctrl_zns *id;
25         int status;
26
27         id = kzalloc(sizeof(*id), GFP_KERNEL);
28         if (!id)
29                 return -ENOMEM;
30
31         c.identify.opcode = nvme_admin_identify;
32         c.identify.cns = NVME_ID_CNS_CS_CTRL;
33         c.identify.csi = NVME_CSI_ZNS;
34
35         status = nvme_submit_sync_cmd(ctrl->admin_q, &c, id, sizeof(*id));
36         if (status) {
37                 kfree(id);
38                 return status;
39         }
40
41         if (id->zasl)
42                 ctrl->max_zone_append = 1 << (id->zasl + 3);
43         else
44                 ctrl->max_zone_append = ctrl->max_hw_sectors;
45         kfree(id);
46         return 0;
47 }
48
49 int nvme_update_zone_info(struct nvme_ns *ns, unsigned lbaf)
50 {
51         struct nvme_effects_log *log = ns->head->effects;
52         struct request_queue *q = ns->queue;
53         struct nvme_command c = { };
54         struct nvme_id_ns_zns *id;
55         int status;
56
57         /* Driver requires zone append support */
58         if (!(le32_to_cpu(log->iocs[nvme_cmd_zone_append]) &
59                         NVME_CMD_EFFECTS_CSUPP)) {
60                 dev_warn(ns->ctrl->device,
61                         "append not supported for zoned namespace:%d\n",
62                         ns->head->ns_id);
63                 return -EINVAL;
64         }
65
66         /* Lazily query controller append limit for the first zoned namespace */
67         if (!ns->ctrl->max_zone_append) {
68                 status = nvme_set_max_append(ns->ctrl);
69                 if (status)
70                         return status;
71         }
72
73         id = kzalloc(sizeof(*id), GFP_KERNEL);
74         if (!id)
75                 return -ENOMEM;
76
77         c.identify.opcode = nvme_admin_identify;
78         c.identify.nsid = cpu_to_le32(ns->head->ns_id);
79         c.identify.cns = NVME_ID_CNS_CS_NS;
80         c.identify.csi = NVME_CSI_ZNS;
81
82         status = nvme_submit_sync_cmd(ns->ctrl->admin_q, &c, id, sizeof(*id));
83         if (status)
84                 goto free_data;
85
86         /*
87          * We currently do not handle devices requiring any of the zoned
88          * operation characteristics.
89          */
90         if (id->zoc) {
91                 dev_warn(ns->ctrl->device,
92                         "zone operations:%x not supported for namespace:%u\n",
93                         le16_to_cpu(id->zoc), ns->head->ns_id);
94                 status = -EINVAL;
95                 goto free_data;
96         }
97
98         ns->zsze = nvme_lba_to_sect(ns, le64_to_cpu(id->lbafe[lbaf].zsze));
99         if (!is_power_of_2(ns->zsze)) {
100                 dev_warn(ns->ctrl->device,
101                         "invalid zone size:%llu for namespace:%u\n",
102                         ns->zsze, ns->head->ns_id);
103                 status = -EINVAL;
104                 goto free_data;
105         }
106
107         q->limits.zoned = BLK_ZONED_HM;
108         blk_queue_flag_set(QUEUE_FLAG_ZONE_RESETALL, q);
109         blk_queue_max_open_zones(q, le32_to_cpu(id->mor) + 1);
110         blk_queue_max_active_zones(q, le32_to_cpu(id->mar) + 1);
111 free_data:
112         kfree(id);
113         return status;
114 }
115
116 static void *nvme_zns_alloc_report_buffer(struct nvme_ns *ns,
117                                           unsigned int nr_zones, size_t *buflen)
118 {
119         struct request_queue *q = ns->disk->queue;
120         size_t bufsize;
121         void *buf;
122
123         const size_t min_bufsize = sizeof(struct nvme_zone_report) +
124                                    sizeof(struct nvme_zone_descriptor);
125
126         nr_zones = min_t(unsigned int, nr_zones,
127                          get_capacity(ns->disk) >> ilog2(ns->zsze));
128
129         bufsize = sizeof(struct nvme_zone_report) +
130                 nr_zones * sizeof(struct nvme_zone_descriptor);
131         bufsize = min_t(size_t, bufsize,
132                         queue_max_hw_sectors(q) << SECTOR_SHIFT);
133         bufsize = min_t(size_t, bufsize, queue_max_segments(q) << PAGE_SHIFT);
134
135         while (bufsize >= min_bufsize) {
136                 buf = __vmalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
137                 if (buf) {
138                         *buflen = bufsize;
139                         return buf;
140                 }
141                 bufsize >>= 1;
142         }
143         return NULL;
144 }
145
146 static int nvme_zone_parse_entry(struct nvme_ns *ns,
147                                  struct nvme_zone_descriptor *entry,
148                                  unsigned int idx, report_zones_cb cb,
149                                  void *data)
150 {
151         struct blk_zone zone = { };
152
153         if ((entry->zt & 0xf) != NVME_ZONE_TYPE_SEQWRITE_REQ) {
154                 dev_err(ns->ctrl->device, "invalid zone type %#x\n",
155                                 entry->zt);
156                 return -EINVAL;
157         }
158
159         zone.type = BLK_ZONE_TYPE_SEQWRITE_REQ;
160         zone.cond = entry->zs >> 4;
161         zone.len = ns->zsze;
162         zone.capacity = nvme_lba_to_sect(ns, le64_to_cpu(entry->zcap));
163         zone.start = nvme_lba_to_sect(ns, le64_to_cpu(entry->zslba));
164         zone.wp = nvme_lba_to_sect(ns, le64_to_cpu(entry->wp));
165
166         return cb(&zone, idx, data);
167 }
168
169 static int nvme_ns_report_zones(struct nvme_ns *ns, sector_t sector,
170                         unsigned int nr_zones, report_zones_cb cb, void *data)
171 {
172         struct nvme_zone_report *report;
173         struct nvme_command c = { };
174         int ret, zone_idx = 0;
175         unsigned int nz, i;
176         size_t buflen;
177
178         report = nvme_zns_alloc_report_buffer(ns, nr_zones, &buflen);
179         if (!report)
180                 return -ENOMEM;
181
182         c.zmr.opcode = nvme_cmd_zone_mgmt_recv;
183         c.zmr.nsid = cpu_to_le32(ns->head->ns_id);
184         c.zmr.numd = cpu_to_le32(nvme_bytes_to_numd(buflen));
185         c.zmr.zra = NVME_ZRA_ZONE_REPORT;
186         c.zmr.zrasf = NVME_ZRASF_ZONE_REPORT_ALL;
187         c.zmr.pr = NVME_REPORT_ZONE_PARTIAL;
188
189         sector &= ~(ns->zsze - 1);
190         while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) {
191                 memset(report, 0, buflen);
192
193                 c.zmr.slba = cpu_to_le64(nvme_sect_to_lba(ns, sector));
194                 ret = nvme_submit_sync_cmd(ns->queue, &c, report, buflen);
195                 if (ret) {
196                         if (ret > 0)
197                                 ret = -EIO;
198                         goto out_free;
199                 }
200
201                 nz = min((unsigned int)le64_to_cpu(report->nr_zones), nr_zones);
202                 if (!nz)
203                         break;
204
205                 for (i = 0; i < nz && zone_idx < nr_zones; i++) {
206                         ret = nvme_zone_parse_entry(ns, &report->entries[i],
207                                                     zone_idx, cb, data);
208                         if (ret)
209                                 goto out_free;
210                         zone_idx++;
211                 }
212
213                 sector += ns->zsze * nz;
214         }
215
216         if (zone_idx > 0)
217                 ret = zone_idx;
218         else
219                 ret = -EINVAL;
220 out_free:
221         kvfree(report);
222         return ret;
223 }
224
225 int nvme_report_zones(struct gendisk *disk, sector_t sector,
226                       unsigned int nr_zones, report_zones_cb cb, void *data)
227 {
228         struct nvme_ns_head *head = NULL;
229         struct nvme_ns *ns;
230         int srcu_idx, ret;
231
232         ns = nvme_get_ns_from_disk(disk, &head, &srcu_idx);
233         if (unlikely(!ns))
234                 return -EWOULDBLOCK;
235
236         if (ns->head->ids.csi == NVME_CSI_ZNS)
237                 ret = nvme_ns_report_zones(ns, sector, nr_zones, cb, data);
238         else
239                 ret = -EINVAL;
240         nvme_put_ns_from_disk(head, srcu_idx);
241
242         return ret;
243 }
244
245 blk_status_t nvme_setup_zone_mgmt_send(struct nvme_ns *ns, struct request *req,
246                 struct nvme_command *c, enum nvme_zone_mgmt_action action)
247 {
248         c->zms.opcode = nvme_cmd_zone_mgmt_send;
249         c->zms.nsid = cpu_to_le32(ns->head->ns_id);
250         c->zms.slba = cpu_to_le64(nvme_sect_to_lba(ns, blk_rq_pos(req)));
251         c->zms.zsa = action;
252
253         if (req_op(req) == REQ_OP_ZONE_RESET_ALL)
254                 c->zms.select_all = 1;
255
256         return BLK_STS_OK;
257 }