]> Git Repo - linux.git/blob - drivers/nvme/host/zns.c
Linux 6.14-rc3
[linux.git] / drivers / nvme / host / zns.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2020 Western Digital Corporation or its affiliates.
4  */
5
6 #include <linux/blkdev.h>
7 #include <linux/vmalloc.h>
8 #include "nvme.h"
9
10 static int nvme_set_max_append(struct nvme_ctrl *ctrl)
11 {
12         struct nvme_command c = { };
13         struct nvme_id_ctrl_zns *id;
14         int status;
15
16         id = kzalloc(sizeof(*id), GFP_KERNEL);
17         if (!id)
18                 return -ENOMEM;
19
20         c.identify.opcode = nvme_admin_identify;
21         c.identify.cns = NVME_ID_CNS_CS_CTRL;
22         c.identify.csi = NVME_CSI_ZNS;
23
24         status = nvme_submit_sync_cmd(ctrl->admin_q, &c, id, sizeof(*id));
25         if (status) {
26                 kfree(id);
27                 return status;
28         }
29
30         if (id->zasl)
31                 ctrl->max_zone_append = 1 << (id->zasl + 3);
32         else
33                 ctrl->max_zone_append = ctrl->max_hw_sectors;
34         kfree(id);
35         return 0;
36 }
37
38 int nvme_query_zone_info(struct nvme_ns *ns, unsigned lbaf,
39                 struct nvme_zone_info *zi)
40 {
41         struct nvme_effects_log *log = ns->head->effects;
42         struct nvme_command c = { };
43         struct nvme_id_ns_zns *id;
44         int status;
45
46         /* Driver requires zone append support */
47         if ((le32_to_cpu(log->iocs[nvme_cmd_zone_append]) &
48                         NVME_CMD_EFFECTS_CSUPP)) {
49                 if (test_and_clear_bit(NVME_NS_FORCE_RO, &ns->flags))
50                         dev_warn(ns->ctrl->device,
51                                  "Zone Append supported for zoned namespace:%d. Remove read-only mode\n",
52                                  ns->head->ns_id);
53         } else {
54                 set_bit(NVME_NS_FORCE_RO, &ns->flags);
55                 dev_warn(ns->ctrl->device,
56                          "Zone Append not supported for zoned namespace:%d. Forcing to read-only mode\n",
57                          ns->head->ns_id);
58         }
59
60         /* Lazily query controller append limit for the first zoned namespace */
61         if (!ns->ctrl->max_zone_append) {
62                 status = nvme_set_max_append(ns->ctrl);
63                 if (status)
64                         return status;
65         }
66
67         id = kzalloc(sizeof(*id), GFP_KERNEL);
68         if (!id)
69                 return -ENOMEM;
70
71         c.identify.opcode = nvme_admin_identify;
72         c.identify.nsid = cpu_to_le32(ns->head->ns_id);
73         c.identify.cns = NVME_ID_CNS_CS_NS;
74         c.identify.csi = NVME_CSI_ZNS;
75
76         status = nvme_submit_sync_cmd(ns->ctrl->admin_q, &c, id, sizeof(*id));
77         if (status)
78                 goto free_data;
79
80         /*
81          * We currently do not handle devices requiring any of the zoned
82          * operation characteristics.
83          */
84         if (id->zoc) {
85                 dev_warn(ns->ctrl->device,
86                         "zone operations:%x not supported for namespace:%u\n",
87                         le16_to_cpu(id->zoc), ns->head->ns_id);
88                 status = -ENODEV;
89                 goto free_data;
90         }
91
92         zi->zone_size = le64_to_cpu(id->lbafe[lbaf].zsze);
93         if (!is_power_of_2(zi->zone_size)) {
94                 dev_warn(ns->ctrl->device,
95                         "invalid zone size: %llu for namespace: %u\n",
96                         zi->zone_size, ns->head->ns_id);
97                 status = -ENODEV;
98                 goto free_data;
99         }
100         zi->max_open_zones = le32_to_cpu(id->mor) + 1;
101         zi->max_active_zones = le32_to_cpu(id->mar) + 1;
102
103 free_data:
104         kfree(id);
105         return status;
106 }
107
108 void nvme_update_zone_info(struct nvme_ns *ns, struct queue_limits *lim,
109                 struct nvme_zone_info *zi)
110 {
111         lim->features |= BLK_FEAT_ZONED;
112         lim->max_open_zones = zi->max_open_zones;
113         lim->max_active_zones = zi->max_active_zones;
114         lim->max_hw_zone_append_sectors = ns->ctrl->max_zone_append;
115         lim->chunk_sectors = ns->head->zsze =
116                 nvme_lba_to_sect(ns->head, zi->zone_size);
117 }
118
119 static void *nvme_zns_alloc_report_buffer(struct nvme_ns *ns,
120                                           unsigned int nr_zones, size_t *buflen)
121 {
122         struct request_queue *q = ns->disk->queue;
123         size_t bufsize;
124         void *buf;
125
126         const size_t min_bufsize = sizeof(struct nvme_zone_report) +
127                                    sizeof(struct nvme_zone_descriptor);
128
129         nr_zones = min_t(unsigned int, nr_zones,
130                          get_capacity(ns->disk) >> ilog2(ns->head->zsze));
131
132         bufsize = sizeof(struct nvme_zone_report) +
133                 nr_zones * sizeof(struct nvme_zone_descriptor);
134         bufsize = min_t(size_t, bufsize,
135                         queue_max_hw_sectors(q) << SECTOR_SHIFT);
136         bufsize = min_t(size_t, bufsize, queue_max_segments(q) << PAGE_SHIFT);
137
138         while (bufsize >= min_bufsize) {
139                 buf = __vmalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
140                 if (buf) {
141                         *buflen = bufsize;
142                         return buf;
143                 }
144                 bufsize >>= 1;
145         }
146         return NULL;
147 }
148
149 static int nvme_zone_parse_entry(struct nvme_ctrl *ctrl,
150                                  struct nvme_ns_head *head,
151                                  struct nvme_zone_descriptor *entry,
152                                  unsigned int idx, report_zones_cb cb,
153                                  void *data)
154 {
155         struct blk_zone zone = { };
156
157         if ((entry->zt & 0xf) != NVME_ZONE_TYPE_SEQWRITE_REQ) {
158                 dev_err(ctrl->device, "invalid zone type %#x\n",
159                                 entry->zt);
160                 return -EINVAL;
161         }
162
163         zone.type = BLK_ZONE_TYPE_SEQWRITE_REQ;
164         zone.cond = entry->zs >> 4;
165         zone.len = head->zsze;
166         zone.capacity = nvme_lba_to_sect(head, le64_to_cpu(entry->zcap));
167         zone.start = nvme_lba_to_sect(head, le64_to_cpu(entry->zslba));
168         if (zone.cond == BLK_ZONE_COND_FULL)
169                 zone.wp = zone.start + zone.len;
170         else
171                 zone.wp = nvme_lba_to_sect(head, le64_to_cpu(entry->wp));
172
173         return cb(&zone, idx, data);
174 }
175
176 int nvme_ns_report_zones(struct nvme_ns *ns, sector_t sector,
177                 unsigned int nr_zones, report_zones_cb cb, void *data)
178 {
179         struct nvme_zone_report *report;
180         struct nvme_command c = { };
181         int ret, zone_idx = 0;
182         unsigned int nz, i;
183         size_t buflen;
184
185         if (ns->head->ids.csi != NVME_CSI_ZNS)
186                 return -EINVAL;
187
188         report = nvme_zns_alloc_report_buffer(ns, nr_zones, &buflen);
189         if (!report)
190                 return -ENOMEM;
191
192         c.zmr.opcode = nvme_cmd_zone_mgmt_recv;
193         c.zmr.nsid = cpu_to_le32(ns->head->ns_id);
194         c.zmr.numd = cpu_to_le32(nvme_bytes_to_numd(buflen));
195         c.zmr.zra = NVME_ZRA_ZONE_REPORT;
196         c.zmr.zrasf = NVME_ZRASF_ZONE_REPORT_ALL;
197         c.zmr.pr = NVME_REPORT_ZONE_PARTIAL;
198
199         sector &= ~(ns->head->zsze - 1);
200         while (zone_idx < nr_zones && sector < get_capacity(ns->disk)) {
201                 memset(report, 0, buflen);
202
203                 c.zmr.slba = cpu_to_le64(nvme_sect_to_lba(ns->head, sector));
204                 ret = nvme_submit_sync_cmd(ns->queue, &c, report, buflen);
205                 if (ret) {
206                         if (ret > 0)
207                                 ret = -EIO;
208                         goto out_free;
209                 }
210
211                 nz = min((unsigned int)le64_to_cpu(report->nr_zones), nr_zones);
212                 if (!nz)
213                         break;
214
215                 for (i = 0; i < nz && zone_idx < nr_zones; i++) {
216                         ret = nvme_zone_parse_entry(ns->ctrl, ns->head,
217                                                     &report->entries[i],
218                                                     zone_idx, cb, data);
219                         if (ret)
220                                 goto out_free;
221                         zone_idx++;
222                 }
223
224                 sector += ns->head->zsze * nz;
225         }
226
227         if (zone_idx > 0)
228                 ret = zone_idx;
229         else
230                 ret = -EINVAL;
231 out_free:
232         kvfree(report);
233         return ret;
234 }
235
236 blk_status_t nvme_setup_zone_mgmt_send(struct nvme_ns *ns, struct request *req,
237                 struct nvme_command *c, enum nvme_zone_mgmt_action action)
238 {
239         memset(c, 0, sizeof(*c));
240
241         c->zms.opcode = nvme_cmd_zone_mgmt_send;
242         c->zms.nsid = cpu_to_le32(ns->head->ns_id);
243         c->zms.slba = cpu_to_le64(nvme_sect_to_lba(ns->head, blk_rq_pos(req)));
244         c->zms.zsa = action;
245
246         if (req_op(req) == REQ_OP_ZONE_RESET_ALL)
247                 c->zms.select_all = 1;
248
249         return BLK_STS_OK;
250 }
This page took 0.045201 seconds and 4 git commands to generate.