1 | /* $NetBSD: nvmereg.h,v 1.13 2018/12/01 18:25:39 jdolecek Exp $ */ |
2 | /* $OpenBSD: nvmereg.h,v 1.10 2016/04/14 11:18:32 dlg Exp $ */ |
3 | |
4 | /* |
5 | * Copyright (c) 2014 David Gwynne <dlg@openbsd.org> |
6 | * |
7 | * Permission to use, copy, modify, and distribute this software for any |
8 | * purpose with or without fee is hereby granted, provided that the above |
9 | * copyright notice and this permission notice appear in all copies. |
10 | * |
11 | * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES |
12 | * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF |
13 | * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR |
14 | * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES |
15 | * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN |
16 | * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF |
17 | * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
18 | */ |
19 | |
20 | #ifndef __NVMEREG_H__ |
21 | #define __NVMEREG_H__ |
22 | |
23 | #ifndef NVME_CTASSERT |
24 | #define NVME_CTASSERT(x, s) __CTASSERT(x) |
25 | #endif |
26 | |
27 | #define NVME_CAP 0x0000 /* Controller Capabilities */ |
28 | #define NVME_CAP_MPSMAX(_r) (12 + (((_r) >> 52) & 0xf)) /* shift */ |
29 | #define NVME_CAP_MPSMIN(_r) (12 + (((_r) >> 48) & 0xf)) /* shift */ |
30 | #define NVME_CAP_CSS(_r) (((_r) >> 37) & 0x7f) |
31 | #define NVME_CAP_CSS_NVM __BIT(0) |
32 | #define NVME_CAP_NSSRS(_r) ISSET((_r), __BIT(36)) |
33 | #define NVME_CAP_DSTRD(_r) __BIT(2 + (((_r) >> 32) & 0xf)) /* bytes */ |
34 | #define NVME_CAP_TO(_r) (500 * (((_r) >> 24) & 0xff)) /* ms */ |
35 | #define NVME_CAP_AMS(_r) (((_r) >> 17) & 0x3) |
36 | #define NVME_CAP_AMS_WRR __BIT(0) |
37 | #define NVME_CAP_AMS_VENDOR __BIT(1) |
38 | #define NVME_CAP_CQR(_r) ISSET((_r), __BIT(16)) |
39 | #define NVME_CAP_MQES(_r) (((_r) & 0xffff) + 1) |
40 | #define NVME_CAP_LO 0x0000 |
41 | #define NVME_CAP_HI 0x0004 |
42 | #define NVME_VS 0x0008 /* Version */ |
43 | #define NVME_VS_MJR(_r) (((_r) >> 16) & 0xffff) |
44 | #define NVME_VS_MNR(_r) (((_r) >> 8) & 0xff) |
45 | #define NVME_VS_TER(_r) ((_r) & 0xff) |
46 | #define NVME_INTMS 0x000c /* Interrupt Mask Set */ |
47 | #define NVME_INTMC 0x0010 /* Interrupt Mask Clear */ |
48 | #define NVME_CC 0x0014 /* Controller Configuration */ |
49 | #define NVME_CC_IOCQES(_v) (((_v) & 0xf) << 20) |
50 | #define NVME_CC_IOCQES_MASK NVME_CC_IOCQES(0xf) |
51 | #define NVME_CC_IOCQES_R(_v) (((_v) >> 20) & 0xf) |
52 | #define NVME_CC_IOSQES(_v) (((_v) & 0xf) << 16) |
53 | #define NVME_CC_IOSQES_MASK NVME_CC_IOSQES(0xf) |
54 | #define NVME_CC_IOSQES_R(_v) (((_v) >> 16) & 0xf) |
55 | #define NVME_CC_SHN(_v) (((_v) & 0x3) << 14) |
56 | #define NVME_CC_SHN_MASK NVME_CC_SHN(0x3) |
57 | #define NVME_CC_SHN_R(_v) (((_v) >> 15) & 0x3) |
58 | #define NVME_CC_SHN_NONE 0 |
59 | #define NVME_CC_SHN_NORMAL 1 |
60 | #define NVME_CC_SHN_ABRUPT 2 |
61 | #define NVME_CC_AMS(_v) (((_v) & 0x7) << 11) |
62 | #define NVME_CC_AMS_MASK NVME_CC_AMS(0x7) |
63 | #define NVME_CC_AMS_R(_v) (((_v) >> 11) & 0xf) |
64 | #define NVME_CC_AMS_RR 0 /* round-robin */ |
65 | #define NVME_CC_AMS_WRR_U 1 /* weighted round-robin w/ urgent */ |
66 | #define NVME_CC_AMS_VENDOR 7 /* vendor */ |
67 | #define NVME_CC_MPS(_v) ((((_v) - 12) & 0xf) << 7) |
68 | #define NVME_CC_MPS_MASK (0xf << 7) |
69 | #define NVME_CC_MPS_R(_v) (12 + (((_v) >> 7) & 0xf)) |
70 | #define NVME_CC_CSS(_v) (((_v) & 0x7) << 4) |
71 | #define NVME_CC_CSS_MASK NVME_CC_CSS(0x7) |
72 | #define NVME_CC_CSS_R(_v) (((_v) >> 4) & 0x7) |
73 | #define NVME_CC_CSS_NVM 0 |
74 | #define NVME_CC_EN __BIT(0) |
75 | #define NVME_CSTS 0x001c /* Controller Status */ |
76 | #define NVME_CSTS_SHST_MASK (0x3 << 2) |
77 | #define NVME_CSTS_SHST_NONE (0x0 << 2) /* normal operation */ |
78 | #define NVME_CSTS_SHST_WAIT (0x1 << 2) /* shutdown processing occurring */ |
79 | #define NVME_CSTS_SHST_DONE (0x2 << 2) /* shutdown processing complete */ |
80 | #define NVME_CSTS_CFS (1 << 1) |
81 | #define NVME_CSTS_RDY (1 << 0) |
82 | #define NVME_NSSR 0x0020 /* NVM Subsystem Reset (Optional) */ |
83 | #define NVME_AQA 0x0024 /* Admin Queue Attributes */ |
84 | /* Admin Completion Queue Size */ |
85 | #define NVME_AQA_ACQS(_v) (((_v) - 1) << 16) |
86 | #define NVME_AQA_ACQS_R(_v) ((_v >> 16) & ((1 << 12) - 1)) |
87 | /* Admin Submission Queue Size */ |
88 | #define NVME_AQA_ASQS(_v) (((_v) - 1) << 0) |
89 | #define NVME_AQA_ASQS_R(_v) (_v & ((1 << 12) - 1)) |
90 | #define NVME_ASQ 0x0028 /* Admin Submission Queue Base Address */ |
91 | #define NVME_ACQ 0x0030 /* Admin Completion Queue Base Address */ |
92 | |
93 | #define NVME_ADMIN_Q 0 |
94 | /* Submission Queue Tail Doorbell */ |
95 | #define NVME_SQTDBL(_q, _s) (0x1000 + (2 * (_q) + 0) * (_s)) |
96 | /* Completion Queue Head Doorbell */ |
97 | #define NVME_CQHDBL(_q, _s) (0x1000 + (2 * (_q) + 1) * (_s)) |
98 | |
99 | struct nvme_sge { |
100 | uint8_t id; |
101 | uint8_t _reserved[15]; |
102 | } __packed __aligned(8); |
103 | NVME_CTASSERT(sizeof(struct nvme_sge) == 16, "bad size for nvme_sge" ); |
104 | |
105 | struct nvme_sge_data { |
106 | uint8_t id; |
107 | uint8_t _reserved[3]; |
108 | |
109 | uint32_t length; |
110 | |
111 | uint64_t address; |
112 | } __packed __aligned(8); |
113 | NVME_CTASSERT(sizeof(struct nvme_sge_data) == 16, "bad size for nvme_sge_data" ); |
114 | |
115 | struct nvme_sge_bit_bucket { |
116 | uint8_t id; |
117 | uint8_t _reserved[3]; |
118 | |
119 | uint32_t length; |
120 | |
121 | uint64_t address; |
122 | } __packed __aligned(8); |
123 | NVME_CTASSERT(sizeof(struct nvme_sge_bit_bucket) == 16, "bad size for nvme_sge_bit_bucket" ); |
124 | |
125 | struct nvme_sqe { |
126 | uint8_t opcode; |
127 | uint8_t flags; |
128 | uint16_t cid; |
129 | |
130 | uint32_t nsid; |
131 | |
132 | uint8_t _reserved[8]; |
133 | |
134 | uint64_t mptr; |
135 | |
136 | union { |
137 | uint64_t prp[2]; |
138 | struct nvme_sge sge; |
139 | } __packed entry; |
140 | |
141 | uint32_t cdw10; |
142 | uint32_t cdw11; |
143 | uint32_t cdw12; |
144 | uint32_t cdw13; |
145 | uint32_t cdw14; |
146 | uint32_t cdw15; |
147 | } __packed __aligned(8); |
148 | NVME_CTASSERT(sizeof(struct nvme_sqe) == 64, "bad size for nvme_sqe" ); |
149 | |
150 | struct nvme_sqe_q { |
151 | uint8_t opcode; |
152 | uint8_t flags; |
153 | uint16_t cid; |
154 | |
155 | uint8_t _reserved1[20]; |
156 | |
157 | uint64_t prp1; |
158 | |
159 | uint8_t _reserved2[8]; |
160 | |
161 | uint16_t qid; |
162 | uint16_t qsize; |
163 | |
164 | uint8_t qflags; |
165 | #define NVM_SQE_SQ_QPRIO_URG (0x0 << 1) |
166 | #define NVM_SQE_SQ_QPRIO_HI (0x1 << 1) |
167 | #define NVM_SQE_SQ_QPRIO_MED (0x2 << 1) |
168 | #define NVM_SQE_SQ_QPRIO_LOW (0x3 << 1) |
169 | #define NVM_SQE_CQ_IEN (1 << 1) |
170 | #define NVM_SQE_Q_PC (1 << 0) |
171 | uint8_t _reserved3; |
172 | uint16_t cqid; /* XXX interrupt vector for cq */ |
173 | |
174 | uint8_t _reserved4[16]; |
175 | } __packed __aligned(8); |
176 | NVME_CTASSERT(sizeof(struct nvme_sqe_q) == 64, "bad size for nvme_sqe_q" ); |
177 | |
178 | struct nvme_sqe_io { |
179 | uint8_t opcode; |
180 | uint8_t flags; |
181 | uint16_t cid; |
182 | |
183 | uint32_t nsid; |
184 | |
185 | uint8_t _reserved[8]; |
186 | |
187 | uint64_t mptr; |
188 | |
189 | union { |
190 | uint64_t prp[2]; |
191 | struct nvme_sge sge; |
192 | } __packed entry; |
193 | |
194 | uint64_t slba; /* Starting LBA */ |
195 | |
196 | uint16_t nlb; /* Number of Logical Blocks */ |
197 | uint16_t ioflags; |
198 | #define NVM_SQE_IO_LR __BIT(15) /* Limited Retry */ |
199 | #define NVM_SQE_IO_FUA __BIT(14) /* Force Unit Access (bypass cache) */ |
200 | |
201 | uint8_t dsm; /* Dataset Management */ |
202 | #define NVM_SQE_IO_INCOMP __BIT(7) /* Incompressible */ |
203 | #define NVM_SQE_IO_SEQ __BIT(6) /* Sequential request */ |
204 | #define NVM_SQE_IO_LAT_MASK __BITS(4, 5) /* Access Latency */ |
205 | #define NVM_SQE_IO_LAT_NONE 0 /* Latency: none */ |
206 | #define NVM_SQE_IO_LAT_IDLE __BIT(4) /* Latency: idle */ |
207 | #define NVM_SQE_IO_LAT_NORMAL __BIT(5) /* Latency: normal */ |
208 | #define NVM_SQE_IO_LAT_LOW __BITS(4, 5) /* Latency: low */ |
209 | #define NVM_SQE_IO_FREQ_MASK __BITS(0, 3) /* Access Frequency */ |
210 | #define NVM_SQE_IO_FREQ_TYPICAL 0x1 /* Typical */ |
211 | #define NVM_SQE_IO_FREQ_INFR_INFW 0x2 /* Infrequent read and writes */ |
212 | #define NVM_SQE_IO_FREQ_FRR_INFW 0x3 /* Frequent read, inf. writes */ |
213 | #define NVM_SQE_IO_FREQ_INFR_FRW 0x4 /* Inf. read, freq. writes */ |
214 | #define NVM_SQE_IO_FREQ_FRR_FRW 0x5 /* Freq. read and writes */ |
215 | #define NVM_SQE_IO_FREQ_ONCE 0x6 /* One time i/o operation */ |
216 | /* Extra Access Frequency bits for read operations */ |
217 | #define NVM_SQE_IO_FREQ_SPEC 0x7 /* Speculative read - prefech */ |
218 | #define NVM_SQE_IO_FREQ_OVERWRITE 0x8 /* Will be overwritten soon */ |
219 | uint8_t _reserved2[3]; |
220 | |
221 | uint32_t eilbrt; /* Expected Initial Logical Block |
222 | Reference Tag */ |
223 | |
224 | uint16_t elbat; /* Expected Logical Block |
225 | Application Tag */ |
226 | uint16_t elbatm; /* Expected Logical Block |
227 | Application Tag Mask */ |
228 | } __packed __aligned(8); |
229 | NVME_CTASSERT(sizeof(struct nvme_sqe_io) == 64, "bad size for nvme_sqe_io" ); |
230 | |
231 | struct nvme_cqe { |
232 | uint32_t cdw0; |
233 | |
234 | uint32_t _reserved; |
235 | |
236 | uint16_t sqhd; /* SQ Head Pointer */ |
237 | uint16_t sqid; /* SQ Identifier */ |
238 | |
239 | uint16_t cid; /* Command Identifier */ |
240 | uint16_t flags; |
241 | #define NVME_CQE_DNR __BIT(15) |
242 | #define NVME_CQE_M __BIT(14) |
243 | #define NVME_CQE_SCT_MASK __BITS(8, 10) |
244 | #define NVME_CQE_SCT(_f) ((_f) & (0x07 << 8)) |
245 | #define NVME_CQE_SCT_GENERIC (0x00 << 8) |
246 | #define NVME_CQE_SCT_COMMAND (0x01 << 8) |
247 | #define NVME_CQE_SCT_MEDIAERR (0x02 << 8) |
248 | #define NVME_CQE_SCT_VENDOR (0x07 << 8) |
249 | #define NVME_CQE_SC_MASK __BITS(1, 7) |
250 | #define NVME_CQE_SC(_f) ((_f) & (0x7f << 1)) |
251 | /* generic command status codes */ |
252 | #define NVME_CQE_SC_SUCCESS (0x00 << 1) |
253 | #define NVME_CQE_SC_INVALID_OPCODE (0x01 << 1) |
254 | #define NVME_CQE_SC_INVALID_FIELD (0x02 << 1) |
255 | #define NVME_CQE_SC_CID_CONFLICT (0x03 << 1) |
256 | #define NVME_CQE_SC_DATA_XFER_ERR (0x04 << 1) |
257 | #define NVME_CQE_SC_ABRT_BY_NO_PWR (0x05 << 1) |
258 | #define NVME_CQE_SC_INTERNAL_DEV_ERR (0x06 << 1) |
259 | #define NVME_CQE_SC_CMD_ABRT_REQD (0x07 << 1) |
260 | #define NVME_CQE_SC_CMD_ABDR_SQ_DEL (0x08 << 1) |
261 | #define NVME_CQE_SC_CMD_ABDR_FUSE_ERR (0x09 << 1) |
262 | #define NVME_CQE_SC_CMD_ABDR_FUSE_MISS (0x0a << 1) |
263 | #define NVME_CQE_SC_INVALID_NS (0x0b << 1) |
264 | #define NVME_CQE_SC_CMD_SEQ_ERR (0x0c << 1) |
265 | #define NVME_CQE_SC_INVALID_LAST_SGL (0x0d << 1) |
266 | #define NVME_CQE_SC_INVALID_NUM_SGL (0x0e << 1) |
267 | #define NVME_CQE_SC_DATA_SGL_LEN (0x0f << 1) |
268 | #define NVME_CQE_SC_MDATA_SGL_LEN (0x10 << 1) |
269 | #define NVME_CQE_SC_SGL_TYPE_INVALID (0x11 << 1) |
270 | #define NVME_CQE_SC_LBA_RANGE (0x80 << 1) |
271 | #define NVME_CQE_SC_CAP_EXCEEDED (0x81 << 1) |
272 | #define NVME_CQE_SC_NS_NOT_RDY (0x82 << 1) |
273 | #define NVME_CQE_SC_RSV_CONFLICT (0x83 << 1) |
274 | /* command specific status codes */ |
275 | #define NVME_CQE_SC_CQE_INVALID (0x00 << 1) |
276 | #define NVME_CQE_SC_INVALID_QID (0x01 << 1) |
277 | #define NVME_CQE_SC_MAX_Q_SIZE (0x02 << 1) |
278 | #define NVME_CQE_SC_ABORT_LIMIT (0x03 << 1) |
279 | #define NVME_CQE_SC_ASYNC_EV_REQ_LIMIT (0x05 << 1) |
280 | #define NVME_CQE_SC_INVALID_FW_SLOT (0x06 << 1) |
281 | #define NVME_CQE_SC_INVALID_FW_IMAGE (0x07 << 1) |
282 | #define NVME_CQE_SC_INVALID_INT_VEC (0x08 << 1) |
283 | #define NVME_CQE_SC_INVALID_LOG_PAGE (0x09 << 1) |
284 | #define NVME_CQE_SC_INVALID_FORMAT (0x0a << 1) |
285 | #define NVME_CQE_SC_FW_REQ_CNV_RESET (0x0b << 1) |
286 | #define NVME_CQE_SC_FW_REQ_NVM_RESET (0x10 << 1) |
287 | #define NVME_CQE_SC_FW_REQ_RESET (0x11 << 1) |
288 | #define NVME_CQE_SC_FW_MAX_TIME_VIO (0x12 << 1) |
289 | #define NVME_CQE_SC_FW_PROHIBIT (0x13 << 1) |
290 | #define NVME_CQE_SC_OVERLAP_RANGE (0x14 << 1) |
291 | #define NVME_CQE_SC_CONFLICT_ATTRS (0x80 << 1) |
292 | #define NVME_CQE_SC_INVALID_PROT_INFO (0x81 << 1) |
293 | #define NVME_CQE_SC_ATT_WR_TO_RO_PAGE (0x82 << 1) |
294 | /* media error status codes */ |
295 | #define NVME_CQE_SC_WRITE_FAULTS (0x80 << 1) |
296 | #define NVME_CQE_SC_UNRECV_READ_ERR (0x81 << 1) |
297 | #define NVME_CQE_SC_GUARD_CHECK_ERR (0x82 << 1) |
298 | #define NVME_CQE_SC_APPL_TAG_CHECK_ERR (0x83 << 1) |
299 | #define NVME_CQE_SC_REF_TAG_CHECK_ERR (0x84 << 1) |
300 | #define NVME_CQE_SC_CMP_FAIL (0x85 << 1) |
301 | #define NVME_CQE_SC_ACCESS_DENIED (0x86 << 1) |
302 | #define NVME_CQE_PHASE __BIT(0) |
303 | } __packed __aligned(8); |
304 | NVME_CTASSERT(sizeof(struct nvme_cqe) == 16, "bad size for nvme_cqe" ); |
305 | |
306 | #define NVM_ADMIN_DEL_IOSQ 0x00 /* Delete I/O Submission Queue */ |
307 | #define NVM_ADMIN_ADD_IOSQ 0x01 /* Create I/O Submission Queue */ |
308 | #define NVM_ADMIN_GET_LOG_PG 0x02 /* Get Log Page */ |
309 | #define NVM_ADMIN_DEL_IOCQ 0x04 /* Delete I/O Completion Queue */ |
310 | #define NVM_ADMIN_ADD_IOCQ 0x05 /* Create I/O Completion Queue */ |
311 | #define NVM_ADMIN_IDENTIFY 0x06 /* Identify */ |
312 | #define NVM_ADMIN_ABORT 0x08 /* Abort */ |
313 | #define NVM_ADMIN_SET_FEATURES 0x09 /* Set Features */ |
314 | #define NVM_ADMIN_GET_FEATURES 0x0a /* Get Features */ |
315 | #define NVM_ADMIN_ASYNC_EV_REQ 0x0c /* Asynchronous Event Request */ |
316 | #define NVM_ADMIN_NS_MANAGEMENT 0x0d /* Namespace Management */ |
317 | /* 0x0e-0x0f - reserved */ |
318 | #define NVM_ADMIN_FW_COMMIT 0x10 /* Firmware Commit */ |
319 | #define NVM_ADMIN_FW_DOWNLOAD 0x11 /* Firmware Image Download */ |
320 | #define NVM_ADMIN_DEV_SELFTEST 0x14 /* Device Self Test */ |
321 | #define NVM_ADMIN_NS_ATTACHMENT 0x15 /* Namespace Attachment */ |
322 | #define NVM_ADMIN_KEEP_ALIVE 0x18 /* Keep Alive */ |
323 | #define NVM_ADMIN_DIRECTIVE_SND 0x19 /* Derective Send */ |
324 | #define NVM_ADMIN_DIRECTIVE_RCV 0x1a /* Derective Receive */ |
325 | #define NVM_ADMIN_VIRT_MGMT 0x1c /* Virtualization Management */ |
326 | #define NVM_ADMIN_NVME_MI_SEND 0x1d /* NVMe-MI Send */ |
327 | #define NVM_ADMIN_NVME_MI_RECV 0x1e /* NVMe-MI Receive */ |
328 | #define NVM_ADMIN_DOORBELL_BC 0x7c /* Doorbell Buffer Config */ |
329 | #define NVM_ADMIN_FORMAT_NVM 0x80 /* Format NVM */ |
330 | #define NVM_ADMIN_SECURITY_SND 0x81 /* Security Send */ |
331 | #define NVM_ADMIN_SECURITY_RCV 0x82 /* Security Receive */ |
332 | #define NVM_ADMIN_SANITIZE 0x84 /* Sanitize */ |
333 | |
334 | #define NVM_CMD_FLUSH 0x00 /* Flush */ |
335 | #define NVM_CMD_WRITE 0x01 /* Write */ |
336 | #define NVM_CMD_READ 0x02 /* Read */ |
337 | #define NVM_CMD_WR_UNCOR 0x04 /* Write Uncorrectable */ |
338 | #define NVM_CMD_COMPARE 0x05 /* Compare */ |
339 | /* 0x06-0x07 - reserved */ |
340 | #define NVM_CMD_WRITE_ZEROES 0x08 /* Write Zeroes */ |
341 | #define NVM_CMD_DSM 0x09 /* Dataset Management */ |
342 | |
343 | /* Features for GET/SET FEATURES */ |
344 | /* 0x00 - reserved */ |
345 | #define NVM_FEAT_ARBITRATION 0x01 |
346 | #define NVM_FEAT_POWER_MANAGEMENT 0x02 |
347 | #define NVM_FEAT_LBA_RANGE_TYPE 0x03 |
348 | #define NVM_FEAT_TEMPERATURE_THRESHOLD 0x04 |
349 | #define NVM_FEAT_ERROR_RECOVERY 0x05 |
350 | #define NVM_FEATURE_VOLATILE_WRITE_CACHE 0x06 /* optional */ |
351 | #define NVM_FEATURE_NUMBER_OF_QUEUES 0x07 /* mandatory */ |
352 | #define NVM_FEAT_INTERRUPT_COALESCING 0x08 |
353 | #define NVM_FEAT_INTERRUPT_VECTOR_CONFIGURATION 0x09 |
354 | #define NVM_FEAT_WRITE_ATOMICITY 0x0a |
355 | #define NVM_FEAT_ASYNC_EVENT_CONFIGURATION 0x0b |
356 | #define NVM_FEAT_AUTONOMOUS_POWER_STATE_TRANSITION 0x0c |
357 | #define NVM_FEAT_HOST_MEMORY_BUFFER 0x0d |
358 | #define NVM_FEAT_TIMESTAMP 0x0e |
359 | #define NVM_FEAT_KEEP_ALIVE_TIMER 0x0f |
360 | #define NVM_FEAT_HOST_CONTROLLED_THERMAL_MGMT 0x10 |
361 | #define NVM_FEAT_NON_OP_POWER_STATE_CONFIG 0x11 |
362 | /* 0x12-0x77 - reserved */ |
363 | /* 0x78-0x7f - NVMe Management Interface */ |
364 | #define NVM_FEAT_SOFTWARE_PROGRESS_MARKER 0x80 |
365 | #define NVM_FEAT_HOST_IDENTIFIER 0x81 |
366 | #define NVM_FEAT_RESERVATION_NOTIFICATION_MASK 0x82 |
367 | #define NVM_FEAT_RESERVATION_PERSISTANCE 0x83 |
368 | /* 0x84-0xBF - command set specific (reserved) */ |
369 | /* 0xC0-0xFF - vendor specific */ |
370 | |
371 | #define NVM_SET_FEATURES_SV __BIT(31) /* Persist */ |
372 | |
373 | #define NVM_VOLATILE_WRITE_CACHE_WCE __BIT(0) /* Write Cache Enable */ |
374 | |
375 | /* Power State Descriptor Data */ |
376 | struct nvm_identify_psd { |
377 | uint16_t mp; /* Max Power */ |
378 | uint8_t _reserved1; |
379 | uint8_t flags; |
380 | #define NVME_PSD_NOPS __BIT(1) |
381 | #define NVME_PSD_MPS __BIT(0) |
382 | |
383 | uint32_t enlat; /* Entry Latency */ |
384 | |
385 | uint32_t exlat; /* Exit Latency */ |
386 | |
387 | uint8_t rrt; /* Relative Read Throughput */ |
388 | #define NVME_PSD_RRT_MASK __BITS(0, 4) |
389 | uint8_t rrl; /* Relative Read Latency */ |
390 | #define NVME_PSD_RRL_MASK __BITS(0, 4) |
391 | uint8_t rwt; /* Relative Write Throughput */ |
392 | #define NVME_PSD_RWT_MASK __BITS(0, 4) |
393 | uint8_t rwl; /* Relative Write Latency */ |
394 | #define NVME_PSD_RWL_MASK __BITS(0, 4) |
395 | |
396 | uint16_t idlp; /* Idle Power */ |
397 | uint8_t ips; /* Idle Power Scale */ |
398 | #define NVME_PSD_IPS_MASK __BITS(0, 1) |
399 | uint8_t _reserved2; |
400 | uint16_t actp; /* Active Power */ |
401 | uint16_t ap; /* Active Power Workload/Scale */ |
402 | #define NVME_PSD_APW_MASK __BITS(0, 2) |
403 | #define NVME_PSD_APS_MASK __BITS(6, 7) |
404 | |
405 | uint8_t _reserved[8]; |
406 | } __packed __aligned(8); |
407 | NVME_CTASSERT(sizeof(struct nvm_identify_psd) == 32, "bad size for nvm_identify_psd" ); |
408 | |
409 | struct nvm_identify_controller { |
410 | /* Controller Capabilities and Features */ |
411 | |
412 | uint16_t vid; /* PCI Vendor ID */ |
413 | uint16_t ssvid; /* PCI Subsystem Vendor ID */ |
414 | |
415 | uint8_t sn[20]; /* Serial Number */ |
416 | uint8_t mn[40]; /* Model Number */ |
417 | uint8_t fr[8]; /* Firmware Revision */ |
418 | |
419 | uint8_t rab; /* Recommended Arbitration Burst */ |
420 | uint8_t ieee[3]; /* IEEE OUI Identifier */ |
421 | |
422 | uint8_t cmic; /* Controller Multi-Path I/O and |
423 | Namespace Sharing Capabilities */ |
424 | uint8_t mdts; /* Maximum Data Transfer Size */ |
425 | |
426 | uint16_t cntlid; /* Controller ID */ |
427 | uint32_t ver; /* Version */ |
428 | |
429 | uint32_t rtd3r; /* RTD3 Resume Latency */ |
430 | uint32_t rtd3e; /* RTD3 Enter Latency */ |
431 | |
432 | uint32_t oaes; /* Optional Asynchronous Events Supported */ |
433 | uint32_t ctrattr; /* Controller Attributes */ |
434 | |
435 | uint8_t _reserved1[12]; |
436 | |
437 | uint8_t fguid[16]; /* FRU Globally Unique Identifier */ |
438 | |
439 | uint8_t _reserved2[128]; |
440 | |
441 | /* Admin Command Set Attributes & Optional Controller Capabilities */ |
442 | |
443 | uint16_t oacs; /* Optional Admin Command Support */ |
444 | #define NVME_ID_CTRLR_OACS_DOORBELL_BC __BIT(8) |
445 | #define NVME_ID_CTRLR_OACS_VIRT_MGMT __BIT(7) |
446 | #define NVME_ID_CTRLR_OACS_NVME_MI __BIT(6) |
447 | #define NVME_ID_CTRLR_OACS_DIRECTIVES __BIT(5) |
448 | #define NVME_ID_CTRLR_OACS_DEV_SELFTEST __BIT(4) |
449 | #define NVME_ID_CTRLR_OACS_NS __BIT(3) |
450 | #define NVME_ID_CTRLR_OACS_FW __BIT(2) |
451 | #define NVME_ID_CTRLR_OACS_FORMAT __BIT(1) |
452 | #define NVME_ID_CTRLR_OACS_SECURITY __BIT(0) |
453 | uint8_t acl; /* Abort Command Limit */ |
454 | uint8_t aerl; /* Asynchronous Event Request Limit */ |
455 | |
456 | uint8_t frmw; /* Firmware Updates */ |
457 | #define NVME_ID_CTRLR_FRMW_NOREQ_RESET __BIT(4) |
458 | #define NVME_ID_CTRLR_FRMW_NSLOT __BITS(1, 3) |
459 | #define NVME_ID_CTRLR_FRMW_SLOT1_RO __BIT(0) |
460 | uint8_t lpa; /* Log Page Attributes */ |
461 | #define NVME_ID_CTRLR_LPA_CMD_EFFECT __BIT(1) |
462 | #define NVME_ID_CTRLR_LPA_NS_SMART __BIT(0) |
463 | uint8_t elpe; /* Error Log Page Entries */ |
464 | uint8_t npss; /* Number of Power States Support */ |
465 | |
466 | uint8_t avscc; /* Admin Vendor Specific Command |
467 | Configuration */ |
468 | uint8_t apsta; /* Autonomous Power State Transition |
469 | Attributes */ |
470 | |
471 | uint16_t wctemp; /* Warning Composite Temperature |
472 | Threshold */ |
473 | uint16_t cctemp; /* Critical Composite Temperature |
474 | Threshold */ |
475 | |
476 | uint16_t mtfa; /* Maximum Time for Firmware Activation */ |
477 | |
478 | uint32_t hmpre; /* Host Memory Buffer Preferred Size */ |
479 | uint32_t hmmin; /* Host Memory Buffer Minimum Size */ |
480 | |
481 | struct { |
482 | uint64_t tnvmcap[2]; |
483 | uint64_t unvmcap[2]; |
484 | } __packed untncap; /* Name space capabilities: |
485 | if NVME_ID_CTRLR_OACS_NS, |
486 | report tnvmcap and unvmcap */ |
487 | |
488 | uint32_t rpmbs; /* Replay Protected Memory Block Support */ |
489 | |
490 | uint16_t edstt; /* Extended Device Self-test Time */ |
491 | uint8_t dsto; /* Device Self-test Options */ |
492 | |
493 | uint8_t fwug; /* Firmware Update Granularity */ |
494 | |
495 | uint16_t kas; /* Keep Alive Support */ |
496 | |
497 | uint16_t hctma; /* Host Controlled Thermal Management |
498 | Attributes */ |
499 | uint16_t mntmt; /* Minimum Thermal Management Temperature */ |
500 | uint16_t mxtmt; /* Maximum Thermal Management Temperature */ |
501 | |
502 | uint32_t sanicap; /* Sanitize Capabilities */ |
503 | |
504 | uint8_t _reserved3[180]; |
505 | |
506 | /* NVM Command Set Attributes */ |
507 | |
508 | uint8_t sqes; /* Submission Queue Entry Size */ |
509 | #define NVME_ID_CTRLR_SQES_MAX __BITS(4, 7) |
510 | #define NVME_ID_CTRLR_SQES_MIN __BITS(0, 3) |
511 | uint8_t cqes; /* Completion Queue Entry Size */ |
512 | #define NVME_ID_CTRLR_CQES_MAX __BITS(4, 7) |
513 | #define NVME_ID_CTRLR_CQES_MIN __BITS(0, 3) |
514 | |
515 | uint16_t maxcmd; /* Maximum Outstanding Commands */ |
516 | |
517 | uint32_t nn; /* Number of Namespaces */ |
518 | |
519 | uint16_t oncs; /* Optional NVM Command Support */ |
520 | #define NVME_ID_CTRLR_ONCS_TIMESTAMP __BIT(6) |
521 | #define NVME_ID_CTRLR_ONCS_RESERVATION __BIT(5) |
522 | #define NVME_ID_CTRLR_ONCS_SET_FEATURES __BIT(4) |
523 | #define NVME_ID_CTRLR_ONCS_WRITE_ZERO __BIT(3) |
524 | #define NVME_ID_CTRLR_ONCS_DSM __BIT(2) |
525 | #define NVME_ID_CTRLR_ONCS_WRITE_UNC __BIT(1) |
526 | #define NVME_ID_CTRLR_ONCS_COMPARE __BIT(0) |
527 | uint16_t fuses; /* Fused Operation Support */ |
528 | |
529 | uint8_t fna; /* Format NVM Attributes */ |
530 | #define NVME_ID_CTRLR_FNA_CRYPTO_ERASE __BIT(2) |
531 | #define NVME_ID_CTRLR_FNA_ERASE_ALL __BIT(1) |
532 | #define NVME_ID_CTRLR_FNA_FORMAT_ALL __BIT(0) |
533 | uint8_t vwc; /* Volatile Write Cache */ |
534 | #define NVME_ID_CTRLR_VWC_PRESENT __BIT(0) |
535 | uint16_t awun; /* Atomic Write Unit Normal */ |
536 | uint16_t awupf; /* Atomic Write Unit Power Fail */ |
537 | |
538 | uint8_t nvscc; /* NVM Vendor Specific Command */ |
539 | uint8_t _reserved4[1]; |
540 | |
541 | uint16_t acwu; /* Atomic Compare & Write Unit */ |
542 | uint8_t _reserved5[2]; |
543 | |
544 | uint32_t sgls; /* SGL Support */ |
545 | |
546 | uint8_t _reserved6[228]; |
547 | |
548 | uint8_t subnqn[256]; /* NVM Subsystem NVMe Qualified Name */ |
549 | |
550 | uint8_t _reserved7[768]; |
551 | |
552 | uint8_t _reserved8[256]; /* NVMe over Fabrics specification */ |
553 | |
554 | struct nvm_identify_psd psd[32]; /* Power State Descriptors */ |
555 | |
556 | uint8_t vs[1024]; /* Vendor Specific */ |
557 | } __packed __aligned(8); |
558 | NVME_CTASSERT(sizeof(struct nvm_identify_controller) == 4096, "bad size for nvm_identify_controller" ); |
559 | |
560 | struct nvm_namespace_format { |
561 | uint16_t ms; /* Metadata Size */ |
562 | uint8_t lbads; /* LBA Data Size */ |
563 | uint8_t rp; /* Relative Performance */ |
564 | } __packed __aligned(4); |
565 | NVME_CTASSERT(sizeof(struct nvm_namespace_format) == 4, "bad size for nvm_namespace_format" ); |
566 | |
567 | struct nvm_identify_namespace { |
568 | uint64_t nsze; /* Namespace Size */ |
569 | |
570 | uint64_t ncap; /* Namespace Capacity */ |
571 | |
572 | uint64_t nuse; /* Namespace Utilization */ |
573 | |
574 | uint8_t nsfeat; /* Namespace Features */ |
575 | #define NVME_ID_NS_NSFEAT_LOGICAL_BLK_ERR __BIT(2) |
576 | #define NVME_ID_NS_NSFEAT_NS __BIT(1) |
577 | #define NVME_ID_NS_NSFEAT_THIN_PROV __BIT(0) |
578 | uint8_t nlbaf; /* Number of LBA Formats */ |
579 | uint8_t flbas; /* Formatted LBA Size */ |
580 | #define NVME_ID_NS_FLBAS(_f) ((_f) & 0x0f) |
581 | #define NVME_ID_NS_FLBAS_MD 0x10 |
582 | uint8_t mc; /* Metadata Capabilities */ |
583 | uint8_t dpc; /* End-to-end Data Protection |
584 | Capabilities */ |
585 | uint8_t dps; /* End-to-end Data Protection Type |
586 | Settings */ |
587 | #define NVME_ID_NS_DPS_MD_START __BIT(3) |
588 | #define NVME_ID_NS_DPS_PIT(_f) ((_f) & 0x7) |
589 | |
590 | uint8_t nmic; /* Namespace Multi-path I/O and Namespace |
591 | Sharing Capabilities */ |
592 | |
593 | uint8_t rescap; /* Reservation Capabilities */ |
594 | |
595 | uint8_t fpi; /* Format Progress Indicator */ |
596 | |
597 | uint8_t dlfeat; /* Deallocate Logical Block Features */ |
598 | |
599 | uint16_t nawun; /* Namespace Atomic Write Unit Normal */ |
600 | uint16_t nawupf; /* Namespace Atomic Write Unit Power Fail */ |
601 | uint16_t nacwu; /* Namespace Atomic Compare & Write Unit */ |
602 | uint16_t nabsn; /* Namespace Atomic Boundary Size Normal */ |
603 | uint16_t nabo; /* Namespace Atomic Boundary Offset */ |
604 | uint16_t nabspf; /* Namespace Atomic Boundary Size Power |
605 | Fail */ |
606 | uint16_t noiob; /* Namespace Optimal IO Boundary */ |
607 | |
608 | uint8_t nvmcap[16]; /* NVM Capacity */ |
609 | |
610 | uint8_t _reserved1[40]; /* bytes 64-103: Reserved */ |
611 | |
612 | uint8_t nguid[16]; /* Namespace Globally Unique Identifier */ |
613 | uint8_t eui64[8]; /* IEEE Extended Unique Identifier */ |
614 | |
615 | struct nvm_namespace_format |
616 | lbaf[16]; /* LBA Format Support */ |
617 | |
618 | uint8_t _reserved2[192]; |
619 | |
620 | uint8_t vs[3712]; |
621 | } __packed __aligned(8); |
622 | NVME_CTASSERT(sizeof(struct nvm_identify_namespace) == 4096, "bad size for nvm_identify_namespace" ); |
623 | |
624 | #endif /* __NVMEREG_H__ */ |
625 | |