| 1 | /*	$NetBSD: nvmereg.h,v 1.13 2018/12/01 18:25:39 jdolecek Exp $	*/ | 
| 2 | /*	$OpenBSD: nvmereg.h,v 1.10 2016/04/14 11:18:32 dlg Exp $ */ | 
| 3 |  | 
| 4 | /* | 
| 5 |  * Copyright (c) 2014 David Gwynne <dlg@openbsd.org> | 
| 6 |  * | 
| 7 |  * Permission to use, copy, modify, and distribute this software for any | 
| 8 |  * purpose with or without fee is hereby granted, provided that the above | 
| 9 |  * copyright notice and this permission notice appear in all copies. | 
| 10 |  * | 
| 11 |  * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES | 
| 12 |  * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF | 
| 13 |  * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR | 
| 14 |  * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES | 
| 15 |  * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN | 
| 16 |  * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF | 
| 17 |  * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. | 
| 18 |  */ | 
| 19 |  | 
| 20 | #ifndef	__NVMEREG_H__ | 
| 21 | #define	__NVMEREG_H__ | 
| 22 |  | 
| 23 | #ifndef	NVME_CTASSERT | 
| 24 | #define	NVME_CTASSERT(x, s)	__CTASSERT(x) | 
| 25 | #endif | 
| 26 |  | 
| 27 | #define NVME_CAP	0x0000	/* Controller Capabilities */ | 
| 28 | #define  NVME_CAP_MPSMAX(_r)	(12 + (((_r) >> 52) & 0xf)) /* shift */ | 
| 29 | #define  NVME_CAP_MPSMIN(_r)	(12 + (((_r) >> 48) & 0xf)) /* shift */ | 
| 30 | #define  NVME_CAP_CSS(_r)	(((_r) >> 37) & 0x7f) | 
| 31 | #define  NVME_CAP_CSS_NVM	__BIT(0) | 
| 32 | #define  NVME_CAP_NSSRS(_r)	ISSET((_r), __BIT(36)) | 
| 33 | #define  NVME_CAP_DSTRD(_r)	__BIT(2 + (((_r) >> 32) & 0xf)) /* bytes */ | 
| 34 | #define  NVME_CAP_TO(_r)	(500 * (((_r) >> 24) & 0xff)) /* ms */ | 
| 35 | #define  NVME_CAP_AMS(_r)	(((_r) >> 17) & 0x3) | 
| 36 | #define  NVME_CAP_AMS_WRR	__BIT(0) | 
| 37 | #define  NVME_CAP_AMS_VENDOR	__BIT(1) | 
| 38 | #define  NVME_CAP_CQR(_r)	ISSET((_r), __BIT(16)) | 
| 39 | #define  NVME_CAP_MQES(_r)	(((_r) & 0xffff) + 1) | 
| 40 | #define NVME_CAP_LO	0x0000 | 
| 41 | #define NVME_CAP_HI	0x0004 | 
| 42 | #define NVME_VS		0x0008	/* Version */ | 
| 43 | #define  NVME_VS_MJR(_r)	(((_r) >> 16) & 0xffff) | 
| 44 | #define  NVME_VS_MNR(_r)	(((_r) >> 8) & 0xff) | 
| 45 | #define  NVME_VS_TER(_r)	((_r) & 0xff) | 
| 46 | #define NVME_INTMS	0x000c	/* Interrupt Mask Set */ | 
| 47 | #define NVME_INTMC	0x0010	/* Interrupt Mask Clear */ | 
| 48 | #define NVME_CC		0x0014	/* Controller Configuration */ | 
| 49 | #define  NVME_CC_IOCQES(_v)	(((_v) & 0xf) << 20) | 
| 50 | #define  NVME_CC_IOCQES_MASK	NVME_CC_IOCQES(0xf) | 
| 51 | #define  NVME_CC_IOCQES_R(_v)	(((_v) >> 20) & 0xf) | 
| 52 | #define  NVME_CC_IOSQES(_v)	(((_v) & 0xf) << 16) | 
| 53 | #define  NVME_CC_IOSQES_MASK	NVME_CC_IOSQES(0xf) | 
| 54 | #define  NVME_CC_IOSQES_R(_v)	(((_v) >> 16) & 0xf) | 
| 55 | #define  NVME_CC_SHN(_v)	(((_v) & 0x3) << 14) | 
| 56 | #define  NVME_CC_SHN_MASK	NVME_CC_SHN(0x3) | 
| 57 | #define  NVME_CC_SHN_R(_v)	(((_v) >> 15) & 0x3) | 
| 58 | #define  NVME_CC_SHN_NONE	0 | 
| 59 | #define  NVME_CC_SHN_NORMAL	1 | 
| 60 | #define  NVME_CC_SHN_ABRUPT	2 | 
| 61 | #define  NVME_CC_AMS(_v)	(((_v) & 0x7) << 11) | 
| 62 | #define  NVME_CC_AMS_MASK	NVME_CC_AMS(0x7) | 
| 63 | #define  NVME_CC_AMS_R(_v)	(((_v) >> 11) & 0xf) | 
| 64 | #define  NVME_CC_AMS_RR		0 /* round-robin */ | 
| 65 | #define  NVME_CC_AMS_WRR_U	1 /* weighted round-robin w/ urgent */ | 
| 66 | #define  NVME_CC_AMS_VENDOR	7 /* vendor */ | 
| 67 | #define  NVME_CC_MPS(_v)	((((_v) - 12) & 0xf) << 7) | 
| 68 | #define  NVME_CC_MPS_MASK	(0xf << 7) | 
| 69 | #define  NVME_CC_MPS_R(_v)	(12 + (((_v) >> 7) & 0xf)) | 
| 70 | #define  NVME_CC_CSS(_v)	(((_v) & 0x7) << 4) | 
| 71 | #define  NVME_CC_CSS_MASK	NVME_CC_CSS(0x7) | 
| 72 | #define  NVME_CC_CSS_R(_v)	(((_v) >> 4) & 0x7) | 
| 73 | #define  NVME_CC_CSS_NVM	0 | 
| 74 | #define  NVME_CC_EN		__BIT(0) | 
| 75 | #define NVME_CSTS	0x001c	/* Controller Status */ | 
| 76 | #define  NVME_CSTS_SHST_MASK	(0x3 << 2) | 
| 77 | #define  NVME_CSTS_SHST_NONE	(0x0 << 2) /* normal operation */ | 
| 78 | #define  NVME_CSTS_SHST_WAIT	(0x1 << 2) /* shutdown processing occurring */ | 
| 79 | #define  NVME_CSTS_SHST_DONE	(0x2 << 2) /* shutdown processing complete */ | 
| 80 | #define  NVME_CSTS_CFS		(1 << 1) | 
| 81 | #define  NVME_CSTS_RDY		(1 << 0) | 
| 82 | #define NVME_NSSR	0x0020	/* NVM Subsystem Reset (Optional) */ | 
| 83 | #define NVME_AQA	0x0024	/* Admin Queue Attributes */ | 
| 84 | 				/* Admin Completion Queue Size */ | 
| 85 | #define  NVME_AQA_ACQS(_v)	(((_v) - 1) << 16) | 
| 86 | #define  NVME_AQA_ACQS_R(_v)	((_v >> 16) & ((1 << 12) - 1)) | 
| 87 | 				/* Admin Submission Queue Size */ | 
| 88 | #define  NVME_AQA_ASQS(_v)	(((_v) - 1) << 0) | 
| 89 | #define  NVME_AQA_ASQS_R(_v)	(_v & ((1 << 12) - 1)) | 
| 90 | #define NVME_ASQ	0x0028	/* Admin Submission Queue Base Address */ | 
| 91 | #define NVME_ACQ	0x0030	/* Admin Completion Queue Base Address */ | 
| 92 |  | 
| 93 | #define NVME_ADMIN_Q		0 | 
| 94 | /* Submission Queue Tail Doorbell */ | 
| 95 | #define NVME_SQTDBL(_q, _s)	(0x1000 + (2 * (_q) + 0) * (_s)) | 
| 96 | /* Completion Queue Head Doorbell */ | 
| 97 | #define NVME_CQHDBL(_q, _s)	(0x1000 + (2 * (_q) + 1) * (_s)) | 
| 98 |  | 
| 99 | struct nvme_sge { | 
| 100 | 	uint8_t		id; | 
| 101 | 	uint8_t		_reserved[15]; | 
| 102 | } __packed __aligned(8); | 
| 103 | NVME_CTASSERT(sizeof(struct nvme_sge) == 16, "bad size for nvme_sge" ); | 
| 104 |  | 
| 105 | struct nvme_sge_data { | 
| 106 | 	uint8_t		id; | 
| 107 | 	uint8_t		_reserved[3]; | 
| 108 |  | 
| 109 | 	uint32_t	length; | 
| 110 |  | 
| 111 | 	uint64_t	address; | 
| 112 | } __packed __aligned(8); | 
| 113 | NVME_CTASSERT(sizeof(struct nvme_sge_data) == 16, "bad size for nvme_sge_data" ); | 
| 114 |  | 
| 115 | struct nvme_sge_bit_bucket { | 
| 116 | 	uint8_t		id; | 
| 117 | 	uint8_t		_reserved[3]; | 
| 118 |  | 
| 119 | 	uint32_t	length; | 
| 120 |  | 
| 121 | 	uint64_t	address; | 
| 122 | } __packed __aligned(8); | 
| 123 | NVME_CTASSERT(sizeof(struct nvme_sge_bit_bucket) == 16, "bad size for nvme_sge_bit_bucket" ); | 
| 124 |  | 
| 125 | struct nvme_sqe { | 
| 126 | 	uint8_t		opcode; | 
| 127 | 	uint8_t		flags; | 
| 128 | 	uint16_t	cid; | 
| 129 |  | 
| 130 | 	uint32_t	nsid; | 
| 131 |  | 
| 132 | 	uint8_t		_reserved[8]; | 
| 133 |  | 
| 134 | 	uint64_t	mptr; | 
| 135 |  | 
| 136 | 	union { | 
| 137 | 		uint64_t	prp[2]; | 
| 138 | 		struct nvme_sge	sge; | 
| 139 | 	} __packed	entry; | 
| 140 |  | 
| 141 | 	uint32_t	cdw10; | 
| 142 | 	uint32_t	cdw11; | 
| 143 | 	uint32_t	cdw12; | 
| 144 | 	uint32_t	cdw13; | 
| 145 | 	uint32_t	cdw14; | 
| 146 | 	uint32_t	cdw15; | 
| 147 | } __packed __aligned(8); | 
| 148 | NVME_CTASSERT(sizeof(struct nvme_sqe) == 64, "bad size for nvme_sqe" ); | 
| 149 |  | 
| 150 | struct nvme_sqe_q { | 
| 151 | 	uint8_t		opcode; | 
| 152 | 	uint8_t		flags; | 
| 153 | 	uint16_t	cid; | 
| 154 |  | 
| 155 | 	uint8_t		_reserved1[20]; | 
| 156 |  | 
| 157 | 	uint64_t	prp1; | 
| 158 |  | 
| 159 | 	uint8_t		_reserved2[8]; | 
| 160 |  | 
| 161 | 	uint16_t	qid; | 
| 162 | 	uint16_t	qsize; | 
| 163 |  | 
| 164 | 	uint8_t		qflags; | 
| 165 | #define NVM_SQE_SQ_QPRIO_URG	(0x0 << 1) | 
| 166 | #define NVM_SQE_SQ_QPRIO_HI	(0x1 << 1) | 
| 167 | #define NVM_SQE_SQ_QPRIO_MED	(0x2 << 1) | 
| 168 | #define NVM_SQE_SQ_QPRIO_LOW	(0x3 << 1) | 
| 169 | #define NVM_SQE_CQ_IEN		(1 << 1) | 
| 170 | #define NVM_SQE_Q_PC		(1 << 0) | 
| 171 | 	uint8_t		_reserved3; | 
| 172 | 	uint16_t	cqid; /* XXX interrupt vector for cq */ | 
| 173 |  | 
| 174 | 	uint8_t		_reserved4[16]; | 
| 175 | } __packed __aligned(8); | 
| 176 | NVME_CTASSERT(sizeof(struct nvme_sqe_q) == 64, "bad size for nvme_sqe_q" ); | 
| 177 |  | 
| 178 | struct nvme_sqe_io { | 
| 179 | 	uint8_t		opcode; | 
| 180 | 	uint8_t		flags; | 
| 181 | 	uint16_t	cid; | 
| 182 |  | 
| 183 | 	uint32_t	nsid; | 
| 184 |  | 
| 185 | 	uint8_t		_reserved[8]; | 
| 186 |  | 
| 187 | 	uint64_t	mptr; | 
| 188 |  | 
| 189 | 	union { | 
| 190 | 		uint64_t	prp[2]; | 
| 191 | 		struct nvme_sge	sge; | 
| 192 | 	} __packed	entry; | 
| 193 |  | 
| 194 | 	uint64_t	slba;	/* Starting LBA */ | 
| 195 |  | 
| 196 | 	uint16_t	nlb;	/* Number of Logical Blocks */ | 
| 197 | 	uint16_t	ioflags; | 
| 198 | #define NVM_SQE_IO_LR	__BIT(15)	/* Limited Retry */ | 
| 199 | #define NVM_SQE_IO_FUA	__BIT(14)	/* Force Unit Access (bypass cache) */ | 
| 200 |  | 
| 201 | 	uint8_t		dsm;	/* Dataset Management */ | 
| 202 | #define NVM_SQE_IO_INCOMP	__BIT(7)	/* Incompressible */ | 
| 203 | #define NVM_SQE_IO_SEQ		__BIT(6)	/* Sequential request */ | 
| 204 | #define NVM_SQE_IO_LAT_MASK	__BITS(4, 5)	/* Access Latency */ | 
| 205 | #define  NVM_SQE_IO_LAT_NONE	0		/* Latency: none */ | 
| 206 | #define  NVM_SQE_IO_LAT_IDLE	__BIT(4)	/* Latency: idle */ | 
| 207 | #define  NVM_SQE_IO_LAT_NORMAL	__BIT(5)	/* Latency: normal */ | 
| 208 | #define  NVM_SQE_IO_LAT_LOW	__BITS(4, 5)	/* Latency: low */ | 
| 209 | #define NVM_SQE_IO_FREQ_MASK	__BITS(0, 3)	/* Access Frequency */ | 
| 210 | #define  NVM_SQE_IO_FREQ_TYPICAL	0x1	/* Typical */ | 
| 211 | #define  NVM_SQE_IO_FREQ_INFR_INFW	0x2	/* Infrequent read and writes */ | 
| 212 | #define  NVM_SQE_IO_FREQ_FRR_INFW	0x3	/* Frequent read, inf. writes */ | 
| 213 | #define  NVM_SQE_IO_FREQ_INFR_FRW	0x4	/* Inf. read, freq. writes */ | 
| 214 | #define  NVM_SQE_IO_FREQ_FRR_FRW	0x5	/* Freq. read and writes */ | 
| 215 | #define  NVM_SQE_IO_FREQ_ONCE		0x6	/* One time i/o operation */ | 
| 216 | /* Extra Access Frequency bits for read operations */ | 
| 217 | #define  NVM_SQE_IO_FREQ_SPEC		0x7	/* Speculative read - prefech */ | 
| 218 | #define  NVM_SQE_IO_FREQ_OVERWRITE	0x8	/* Will be overwritten soon */ | 
| 219 | 	uint8_t		_reserved2[3]; | 
| 220 |  | 
| 221 | 	uint32_t	eilbrt;	/* Expected Initial Logical Block | 
| 222 | 				   Reference Tag */ | 
| 223 |  | 
| 224 | 	uint16_t	elbat;	/* Expected Logical Block | 
| 225 | 				   Application Tag */ | 
| 226 | 	uint16_t	elbatm;	/* Expected Logical Block | 
| 227 | 				   Application Tag Mask */ | 
| 228 | } __packed __aligned(8); | 
| 229 | NVME_CTASSERT(sizeof(struct nvme_sqe_io) == 64, "bad size for nvme_sqe_io" ); | 
| 230 |  | 
| 231 | struct nvme_cqe { | 
| 232 | 	uint32_t	cdw0; | 
| 233 |  | 
| 234 | 	uint32_t	_reserved; | 
| 235 |  | 
| 236 | 	uint16_t	sqhd; /* SQ Head Pointer */ | 
| 237 | 	uint16_t	sqid; /* SQ Identifier */ | 
| 238 |  | 
| 239 | 	uint16_t	cid; /* Command Identifier */ | 
| 240 | 	uint16_t	flags; | 
| 241 | #define NVME_CQE_DNR		__BIT(15) | 
| 242 | #define NVME_CQE_M		__BIT(14) | 
| 243 | #define NVME_CQE_SCT_MASK	__BITS(8, 10) | 
| 244 | #define NVME_CQE_SCT(_f)	((_f) & (0x07 << 8)) | 
| 245 | #define  NVME_CQE_SCT_GENERIC		(0x00 << 8) | 
| 246 | #define  NVME_CQE_SCT_COMMAND		(0x01 << 8) | 
| 247 | #define  NVME_CQE_SCT_MEDIAERR		(0x02 << 8) | 
| 248 | #define  NVME_CQE_SCT_VENDOR		(0x07 << 8) | 
| 249 | #define NVME_CQE_SC_MASK	__BITS(1, 7) | 
| 250 | #define NVME_CQE_SC(_f)		((_f) & (0x7f << 1)) | 
| 251 | /* generic command status codes */ | 
| 252 | #define  NVME_CQE_SC_SUCCESS		(0x00 << 1) | 
| 253 | #define  NVME_CQE_SC_INVALID_OPCODE	(0x01 << 1) | 
| 254 | #define  NVME_CQE_SC_INVALID_FIELD	(0x02 << 1) | 
| 255 | #define  NVME_CQE_SC_CID_CONFLICT	(0x03 << 1) | 
| 256 | #define  NVME_CQE_SC_DATA_XFER_ERR	(0x04 << 1) | 
| 257 | #define  NVME_CQE_SC_ABRT_BY_NO_PWR	(0x05 << 1) | 
| 258 | #define  NVME_CQE_SC_INTERNAL_DEV_ERR	(0x06 << 1) | 
| 259 | #define  NVME_CQE_SC_CMD_ABRT_REQD	(0x07 << 1) | 
| 260 | #define  NVME_CQE_SC_CMD_ABDR_SQ_DEL	(0x08 << 1) | 
| 261 | #define  NVME_CQE_SC_CMD_ABDR_FUSE_ERR	(0x09 << 1) | 
| 262 | #define  NVME_CQE_SC_CMD_ABDR_FUSE_MISS	(0x0a << 1) | 
| 263 | #define  NVME_CQE_SC_INVALID_NS		(0x0b << 1) | 
| 264 | #define  NVME_CQE_SC_CMD_SEQ_ERR	(0x0c << 1) | 
| 265 | #define  NVME_CQE_SC_INVALID_LAST_SGL	(0x0d << 1) | 
| 266 | #define  NVME_CQE_SC_INVALID_NUM_SGL	(0x0e << 1) | 
| 267 | #define  NVME_CQE_SC_DATA_SGL_LEN	(0x0f << 1) | 
| 268 | #define  NVME_CQE_SC_MDATA_SGL_LEN	(0x10 << 1) | 
| 269 | #define  NVME_CQE_SC_SGL_TYPE_INVALID	(0x11 << 1) | 
| 270 | #define  NVME_CQE_SC_LBA_RANGE		(0x80 << 1) | 
| 271 | #define  NVME_CQE_SC_CAP_EXCEEDED	(0x81 << 1) | 
| 272 | #define  NVME_CQE_SC_NS_NOT_RDY		(0x82 << 1) | 
| 273 | #define  NVME_CQE_SC_RSV_CONFLICT	(0x83 << 1) | 
| 274 | /* command specific status codes */ | 
| 275 | #define  NVME_CQE_SC_CQE_INVALID	(0x00 << 1) | 
| 276 | #define  NVME_CQE_SC_INVALID_QID	(0x01 << 1) | 
| 277 | #define  NVME_CQE_SC_MAX_Q_SIZE		(0x02 << 1) | 
| 278 | #define  NVME_CQE_SC_ABORT_LIMIT	(0x03 << 1) | 
| 279 | #define  NVME_CQE_SC_ASYNC_EV_REQ_LIMIT	(0x05 << 1) | 
| 280 | #define  NVME_CQE_SC_INVALID_FW_SLOT	(0x06 << 1) | 
| 281 | #define  NVME_CQE_SC_INVALID_FW_IMAGE	(0x07 << 1) | 
| 282 | #define  NVME_CQE_SC_INVALID_INT_VEC	(0x08 << 1) | 
| 283 | #define  NVME_CQE_SC_INVALID_LOG_PAGE	(0x09 << 1) | 
| 284 | #define  NVME_CQE_SC_INVALID_FORMAT	(0x0a << 1) | 
| 285 | #define  NVME_CQE_SC_FW_REQ_CNV_RESET	(0x0b << 1) | 
| 286 | #define  NVME_CQE_SC_FW_REQ_NVM_RESET	(0x10 << 1) | 
| 287 | #define  NVME_CQE_SC_FW_REQ_RESET	(0x11 << 1) | 
| 288 | #define  NVME_CQE_SC_FW_MAX_TIME_VIO	(0x12 << 1) | 
| 289 | #define  NVME_CQE_SC_FW_PROHIBIT	(0x13 << 1) | 
| 290 | #define  NVME_CQE_SC_OVERLAP_RANGE	(0x14 << 1) | 
| 291 | #define  NVME_CQE_SC_CONFLICT_ATTRS	(0x80 << 1) | 
| 292 | #define  NVME_CQE_SC_INVALID_PROT_INFO	(0x81 << 1) | 
| 293 | #define  NVME_CQE_SC_ATT_WR_TO_RO_PAGE	(0x82 << 1) | 
| 294 | /* media error status codes */ | 
| 295 | #define  NVME_CQE_SC_WRITE_FAULTS	(0x80 << 1) | 
| 296 | #define  NVME_CQE_SC_UNRECV_READ_ERR	(0x81 << 1) | 
| 297 | #define  NVME_CQE_SC_GUARD_CHECK_ERR	(0x82 << 1) | 
| 298 | #define  NVME_CQE_SC_APPL_TAG_CHECK_ERR	(0x83 << 1) | 
| 299 | #define  NVME_CQE_SC_REF_TAG_CHECK_ERR	(0x84 << 1) | 
| 300 | #define  NVME_CQE_SC_CMP_FAIL		(0x85 << 1) | 
| 301 | #define  NVME_CQE_SC_ACCESS_DENIED	(0x86 << 1) | 
| 302 | #define NVME_CQE_PHASE		__BIT(0) | 
| 303 | } __packed __aligned(8); | 
| 304 | NVME_CTASSERT(sizeof(struct nvme_cqe) == 16, "bad size for nvme_cqe" ); | 
| 305 |  | 
| 306 | #define NVM_ADMIN_DEL_IOSQ	0x00 /* Delete I/O Submission Queue */ | 
| 307 | #define NVM_ADMIN_ADD_IOSQ	0x01 /* Create I/O Submission Queue */ | 
| 308 | #define NVM_ADMIN_GET_LOG_PG	0x02 /* Get Log Page */ | 
| 309 | #define NVM_ADMIN_DEL_IOCQ	0x04 /* Delete I/O Completion Queue */ | 
| 310 | #define NVM_ADMIN_ADD_IOCQ	0x05 /* Create I/O Completion Queue */ | 
| 311 | #define NVM_ADMIN_IDENTIFY	0x06 /* Identify */ | 
| 312 | #define NVM_ADMIN_ABORT		0x08 /* Abort */ | 
| 313 | #define NVM_ADMIN_SET_FEATURES	0x09 /* Set Features */ | 
| 314 | #define NVM_ADMIN_GET_FEATURES	0x0a /* Get Features */ | 
| 315 | #define NVM_ADMIN_ASYNC_EV_REQ	0x0c /* Asynchronous Event Request */ | 
| 316 | #define NVM_ADMIN_NS_MANAGEMENT	0x0d /* Namespace Management */ | 
| 317 | /* 0x0e-0x0f - reserved */ | 
| 318 | #define NVM_ADMIN_FW_COMMIT	0x10 /* Firmware Commit */ | 
| 319 | #define NVM_ADMIN_FW_DOWNLOAD	0x11 /* Firmware Image Download */ | 
| 320 | #define NVM_ADMIN_DEV_SELFTEST	0x14 /* Device Self Test */ | 
| 321 | #define NVM_ADMIN_NS_ATTACHMENT	0x15 /* Namespace Attachment */ | 
| 322 | #define NVM_ADMIN_KEEP_ALIVE	0x18 /* Keep Alive */ | 
| 323 | #define NVM_ADMIN_DIRECTIVE_SND	0x19 /* Derective Send */ | 
| 324 | #define NVM_ADMIN_DIRECTIVE_RCV	0x1a /* Derective Receive */ | 
| 325 | #define NVM_ADMIN_VIRT_MGMT	0x1c /* Virtualization Management */ | 
| 326 | #define NVM_ADMIN_NVME_MI_SEND	0x1d /* NVMe-MI Send */ | 
| 327 | #define NVM_ADMIN_NVME_MI_RECV	0x1e /* NVMe-MI Receive */ | 
| 328 | #define NVM_ADMIN_DOORBELL_BC	0x7c /* Doorbell Buffer Config */ | 
| 329 | #define NVM_ADMIN_FORMAT_NVM	0x80 /* Format NVM */ | 
| 330 | #define NVM_ADMIN_SECURITY_SND	0x81 /* Security Send */ | 
| 331 | #define NVM_ADMIN_SECURITY_RCV	0x82 /* Security Receive */ | 
| 332 | #define NVM_ADMIN_SANITIZE	0x84 /* Sanitize */ | 
| 333 |  | 
| 334 | #define NVM_CMD_FLUSH		0x00 /* Flush */ | 
| 335 | #define NVM_CMD_WRITE		0x01 /* Write */ | 
| 336 | #define NVM_CMD_READ		0x02 /* Read */ | 
| 337 | #define NVM_CMD_WR_UNCOR	0x04 /* Write Uncorrectable */ | 
| 338 | #define NVM_CMD_COMPARE		0x05 /* Compare */ | 
| 339 | /* 0x06-0x07 - reserved */ | 
| 340 | #define NVM_CMD_WRITE_ZEROES	0x08 /* Write Zeroes */ | 
| 341 | #define NVM_CMD_DSM		0x09 /* Dataset Management */ | 
| 342 |  | 
| 343 | /* Features for GET/SET FEATURES */ | 
| 344 | /* 0x00 - reserved */ | 
| 345 | #define NVM_FEAT_ARBITRATION			0x01 | 
| 346 | #define NVM_FEAT_POWER_MANAGEMENT		0x02 | 
| 347 | #define NVM_FEAT_LBA_RANGE_TYPE			0x03 | 
| 348 | #define NVM_FEAT_TEMPERATURE_THRESHOLD		0x04 | 
| 349 | #define NVM_FEAT_ERROR_RECOVERY			0x05 | 
| 350 | #define NVM_FEATURE_VOLATILE_WRITE_CACHE	0x06	/* optional */ | 
| 351 | #define NVM_FEATURE_NUMBER_OF_QUEUES		0x07	/* mandatory */ | 
| 352 | #define NVM_FEAT_INTERRUPT_COALESCING		0x08 | 
| 353 | #define NVM_FEAT_INTERRUPT_VECTOR_CONFIGURATION 0x09 | 
| 354 | #define NVM_FEAT_WRITE_ATOMICITY		0x0a | 
| 355 | #define NVM_FEAT_ASYNC_EVENT_CONFIGURATION	0x0b | 
| 356 | #define NVM_FEAT_AUTONOMOUS_POWER_STATE_TRANSITION 0x0c | 
| 357 | #define NVM_FEAT_HOST_MEMORY_BUFFER		0x0d | 
| 358 | #define NVM_FEAT_TIMESTAMP			0x0e | 
| 359 | #define NVM_FEAT_KEEP_ALIVE_TIMER		0x0f | 
| 360 | #define NVM_FEAT_HOST_CONTROLLED_THERMAL_MGMT	0x10 | 
| 361 | #define NVM_FEAT_NON_OP_POWER_STATE_CONFIG	0x11 | 
| 362 | /* 0x12-0x77 - reserved */ | 
| 363 | /* 0x78-0x7f - NVMe Management Interface */ | 
| 364 | #define NVM_FEAT_SOFTWARE_PROGRESS_MARKER	0x80 | 
| 365 | #define NVM_FEAT_HOST_IDENTIFIER		0x81 | 
| 366 | #define NVM_FEAT_RESERVATION_NOTIFICATION_MASK	0x82 | 
| 367 | #define NVM_FEAT_RESERVATION_PERSISTANCE	0x83 | 
| 368 | /* 0x84-0xBF - command set specific (reserved) */ | 
| 369 | /* 0xC0-0xFF - vendor specific */ | 
| 370 |  | 
| 371 | #define NVM_SET_FEATURES_SV		__BIT(31)	/* Persist */ | 
| 372 |  | 
| 373 | #define NVM_VOLATILE_WRITE_CACHE_WCE	__BIT(0) 	/* Write Cache Enable */ | 
| 374 |  | 
| 375 | /* Power State Descriptor Data */ | 
| 376 | struct nvm_identify_psd { | 
| 377 | 	uint16_t	mp;		/* Max Power */ | 
| 378 | 	uint8_t		_reserved1; | 
| 379 | 	uint8_t		flags; | 
| 380 | #define	NVME_PSD_NOPS		__BIT(1) | 
| 381 | #define	NVME_PSD_MPS		__BIT(0) | 
| 382 |  | 
| 383 | 	uint32_t	enlat;		/* Entry Latency */ | 
| 384 |  | 
| 385 | 	uint32_t	exlat;		/* Exit Latency */ | 
| 386 |  | 
| 387 | 	uint8_t		rrt;		/* Relative Read Throughput */ | 
| 388 | #define	NVME_PSD_RRT_MASK	__BITS(0, 4) | 
| 389 | 	uint8_t		rrl;		/* Relative Read Latency */ | 
| 390 | #define	NVME_PSD_RRL_MASK	__BITS(0, 4) | 
| 391 | 	uint8_t		rwt;		/* Relative Write Throughput */ | 
| 392 | #define	NVME_PSD_RWT_MASK	__BITS(0, 4) | 
| 393 | 	uint8_t		rwl;		/* Relative Write Latency */ | 
| 394 | #define	NVME_PSD_RWL_MASK	__BITS(0, 4) | 
| 395 |  | 
| 396 | 	uint16_t	idlp;		/* Idle Power */ | 
| 397 | 	uint8_t		ips;		/* Idle Power Scale */ | 
| 398 | #define	NVME_PSD_IPS_MASK	__BITS(0, 1) | 
| 399 | 	uint8_t		_reserved2; | 
| 400 | 	uint16_t	actp;		/* Active Power */ | 
| 401 | 	uint16_t	ap;		/* Active Power Workload/Scale */ | 
| 402 | #define	NVME_PSD_APW_MASK	__BITS(0, 2) | 
| 403 | #define	NVME_PSD_APS_MASK	__BITS(6, 7) | 
| 404 |  | 
| 405 | 	uint8_t		_reserved[8]; | 
| 406 | } __packed __aligned(8); | 
| 407 | NVME_CTASSERT(sizeof(struct nvm_identify_psd) == 32, "bad size for nvm_identify_psd" ); | 
| 408 |  | 
| 409 | struct nvm_identify_controller { | 
| 410 | 	/* Controller Capabilities and Features */ | 
| 411 |  | 
| 412 | 	uint16_t	vid;		/* PCI Vendor ID */ | 
| 413 | 	uint16_t	ssvid;		/* PCI Subsystem Vendor ID */ | 
| 414 |  | 
| 415 | 	uint8_t		sn[20];		/* Serial Number */ | 
| 416 | 	uint8_t		mn[40];		/* Model Number */ | 
| 417 | 	uint8_t		fr[8];		/* Firmware Revision */ | 
| 418 |  | 
| 419 | 	uint8_t		rab;		/* Recommended Arbitration Burst */ | 
| 420 | 	uint8_t		ieee[3];	/* IEEE OUI Identifier */ | 
| 421 |  | 
| 422 | 	uint8_t		cmic;		/* Controller Multi-Path I/O and | 
| 423 | 					   Namespace Sharing Capabilities */ | 
| 424 | 	uint8_t		mdts;		/* Maximum Data Transfer Size */ | 
| 425 |  | 
| 426 | 	uint16_t	cntlid;		/* Controller ID */ | 
| 427 | 	uint32_t	ver;		/* Version */ | 
| 428 |  | 
| 429 | 	uint32_t	rtd3r;		/* RTD3 Resume Latency */ | 
| 430 | 	uint32_t	rtd3e;		/* RTD3 Enter Latency */ | 
| 431 |  | 
| 432 | 	uint32_t	oaes;		/* Optional Asynchronous Events Supported */ | 
| 433 | 	uint32_t	ctrattr;	/* Controller Attributes */ | 
| 434 |  | 
| 435 | 	uint8_t		_reserved1[12]; | 
| 436 |  | 
| 437 | 	uint8_t		fguid[16];	/* FRU Globally Unique Identifier */ | 
| 438 |  | 
| 439 | 	uint8_t		_reserved2[128]; | 
| 440 |  | 
| 441 | 	/* Admin Command Set Attributes & Optional Controller Capabilities */ | 
| 442 |  | 
| 443 | 	uint16_t	oacs;		/* Optional Admin Command Support */ | 
| 444 | #define	NVME_ID_CTRLR_OACS_DOORBELL_BC	__BIT(8) | 
| 445 | #define	NVME_ID_CTRLR_OACS_VIRT_MGMT	__BIT(7) | 
| 446 | #define	NVME_ID_CTRLR_OACS_NVME_MI	__BIT(6) | 
| 447 | #define	NVME_ID_CTRLR_OACS_DIRECTIVES	__BIT(5) | 
| 448 | #define	NVME_ID_CTRLR_OACS_DEV_SELFTEST	__BIT(4) | 
| 449 | #define	NVME_ID_CTRLR_OACS_NS		__BIT(3) | 
| 450 | #define	NVME_ID_CTRLR_OACS_FW		__BIT(2) | 
| 451 | #define	NVME_ID_CTRLR_OACS_FORMAT	__BIT(1) | 
| 452 | #define	NVME_ID_CTRLR_OACS_SECURITY	__BIT(0) | 
| 453 | 	uint8_t		acl;		/* Abort Command Limit */ | 
| 454 | 	uint8_t		aerl;		/* Asynchronous Event Request Limit */ | 
| 455 |  | 
| 456 | 	uint8_t		frmw;		/* Firmware Updates */ | 
| 457 | #define	NVME_ID_CTRLR_FRMW_NOREQ_RESET	__BIT(4) | 
| 458 | #define	NVME_ID_CTRLR_FRMW_NSLOT	__BITS(1, 3) | 
| 459 | #define	NVME_ID_CTRLR_FRMW_SLOT1_RO	__BIT(0) | 
| 460 | 	uint8_t		lpa;		/* Log Page Attributes */ | 
| 461 | #define	NVME_ID_CTRLR_LPA_CMD_EFFECT	__BIT(1) | 
| 462 | #define	NVME_ID_CTRLR_LPA_NS_SMART	__BIT(0) | 
| 463 | 	uint8_t		elpe;		/* Error Log Page Entries */ | 
| 464 | 	uint8_t		npss;		/* Number of Power States Support */ | 
| 465 |  | 
| 466 | 	uint8_t		avscc;		/* Admin Vendor Specific Command | 
| 467 | 					   Configuration */ | 
| 468 | 	uint8_t		apsta;		/* Autonomous Power State Transition | 
| 469 | 					   Attributes */ | 
| 470 |  | 
| 471 | 	uint16_t	wctemp;		/* Warning Composite Temperature | 
| 472 | 					   Threshold */ | 
| 473 | 	uint16_t	cctemp;		/* Critical Composite Temperature | 
| 474 | 					   Threshold */ | 
| 475 |  | 
| 476 | 	uint16_t	mtfa;		/* Maximum Time for Firmware Activation */ | 
| 477 |  | 
| 478 | 	uint32_t	hmpre;		/* Host Memory Buffer Preferred Size */ | 
| 479 | 	uint32_t	hmmin;		/* Host Memory Buffer Minimum Size */ | 
| 480 |  | 
| 481 | 	struct { | 
| 482 | 		uint64_t	tnvmcap[2]; | 
| 483 | 		uint64_t	unvmcap[2]; | 
| 484 | 	} __packed untncap;		/* Name space capabilities: | 
| 485 | 					   if NVME_ID_CTRLR_OACS_NS, | 
| 486 | 					   report tnvmcap and unvmcap */ | 
| 487 |  | 
| 488 | 	uint32_t	rpmbs;		/* Replay Protected Memory Block Support */ | 
| 489 |  | 
| 490 | 	uint16_t	edstt;		/* Extended Device Self-test Time */ | 
| 491 | 	uint8_t		dsto;		/* Device Self-test Options */ | 
| 492 |  | 
| 493 | 	uint8_t		fwug;		/* Firmware Update Granularity */ | 
| 494 |  | 
| 495 | 	uint16_t	kas;		/* Keep Alive Support */ | 
| 496 |  | 
| 497 | 	uint16_t	hctma;		/* Host Controlled Thermal Management | 
| 498 | 					   Attributes */ | 
| 499 | 	uint16_t	mntmt;		/* Minimum Thermal Management Temperature */ | 
| 500 | 	uint16_t	mxtmt;		/* Maximum Thermal Management Temperature */ | 
| 501 |  | 
| 502 | 	uint32_t	sanicap;	/* Sanitize Capabilities */ | 
| 503 |  | 
| 504 | 	uint8_t		_reserved3[180]; | 
| 505 |  | 
| 506 | 	/* NVM Command Set Attributes */ | 
| 507 |  | 
| 508 | 	uint8_t		sqes;		/* Submission Queue Entry Size */ | 
| 509 | #define	NVME_ID_CTRLR_SQES_MAX		__BITS(4, 7) | 
| 510 | #define	NVME_ID_CTRLR_SQES_MIN		__BITS(0, 3) | 
| 511 | 	uint8_t		cqes;		/* Completion Queue Entry Size */ | 
| 512 | #define	NVME_ID_CTRLR_CQES_MAX		__BITS(4, 7) | 
| 513 | #define	NVME_ID_CTRLR_CQES_MIN		__BITS(0, 3) | 
| 514 |  | 
| 515 | 	uint16_t	maxcmd;		/* Maximum Outstanding Commands */ | 
| 516 |  | 
| 517 | 	uint32_t	nn;		/* Number of Namespaces */ | 
| 518 |  | 
| 519 | 	uint16_t	oncs;		/* Optional NVM Command Support */ | 
| 520 | #define	NVME_ID_CTRLR_ONCS_TIMESTAMP	__BIT(6) | 
| 521 | #define	NVME_ID_CTRLR_ONCS_RESERVATION	__BIT(5) | 
| 522 | #define	NVME_ID_CTRLR_ONCS_SET_FEATURES	__BIT(4) | 
| 523 | #define	NVME_ID_CTRLR_ONCS_WRITE_ZERO	__BIT(3) | 
| 524 | #define	NVME_ID_CTRLR_ONCS_DSM		__BIT(2) | 
| 525 | #define	NVME_ID_CTRLR_ONCS_WRITE_UNC	__BIT(1) | 
| 526 | #define	NVME_ID_CTRLR_ONCS_COMPARE	__BIT(0) | 
| 527 | 	uint16_t	fuses;		/* Fused Operation Support */ | 
| 528 |  | 
| 529 | 	uint8_t		fna;		/* Format NVM Attributes */ | 
| 530 | #define	NVME_ID_CTRLR_FNA_CRYPTO_ERASE	__BIT(2) | 
| 531 | #define	NVME_ID_CTRLR_FNA_ERASE_ALL	__BIT(1) | 
| 532 | #define	NVME_ID_CTRLR_FNA_FORMAT_ALL	__BIT(0) | 
| 533 | 	uint8_t		vwc;		/* Volatile Write Cache */ | 
| 534 | #define	NVME_ID_CTRLR_VWC_PRESENT	__BIT(0) | 
| 535 | 	uint16_t	awun;		/* Atomic Write Unit Normal */ | 
| 536 | 	uint16_t	awupf;		/* Atomic Write Unit Power Fail */ | 
| 537 |  | 
| 538 | 	uint8_t		nvscc;		/* NVM Vendor Specific Command */ | 
| 539 | 	uint8_t		_reserved4[1]; | 
| 540 |  | 
| 541 | 	uint16_t	acwu;		/* Atomic Compare & Write Unit */ | 
| 542 | 	uint8_t		_reserved5[2]; | 
| 543 |  | 
| 544 | 	uint32_t	sgls;		/* SGL Support */ | 
| 545 |  | 
| 546 | 	uint8_t		_reserved6[228]; | 
| 547 |  | 
| 548 | 	uint8_t		subnqn[256];	/* NVM Subsystem NVMe Qualified Name */ | 
| 549 |  | 
| 550 | 	uint8_t		_reserved7[768]; | 
| 551 |  | 
| 552 | 	uint8_t		_reserved8[256]; /* NVMe over Fabrics specification */ | 
| 553 |  | 
| 554 | 	struct nvm_identify_psd psd[32]; /* Power State Descriptors */ | 
| 555 |  | 
| 556 | 	uint8_t		vs[1024];	/* Vendor Specific */ | 
| 557 | } __packed __aligned(8); | 
| 558 | NVME_CTASSERT(sizeof(struct nvm_identify_controller) == 4096, "bad size for nvm_identify_controller" ); | 
| 559 |  | 
| 560 | struct nvm_namespace_format { | 
| 561 | 	uint16_t	ms;		/* Metadata Size */ | 
| 562 | 	uint8_t		lbads;		/* LBA Data Size */ | 
| 563 | 	uint8_t		rp;		/* Relative Performance */ | 
| 564 | } __packed __aligned(4); | 
| 565 | NVME_CTASSERT(sizeof(struct nvm_namespace_format) == 4, "bad size for nvm_namespace_format" ); | 
| 566 |  | 
| 567 | struct nvm_identify_namespace { | 
| 568 | 	uint64_t	nsze;		/* Namespace Size */ | 
| 569 |  | 
| 570 | 	uint64_t	ncap;		/* Namespace Capacity */ | 
| 571 |  | 
| 572 | 	uint64_t	nuse;		/* Namespace Utilization */ | 
| 573 |  | 
| 574 | 	uint8_t		nsfeat;		/* Namespace Features */ | 
| 575 | #define	NVME_ID_NS_NSFEAT_LOGICAL_BLK_ERR	__BIT(2) | 
| 576 | #define	NVME_ID_NS_NSFEAT_NS			__BIT(1) | 
| 577 | #define	NVME_ID_NS_NSFEAT_THIN_PROV		__BIT(0) | 
| 578 | 	uint8_t		nlbaf;		/* Number of LBA Formats */ | 
| 579 | 	uint8_t		flbas;		/* Formatted LBA Size */ | 
| 580 | #define NVME_ID_NS_FLBAS(_f)			((_f) & 0x0f) | 
| 581 | #define NVME_ID_NS_FLBAS_MD			0x10 | 
| 582 | 	uint8_t		mc;		/* Metadata Capabilities */ | 
| 583 | 	uint8_t		dpc;		/* End-to-end Data Protection | 
| 584 | 					   Capabilities */ | 
| 585 | 	uint8_t		dps;		/* End-to-end Data Protection Type | 
| 586 | 					   Settings */ | 
| 587 | #define	NVME_ID_NS_DPS_MD_START			__BIT(3) | 
| 588 | #define	NVME_ID_NS_DPS_PIT(_f)			((_f) & 0x7) | 
| 589 |  | 
| 590 | 	uint8_t		nmic;		/* Namespace Multi-path I/O and Namespace | 
| 591 | 					   Sharing Capabilities */ | 
| 592 |  | 
| 593 | 	uint8_t		rescap;		/* Reservation Capabilities */ | 
| 594 |  | 
| 595 | 	uint8_t		fpi;		/* Format Progress Indicator */ | 
| 596 |  | 
| 597 | 	uint8_t		dlfeat;		/* Deallocate Logical Block Features */ | 
| 598 |  | 
| 599 | 	uint16_t	nawun;		/* Namespace Atomic Write Unit Normal  */ | 
| 600 | 	uint16_t	nawupf;		/* Namespace Atomic Write Unit Power Fail */ | 
| 601 | 	uint16_t	nacwu;		/* Namespace Atomic Compare & Write Unit */ | 
| 602 | 	uint16_t	nabsn;		/* Namespace Atomic Boundary Size Normal */ | 
| 603 | 	uint16_t	nabo;		/* Namespace Atomic Boundary Offset */ | 
| 604 | 	uint16_t	nabspf;		/* Namespace Atomic Boundary Size Power | 
| 605 | 					   Fail */ | 
| 606 | 	uint16_t	noiob;		/* Namespace Optimal IO Boundary */ | 
| 607 |  | 
| 608 | 	uint8_t		nvmcap[16];	/* NVM Capacity */ | 
| 609 |  | 
| 610 | 	uint8_t		_reserved1[40];	/* bytes 64-103: Reserved */ | 
| 611 |  | 
| 612 | 	uint8_t		nguid[16];	/* Namespace Globally Unique Identifier */ | 
| 613 | 	uint8_t		eui64[8];	/* IEEE Extended Unique Identifier */ | 
| 614 |  | 
| 615 | 	struct nvm_namespace_format | 
| 616 | 			lbaf[16];	/* LBA Format Support */ | 
| 617 |  | 
| 618 | 	uint8_t		_reserved2[192]; | 
| 619 |  | 
| 620 | 	uint8_t		vs[3712]; | 
| 621 | } __packed __aligned(8); | 
| 622 | NVME_CTASSERT(sizeof(struct nvm_identify_namespace) == 4096, "bad size for nvm_identify_namespace" ); | 
| 623 |  | 
| 624 | #endif	/* __NVMEREG_H__ */ | 
| 625 |  |