at v6.19 441 lines 12 kB view raw
1// SPDX-License-Identifier: GPL-2.0 2/* 3 * NVM Express target device driver tracepoints 4 * Copyright (c) 2018 Johannes Thumshirn, SUSE Linux GmbH 5 */ 6 7#include <linux/unaligned.h> 8#include "trace.h" 9 10static const char *nvmet_trace_admin_identify(struct trace_seq *p, u8 *cdw10) 11{ 12 const char *ret = trace_seq_buffer_ptr(p); 13 u8 cns = cdw10[0]; 14 u16 ctrlid = get_unaligned_le16(cdw10 + 2); 15 16 trace_seq_printf(p, "cns=%u, ctrlid=%u", cns, ctrlid); 17 trace_seq_putc(p, 0); 18 19 return ret; 20} 21 22static const char *nvmet_trace_admin_get_features(struct trace_seq *p, 23 u8 *cdw10) 24{ 25 const char *ret = trace_seq_buffer_ptr(p); 26 u8 fid = cdw10[0]; 27 u8 sel = cdw10[1] & 0x7; 28 u32 cdw11 = get_unaligned_le32(cdw10 + 4); 29 30 trace_seq_printf(p, "fid=0x%x, sel=0x%x, cdw11=0x%x", fid, sel, cdw11); 31 trace_seq_putc(p, 0); 32 33 return ret; 34} 35 36static const char *nvmet_trace_get_lba_status(struct trace_seq *p, 37 u8 *cdw10) 38{ 39 const char *ret = trace_seq_buffer_ptr(p); 40 u64 slba = get_unaligned_le64(cdw10); 41 u32 mndw = get_unaligned_le32(cdw10 + 8); 42 u16 rl = get_unaligned_le16(cdw10 + 12); 43 u8 atype = cdw10[15]; 44 45 trace_seq_printf(p, "slba=0x%llx, mndw=0x%x, rl=0x%x, atype=%u", 46 slba, mndw, rl, atype); 47 trace_seq_putc(p, 0); 48 49 return ret; 50} 51 52static const char *nvmet_trace_admin_set_features(struct trace_seq *p, 53 u8 *cdw10) 54{ 55 const char *ret = trace_seq_buffer_ptr(p); 56 u8 fid = cdw10[0]; 57 u8 sv = cdw10[3] & 0x8; 58 u32 cdw11 = get_unaligned_le32(cdw10 + 4); 59 60 trace_seq_printf(p, "fid=0x%x, sv=0x%x, cdw11=0x%x", fid, sv, cdw11); 61 trace_seq_putc(p, 0); 62 63 return ret; 64} 65 66static const char *nvmet_trace_read_write(struct trace_seq *p, u8 *cdw10) 67{ 68 const char *ret = trace_seq_buffer_ptr(p); 69 u64 slba = get_unaligned_le64(cdw10); 70 u16 length = get_unaligned_le16(cdw10 + 8); 71 u16 control = get_unaligned_le16(cdw10 + 10); 72 u32 dsmgmt = get_unaligned_le32(cdw10 + 12); 73 u32 reftag = get_unaligned_le32(cdw10 + 16); 74 75 trace_seq_printf(p, 76 "slba=%llu, len=%u, ctrl=0x%x, dsmgmt=%u, reftag=%u", 77 slba, length, control, dsmgmt, reftag); 78 trace_seq_putc(p, 0); 79 80 return ret; 81} 82 83static const char *nvmet_trace_dsm(struct trace_seq *p, u8 *cdw10) 84{ 85 const char *ret = trace_seq_buffer_ptr(p); 86 87 trace_seq_printf(p, "nr=%u, attributes=%u", 88 get_unaligned_le32(cdw10), 89 get_unaligned_le32(cdw10 + 4)); 90 trace_seq_putc(p, 0); 91 92 return ret; 93} 94 95static const char *nvmet_trace_common(struct trace_seq *p, u8 *cdw10) 96{ 97 const char *ret = trace_seq_buffer_ptr(p); 98 99 trace_seq_printf(p, "cdw10=%*ph", 24, cdw10); 100 trace_seq_putc(p, 0); 101 102 return ret; 103} 104 105const char *nvmet_trace_parse_admin_cmd(struct trace_seq *p, 106 u8 opcode, u8 *cdw10) 107{ 108 switch (opcode) { 109 case nvme_admin_identify: 110 return nvmet_trace_admin_identify(p, cdw10); 111 case nvme_admin_set_features: 112 return nvmet_trace_admin_set_features(p, cdw10); 113 case nvme_admin_get_features: 114 return nvmet_trace_admin_get_features(p, cdw10); 115 case nvme_admin_get_lba_status: 116 return nvmet_trace_get_lba_status(p, cdw10); 117 default: 118 return nvmet_trace_common(p, cdw10); 119 } 120} 121 122static const char *nvmet_trace_zone_mgmt_send(struct trace_seq *p, u8 *cdw10) 123{ 124 static const char * const zsa_strs[] = { 125 [0x01] = "close zone", 126 [0x02] = "finish zone", 127 [0x03] = "open zone", 128 [0x04] = "reset zone", 129 [0x05] = "offline zone", 130 [0x10] = "set zone descriptor extension" 131 }; 132 const char *ret = trace_seq_buffer_ptr(p); 133 u64 slba = get_unaligned_le64(cdw10); 134 const char *zsa_str; 135 u8 zsa = cdw10[12]; 136 u8 all = cdw10[13]; 137 138 if (zsa < ARRAY_SIZE(zsa_strs) && zsa_strs[zsa]) 139 zsa_str = zsa_strs[zsa]; 140 else 141 zsa_str = "reserved"; 142 143 trace_seq_printf(p, "slba=%llu, zsa=%u:%s, all=%u", 144 slba, zsa, zsa_str, all); 145 trace_seq_putc(p, 0); 146 147 return ret; 148} 149 150static const char *nvmet_trace_zone_mgmt_recv(struct trace_seq *p, u8 *cdw10) 151{ 152 static const char * const zrasf_strs[] = { 153 [0x00] = "list all zones", 154 [0x01] = "list the zones in the ZSE: Empty state", 155 [0x02] = "list the zones in the ZSIO: Implicitly Opened state", 156 [0x03] = "list the zones in the ZSEO: Explicitly Opened state", 157 [0x04] = "list the zones in the ZSC: Closed state", 158 [0x05] = "list the zones in the ZSF: Full state", 159 [0x06] = "list the zones in the ZSRO: Read Only state", 160 [0x07] = "list the zones in the ZSO: Offline state", 161 [0x09] = "list the zones that have the zone attribute" 162 }; 163 const char *ret = trace_seq_buffer_ptr(p); 164 u64 slba = get_unaligned_le64(cdw10); 165 u32 numd = get_unaligned_le32(&cdw10[8]); 166 u8 zra = cdw10[12]; 167 u8 zrasf = cdw10[13]; 168 const char *zrasf_str; 169 u8 pr = cdw10[14]; 170 171 if (zrasf < ARRAY_SIZE(zrasf_strs) && zrasf_strs[zrasf]) 172 zrasf_str = zrasf_strs[zrasf]; 173 else 174 zrasf_str = "reserved"; 175 176 trace_seq_printf(p, "slba=%llu, numd=%u, zra=%u, zrasf=%u:%s, pr=%u", 177 slba, numd, zra, zrasf, zrasf_str, pr); 178 trace_seq_putc(p, 0); 179 180 return ret; 181} 182 183static const char *nvmet_trace_resv_reg(struct trace_seq *p, u8 *cdw10) 184{ 185 static const char * const rrega_strs[] = { 186 [0x00] = "register", 187 [0x01] = "unregister", 188 [0x02] = "replace", 189 }; 190 const char *ret = trace_seq_buffer_ptr(p); 191 u8 rrega = cdw10[0] & 0x7; 192 u8 iekey = (cdw10[0] >> 3) & 0x1; 193 u8 ptpl = (cdw10[3] >> 6) & 0x3; 194 const char *rrega_str; 195 196 if (rrega < ARRAY_SIZE(rrega_strs) && rrega_strs[rrega]) 197 rrega_str = rrega_strs[rrega]; 198 else 199 rrega_str = "reserved"; 200 201 trace_seq_printf(p, "rrega=%u:%s, iekey=%u, ptpl=%u", 202 rrega, rrega_str, iekey, ptpl); 203 trace_seq_putc(p, 0); 204 205 return ret; 206} 207 208static const char * const rtype_strs[] = { 209 [0x00] = "reserved", 210 [0x01] = "write exclusive", 211 [0x02] = "exclusive access", 212 [0x03] = "write exclusive registrants only", 213 [0x04] = "exclusive access registrants only", 214 [0x05] = "write exclusive all registrants", 215 [0x06] = "exclusive access all registrants", 216}; 217 218static const char *nvmet_trace_resv_acq(struct trace_seq *p, u8 *cdw10) 219{ 220 static const char * const racqa_strs[] = { 221 [0x00] = "acquire", 222 [0x01] = "preempt", 223 [0x02] = "preempt and abort", 224 }; 225 const char *ret = trace_seq_buffer_ptr(p); 226 u8 racqa = cdw10[0] & 0x7; 227 u8 iekey = (cdw10[0] >> 3) & 0x1; 228 u8 rtype = cdw10[1]; 229 const char *racqa_str = "reserved"; 230 const char *rtype_str = "reserved"; 231 232 if (racqa < ARRAY_SIZE(racqa_strs) && racqa_strs[racqa]) 233 racqa_str = racqa_strs[racqa]; 234 235 if (rtype < ARRAY_SIZE(rtype_strs) && rtype_strs[rtype]) 236 rtype_str = rtype_strs[rtype]; 237 238 trace_seq_printf(p, "racqa=%u:%s, iekey=%u, rtype=%u:%s", 239 racqa, racqa_str, iekey, rtype, rtype_str); 240 trace_seq_putc(p, 0); 241 242 return ret; 243} 244 245static const char *nvmet_trace_resv_rel(struct trace_seq *p, u8 *cdw10) 246{ 247 static const char * const rrela_strs[] = { 248 [0x00] = "release", 249 [0x01] = "clear", 250 }; 251 const char *ret = trace_seq_buffer_ptr(p); 252 u8 rrela = cdw10[0] & 0x7; 253 u8 iekey = (cdw10[0] >> 3) & 0x1; 254 u8 rtype = cdw10[1]; 255 const char *rrela_str = "reserved"; 256 const char *rtype_str = "reserved"; 257 258 if (rrela < ARRAY_SIZE(rrela_strs) && rrela_strs[rrela]) 259 rrela_str = rrela_strs[rrela]; 260 261 if (rtype < ARRAY_SIZE(rtype_strs) && rtype_strs[rtype]) 262 rtype_str = rtype_strs[rtype]; 263 264 trace_seq_printf(p, "rrela=%u:%s, iekey=%u, rtype=%u:%s", 265 rrela, rrela_str, iekey, rtype, rtype_str); 266 trace_seq_putc(p, 0); 267 268 return ret; 269} 270 271static const char *nvmet_trace_resv_report(struct trace_seq *p, u8 *cdw10) 272{ 273 const char *ret = trace_seq_buffer_ptr(p); 274 u32 numd = get_unaligned_le32(cdw10); 275 u8 eds = cdw10[4] & 0x1; 276 277 trace_seq_printf(p, "numd=%u, eds=%u", numd, eds); 278 trace_seq_putc(p, 0); 279 280 return ret; 281} 282 283const char *nvmet_trace_parse_nvm_cmd(struct trace_seq *p, 284 u8 opcode, u8 *cdw10) 285{ 286 switch (opcode) { 287 case nvme_cmd_read: 288 case nvme_cmd_write: 289 case nvme_cmd_write_zeroes: 290 case nvme_cmd_zone_append: 291 return nvmet_trace_read_write(p, cdw10); 292 case nvme_cmd_dsm: 293 return nvmet_trace_dsm(p, cdw10); 294 case nvme_cmd_zone_mgmt_send: 295 return nvmet_trace_zone_mgmt_send(p, cdw10); 296 case nvme_cmd_zone_mgmt_recv: 297 return nvmet_trace_zone_mgmt_recv(p, cdw10); 298 case nvme_cmd_resv_register: 299 return nvmet_trace_resv_reg(p, cdw10); 300 case nvme_cmd_resv_acquire: 301 return nvmet_trace_resv_acq(p, cdw10); 302 case nvme_cmd_resv_release: 303 return nvmet_trace_resv_rel(p, cdw10); 304 case nvme_cmd_resv_report: 305 return nvmet_trace_resv_report(p, cdw10); 306 default: 307 return nvmet_trace_common(p, cdw10); 308 } 309} 310 311static const char *nvmet_trace_fabrics_property_set(struct trace_seq *p, 312 u8 *spc) 313{ 314 const char *ret = trace_seq_buffer_ptr(p); 315 u8 attrib = spc[0]; 316 u32 ofst = get_unaligned_le32(spc + 4); 317 u64 value = get_unaligned_le64(spc + 8); 318 319 trace_seq_printf(p, "attrib=%u, ofst=0x%x, value=0x%llx", 320 attrib, ofst, value); 321 trace_seq_putc(p, 0); 322 return ret; 323} 324 325static const char *nvmet_trace_fabrics_connect(struct trace_seq *p, 326 u8 *spc) 327{ 328 const char *ret = trace_seq_buffer_ptr(p); 329 u16 recfmt = get_unaligned_le16(spc); 330 u16 qid = get_unaligned_le16(spc + 2); 331 u16 sqsize = get_unaligned_le16(spc + 4); 332 u8 cattr = spc[6]; 333 u32 kato = get_unaligned_le32(spc + 8); 334 335 trace_seq_printf(p, "recfmt=%u, qid=%u, sqsize=%u, cattr=%u, kato=%u", 336 recfmt, qid, sqsize, cattr, kato); 337 trace_seq_putc(p, 0); 338 return ret; 339} 340 341static const char *nvmet_trace_fabrics_property_get(struct trace_seq *p, 342 u8 *spc) 343{ 344 const char *ret = trace_seq_buffer_ptr(p); 345 u8 attrib = spc[0]; 346 u32 ofst = get_unaligned_le32(spc + 4); 347 348 trace_seq_printf(p, "attrib=%u, ofst=0x%x", attrib, ofst); 349 trace_seq_putc(p, 0); 350 return ret; 351} 352 353static const char *nvmet_trace_fabrics_auth_send(struct trace_seq *p, u8 *spc) 354{ 355 const char *ret = trace_seq_buffer_ptr(p); 356 u8 spsp0 = spc[1]; 357 u8 spsp1 = spc[2]; 358 u8 secp = spc[3]; 359 u32 tl = get_unaligned_le32(spc + 4); 360 361 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, tl=%u", 362 spsp0, spsp1, secp, tl); 363 trace_seq_putc(p, 0); 364 return ret; 365} 366 367static const char *nvmet_trace_fabrics_auth_receive(struct trace_seq *p, u8 *spc) 368{ 369 const char *ret = trace_seq_buffer_ptr(p); 370 u8 spsp0 = spc[1]; 371 u8 spsp1 = spc[2]; 372 u8 secp = spc[3]; 373 u32 al = get_unaligned_le32(spc + 4); 374 375 trace_seq_printf(p, "spsp0=%02x, spsp1=%02x, secp=%02x, al=%u", 376 spsp0, spsp1, secp, al); 377 trace_seq_putc(p, 0); 378 return ret; 379} 380 381static const char *nvmet_trace_fabrics_common(struct trace_seq *p, u8 *spc) 382{ 383 const char *ret = trace_seq_buffer_ptr(p); 384 385 trace_seq_printf(p, "specific=%*ph", 24, spc); 386 trace_seq_putc(p, 0); 387 return ret; 388} 389 390const char *nvmet_trace_parse_fabrics_cmd(struct trace_seq *p, 391 u8 fctype, u8 *spc) 392{ 393 switch (fctype) { 394 case nvme_fabrics_type_property_set: 395 return nvmet_trace_fabrics_property_set(p, spc); 396 case nvme_fabrics_type_connect: 397 return nvmet_trace_fabrics_connect(p, spc); 398 case nvme_fabrics_type_property_get: 399 return nvmet_trace_fabrics_property_get(p, spc); 400 case nvme_fabrics_type_auth_send: 401 return nvmet_trace_fabrics_auth_send(p, spc); 402 case nvme_fabrics_type_auth_receive: 403 return nvmet_trace_fabrics_auth_receive(p, spc); 404 default: 405 return nvmet_trace_fabrics_common(p, spc); 406 } 407} 408 409const char *nvmet_trace_disk_name(struct trace_seq *p, char *name) 410{ 411 const char *ret = trace_seq_buffer_ptr(p); 412 413 if (*name) 414 trace_seq_printf(p, "disk=%s, ", name); 415 trace_seq_putc(p, 0); 416 417 return ret; 418} 419 420const char *nvmet_trace_ctrl_id(struct trace_seq *p, u16 ctrl_id) 421{ 422 const char *ret = trace_seq_buffer_ptr(p); 423 424 /* 425 * XXX: We don't know the controller instance before executing the 426 * connect command itself because the connect command for the admin 427 * queue will not provide the cntlid which will be allocated in this 428 * command. In case of io queues, the controller instance will be 429 * mapped by the extra data of the connect command. 430 * If we can know the extra data of the connect command in this stage, 431 * we can update this print statement later. 432 */ 433 if (ctrl_id) 434 trace_seq_printf(p, "%d", ctrl_id); 435 else 436 trace_seq_printf(p, "_"); 437 trace_seq_putc(p, 0); 438 439 return ret; 440} 441