LCOV - code coverage report
Current view: top level - lib/nvme - nvme_qpair.c (source / functions) Hit Total Coverage
Test: ut_cov_unit.info Lines: 310 458 67.7 %
Date: 2024-07-15 17:32:15 Functions: 28 37 75.7 %

          Line data    Source code
       1             : /*   SPDX-License-Identifier: BSD-3-Clause
       2             :  *   Copyright (C) 2015 Intel Corporation.
       3             :  *   All rights reserved.
       4             :  *   Copyright (c) 2022 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
       5             :  */
       6             : 
       7             : #include "nvme_internal.h"
       8             : #include "spdk/nvme_ocssd.h"
       9             : #include "spdk/string.h"
      10             : 
      11             : #define NVME_CMD_DPTR_STR_SIZE 256
      12             : 
      13             : static int nvme_qpair_resubmit_request(struct spdk_nvme_qpair *qpair, struct nvme_request *req);
      14             : 
      15             : struct nvme_string {
      16             :         uint16_t        value;
      17             :         const char      *str;
      18             : };
      19             : 
      20             : static const struct nvme_string admin_opcode[] = {
      21             :         { SPDK_NVME_OPC_DELETE_IO_SQ, "DELETE IO SQ" },
      22             :         { SPDK_NVME_OPC_CREATE_IO_SQ, "CREATE IO SQ" },
      23             :         { SPDK_NVME_OPC_GET_LOG_PAGE, "GET LOG PAGE" },
      24             :         { SPDK_NVME_OPC_DELETE_IO_CQ, "DELETE IO CQ" },
      25             :         { SPDK_NVME_OPC_CREATE_IO_CQ, "CREATE IO CQ" },
      26             :         { SPDK_NVME_OPC_IDENTIFY, "IDENTIFY" },
      27             :         { SPDK_NVME_OPC_ABORT, "ABORT" },
      28             :         { SPDK_NVME_OPC_SET_FEATURES, "SET FEATURES" },
      29             :         { SPDK_NVME_OPC_GET_FEATURES, "GET FEATURES" },
      30             :         { SPDK_NVME_OPC_ASYNC_EVENT_REQUEST, "ASYNC EVENT REQUEST" },
      31             :         { SPDK_NVME_OPC_NS_MANAGEMENT, "NAMESPACE MANAGEMENT" },
      32             :         { SPDK_NVME_OPC_FIRMWARE_COMMIT, "FIRMWARE COMMIT" },
      33             :         { SPDK_NVME_OPC_FIRMWARE_IMAGE_DOWNLOAD, "FIRMWARE IMAGE DOWNLOAD" },
      34             :         { SPDK_NVME_OPC_DEVICE_SELF_TEST, "DEVICE SELF-TEST" },
      35             :         { SPDK_NVME_OPC_NS_ATTACHMENT, "NAMESPACE ATTACHMENT" },
      36             :         { SPDK_NVME_OPC_KEEP_ALIVE, "KEEP ALIVE" },
      37             :         { SPDK_NVME_OPC_DIRECTIVE_SEND, "DIRECTIVE SEND" },
      38             :         { SPDK_NVME_OPC_DIRECTIVE_RECEIVE, "DIRECTIVE RECEIVE" },
      39             :         { SPDK_NVME_OPC_VIRTUALIZATION_MANAGEMENT, "VIRTUALIZATION MANAGEMENT" },
      40             :         { SPDK_NVME_OPC_NVME_MI_SEND, "NVME-MI SEND" },
      41             :         { SPDK_NVME_OPC_NVME_MI_RECEIVE, "NVME-MI RECEIVE" },
      42             :         { SPDK_NVME_OPC_DOORBELL_BUFFER_CONFIG, "DOORBELL BUFFER CONFIG" },
      43             :         { SPDK_NVME_OPC_FABRIC, "FABRIC" },
      44             :         { SPDK_NVME_OPC_FORMAT_NVM, "FORMAT NVM" },
      45             :         { SPDK_NVME_OPC_SECURITY_SEND, "SECURITY SEND" },
      46             :         { SPDK_NVME_OPC_SECURITY_RECEIVE, "SECURITY RECEIVE" },
      47             :         { SPDK_NVME_OPC_SANITIZE, "SANITIZE" },
      48             :         { SPDK_NVME_OPC_GET_LBA_STATUS, "GET LBA STATUS" },
      49             :         { SPDK_OCSSD_OPC_GEOMETRY, "OCSSD / GEOMETRY" },
      50             :         { 0xFFFF, "ADMIN COMMAND" }
      51             : };
      52             : 
      53             : static const struct nvme_string fabric_opcode[] = {
      54             :         { SPDK_NVMF_FABRIC_COMMAND_PROPERTY_SET, "PROPERTY SET" },
      55             :         { SPDK_NVMF_FABRIC_COMMAND_CONNECT, "CONNECT" },
      56             :         { SPDK_NVMF_FABRIC_COMMAND_PROPERTY_GET, "PROPERTY GET" },
      57             :         { SPDK_NVMF_FABRIC_COMMAND_AUTHENTICATION_SEND, "AUTHENTICATION SEND" },
      58             :         { SPDK_NVMF_FABRIC_COMMAND_AUTHENTICATION_RECV, "AUTHENTICATION RECV" },
      59             :         { 0xFFFF, "RESERVED / VENDOR SPECIFIC" }
      60             : };
      61             : 
      62             : static const struct nvme_string feat_opcode[] = {
      63             :         { SPDK_NVME_FEAT_ARBITRATION, "ARBITRATION" },
      64             :         { SPDK_NVME_FEAT_POWER_MANAGEMENT, "POWER MANAGEMENT" },
      65             :         { SPDK_NVME_FEAT_LBA_RANGE_TYPE, "LBA RANGE TYPE" },
      66             :         { SPDK_NVME_FEAT_TEMPERATURE_THRESHOLD, "TEMPERATURE THRESHOLD" },
      67             :         { SPDK_NVME_FEAT_ERROR_RECOVERY, "ERROR_RECOVERY" },
      68             :         { SPDK_NVME_FEAT_VOLATILE_WRITE_CACHE, "VOLATILE WRITE CACHE" },
      69             :         { SPDK_NVME_FEAT_NUMBER_OF_QUEUES, "NUMBER OF QUEUES" },
      70             :         { SPDK_NVME_FEAT_INTERRUPT_COALESCING, "INTERRUPT COALESCING" },
      71             :         { SPDK_NVME_FEAT_INTERRUPT_VECTOR_CONFIGURATION, "INTERRUPT VECTOR CONFIGURATION" },
      72             :         { SPDK_NVME_FEAT_WRITE_ATOMICITY, "WRITE ATOMICITY" },
      73             :         { SPDK_NVME_FEAT_ASYNC_EVENT_CONFIGURATION, "ASYNC EVENT CONFIGURATION" },
      74             :         { SPDK_NVME_FEAT_AUTONOMOUS_POWER_STATE_TRANSITION, "AUTONOMOUS POWER STATE TRANSITION" },
      75             :         { SPDK_NVME_FEAT_HOST_MEM_BUFFER, "HOST MEM BUFFER" },
      76             :         { SPDK_NVME_FEAT_TIMESTAMP, "TIMESTAMP" },
      77             :         { SPDK_NVME_FEAT_KEEP_ALIVE_TIMER, "KEEP ALIVE TIMER" },
      78             :         { SPDK_NVME_FEAT_HOST_CONTROLLED_THERMAL_MANAGEMENT, "HOST CONTROLLED THERMAL MANAGEMENT" },
      79             :         { SPDK_NVME_FEAT_NON_OPERATIONAL_POWER_STATE_CONFIG, "NON OPERATIONAL POWER STATE CONFIG" },
      80             :         { SPDK_NVME_FEAT_SOFTWARE_PROGRESS_MARKER, "SOFTWARE PROGRESS MARKER" },
      81             :         { SPDK_NVME_FEAT_HOST_IDENTIFIER, "HOST IDENTIFIER" },
      82             :         { SPDK_NVME_FEAT_HOST_RESERVE_MASK, "HOST RESERVE MASK" },
      83             :         { SPDK_NVME_FEAT_HOST_RESERVE_PERSIST, "HOST RESERVE PERSIST" },
      84             :         { 0xFFFF, "RESERVED" }
      85             : };
      86             : 
      87             : static const struct nvme_string io_opcode[] = {
      88             :         { SPDK_NVME_OPC_FLUSH, "FLUSH" },
      89             :         { SPDK_NVME_OPC_WRITE, "WRITE" },
      90             :         { SPDK_NVME_OPC_READ, "READ" },
      91             :         { SPDK_NVME_OPC_WRITE_UNCORRECTABLE, "WRITE UNCORRECTABLE" },
      92             :         { SPDK_NVME_OPC_COMPARE, "COMPARE" },
      93             :         { SPDK_NVME_OPC_WRITE_ZEROES, "WRITE ZEROES" },
      94             :         { SPDK_NVME_OPC_DATASET_MANAGEMENT, "DATASET MANAGEMENT" },
      95             :         { SPDK_NVME_OPC_RESERVATION_REGISTER, "RESERVATION REGISTER" },
      96             :         { SPDK_NVME_OPC_RESERVATION_REPORT, "RESERVATION REPORT" },
      97             :         { SPDK_NVME_OPC_RESERVATION_ACQUIRE, "RESERVATION ACQUIRE" },
      98             :         { SPDK_NVME_OPC_RESERVATION_RELEASE, "RESERVATION RELEASE" },
      99             :         { SPDK_OCSSD_OPC_VECTOR_RESET, "OCSSD / VECTOR RESET" },
     100             :         { SPDK_OCSSD_OPC_VECTOR_WRITE, "OCSSD / VECTOR WRITE" },
     101             :         { SPDK_OCSSD_OPC_VECTOR_READ, "OCSSD / VECTOR READ" },
     102             :         { SPDK_OCSSD_OPC_VECTOR_COPY, "OCSSD / VECTOR COPY" },
     103             :         { 0xFFFF, "IO COMMAND" }
     104             : };
     105             : 
     106             : static const struct nvme_string sgl_type[] = {
     107             :         { SPDK_NVME_SGL_TYPE_DATA_BLOCK, "DATA BLOCK" },
     108             :         { SPDK_NVME_SGL_TYPE_BIT_BUCKET, "BIT BUCKET" },
     109             :         { SPDK_NVME_SGL_TYPE_SEGMENT, "SEGMENT" },
     110             :         { SPDK_NVME_SGL_TYPE_LAST_SEGMENT, "LAST SEGMENT" },
     111             :         { SPDK_NVME_SGL_TYPE_KEYED_DATA_BLOCK, "KEYED DATA BLOCK" },
     112             :         { SPDK_NVME_SGL_TYPE_TRANSPORT_DATA_BLOCK, "TRANSPORT DATA BLOCK" },
     113             :         { SPDK_NVME_SGL_TYPE_VENDOR_SPECIFIC, "VENDOR SPECIFIC" },
     114             :         { 0xFFFF, "RESERVED" }
     115             : };
     116             : 
     117             : static const struct nvme_string sgl_subtype[] = {
     118             :         { SPDK_NVME_SGL_SUBTYPE_ADDRESS, "ADDRESS" },
     119             :         { SPDK_NVME_SGL_SUBTYPE_OFFSET, "OFFSET" },
     120             :         { SPDK_NVME_SGL_SUBTYPE_TRANSPORT, "TRANSPORT" },
     121             :         { SPDK_NVME_SGL_SUBTYPE_INVALIDATE_KEY, "INVALIDATE KEY" },
     122             :         { 0xFFFF, "RESERVED" }
     123             : };
     124             : 
     125             : static const char *
     126          17 : nvme_get_string(const struct nvme_string *strings, uint16_t value)
     127             : {
     128             :         const struct nvme_string *entry;
     129             : 
     130          17 :         entry = strings;
     131             : 
     132          59 :         while (entry->value != 0xFFFF) {
     133          59 :                 if (entry->value == value) {
     134          17 :                         return entry->str;
     135             :                 }
     136          42 :                 entry++;
     137             :         }
     138           0 :         return entry->str;
     139             : }
     140             : 
     141             : static void
     142           2 : nvme_get_sgl_unkeyed(char *buf, size_t size, struct spdk_nvme_cmd *cmd)
     143             : {
     144           2 :         struct spdk_nvme_sgl_descriptor *sgl = &cmd->dptr.sgl1;
     145             : 
     146           2 :         snprintf(buf, size, " len:0x%x", sgl->unkeyed.length);
     147           2 : }
     148             : 
     149             : static void
     150           2 : nvme_get_sgl_keyed(char *buf, size_t size, struct spdk_nvme_cmd *cmd)
     151             : {
     152           2 :         struct spdk_nvme_sgl_descriptor *sgl = &cmd->dptr.sgl1;
     153             : 
     154           2 :         snprintf(buf, size, " len:0x%x key:0x%x", sgl->keyed.length, sgl->keyed.key);
     155           2 : }
     156             : 
     157             : static void
     158           2 : nvme_get_sgl(char *buf, size_t size, struct spdk_nvme_cmd *cmd)
     159             : {
     160           2 :         struct spdk_nvme_sgl_descriptor *sgl = &cmd->dptr.sgl1;
     161             :         int c;
     162             : 
     163           2 :         c = snprintf(buf, size, "SGL %s %s 0x%" PRIx64, nvme_get_string(sgl_type, sgl->generic.type),
     164           2 :                      nvme_get_string(sgl_subtype, sgl->generic.subtype), sgl->address);
     165           2 :         assert(c >= 0 && (size_t)c < size);
     166             : 
     167           2 :         if (sgl->generic.type == SPDK_NVME_SGL_TYPE_DATA_BLOCK) {
     168           1 :                 nvme_get_sgl_unkeyed(buf + c, size - c, cmd);
     169             :         }
     170             : 
     171           2 :         if (sgl->generic.type == SPDK_NVME_SGL_TYPE_KEYED_DATA_BLOCK) {
     172           1 :                 nvme_get_sgl_keyed(buf + c, size - c, cmd);
     173             :         }
     174           2 : }
     175             : 
     176             : static void
     177           0 : nvme_get_prp(char *buf, size_t size, struct spdk_nvme_cmd *cmd)
     178             : {
     179           0 :         snprintf(buf, size, "PRP1 0x%" PRIx64 " PRP2 0x%" PRIx64, cmd->dptr.prp.prp1, cmd->dptr.prp.prp2);
     180           0 : }
     181             : 
     182             : static void
     183           5 : nvme_get_dptr(char *buf, size_t size, struct spdk_nvme_cmd *cmd)
     184             : {
     185           5 :         if (spdk_nvme_opc_get_data_transfer(cmd->opc) != SPDK_NVME_DATA_NONE) {
     186           0 :                 switch (cmd->psdt) {
     187           0 :                 case SPDK_NVME_PSDT_PRP:
     188           0 :                         nvme_get_prp(buf, size, cmd);
     189           0 :                         break;
     190           0 :                 case SPDK_NVME_PSDT_SGL_MPTR_CONTIG:
     191             :                 case SPDK_NVME_PSDT_SGL_MPTR_SGL:
     192           0 :                         nvme_get_sgl(buf, size, cmd);
     193           0 :                         break;
     194           5 :                 default:
     195             :                         ;
     196             :                 }
     197             :         }
     198           5 : }
     199             : 
     200             : static void
     201           0 : nvme_admin_qpair_print_command(uint16_t qid, struct spdk_nvme_cmd *cmd)
     202             : {
     203           0 :         struct spdk_nvmf_capsule_cmd *fcmd = (void *)cmd;
     204           0 :         char dptr[NVME_CMD_DPTR_STR_SIZE] = {'\0'};
     205             : 
     206           0 :         assert(cmd != NULL);
     207             : 
     208           0 :         nvme_get_dptr(dptr, sizeof(dptr), cmd);
     209             : 
     210           0 :         switch ((int)cmd->opc) {
     211           0 :         case SPDK_NVME_OPC_SET_FEATURES:
     212             :         case SPDK_NVME_OPC_GET_FEATURES:
     213           0 :                 SPDK_NOTICELOG("%s %s cid:%d cdw10:%08x %s\n",
     214             :                                nvme_get_string(admin_opcode, cmd->opc), nvme_get_string(feat_opcode,
     215             :                                                cmd->cdw10_bits.set_features.fid), cmd->cid, cmd->cdw10, dptr);
     216           0 :                 break;
     217           0 :         case SPDK_NVME_OPC_FABRIC:
     218           0 :                 SPDK_NOTICELOG("%s %s qid:%d cid:%d %s\n",
     219             :                                nvme_get_string(admin_opcode, cmd->opc), nvme_get_string(fabric_opcode, fcmd->fctype), qid,
     220             :                                fcmd->cid, dptr);
     221           0 :                 break;
     222           0 :         default:
     223           0 :                 SPDK_NOTICELOG("%s (%02x) qid:%d cid:%d nsid:%x cdw10:%08x cdw11:%08x %s\n",
     224             :                                nvme_get_string(admin_opcode, cmd->opc), cmd->opc, qid, cmd->cid, cmd->nsid, cmd->cdw10,
     225             :                                cmd->cdw11, dptr);
     226             :         }
     227           0 : }
     228             : 
     229             : static void
     230           5 : nvme_io_qpair_print_command(uint16_t qid, struct spdk_nvme_cmd *cmd)
     231             : {
     232           5 :         char dptr[NVME_CMD_DPTR_STR_SIZE] = {'\0'};
     233             : 
     234           5 :         assert(cmd != NULL);
     235             : 
     236           5 :         nvme_get_dptr(dptr, sizeof(dptr), cmd);
     237             : 
     238           5 :         switch ((int)cmd->opc) {
     239           0 :         case SPDK_NVME_OPC_WRITE:
     240             :         case SPDK_NVME_OPC_READ:
     241             :         case SPDK_NVME_OPC_WRITE_UNCORRECTABLE:
     242             :         case SPDK_NVME_OPC_COMPARE:
     243           0 :                 SPDK_NOTICELOG("%s sqid:%d cid:%d nsid:%d "
     244             :                                "lba:%llu len:%d %s\n",
     245             :                                nvme_get_string(io_opcode, cmd->opc), qid, cmd->cid, cmd->nsid,
     246             :                                ((unsigned long long)cmd->cdw11 << 32) + cmd->cdw10,
     247             :                                (cmd->cdw12 & 0xFFFF) + 1, dptr);
     248           0 :                 break;
     249           5 :         case SPDK_NVME_OPC_FLUSH:
     250             :         case SPDK_NVME_OPC_DATASET_MANAGEMENT:
     251           5 :                 SPDK_NOTICELOG("%s sqid:%d cid:%d nsid:%d\n",
     252             :                                nvme_get_string(io_opcode, cmd->opc), qid, cmd->cid, cmd->nsid);
     253           5 :                 break;
     254           0 :         default:
     255           0 :                 SPDK_NOTICELOG("%s (%02x) sqid:%d cid:%d nsid:%d\n",
     256             :                                nvme_get_string(io_opcode, cmd->opc), cmd->opc, qid, cmd->cid, cmd->nsid);
     257           0 :                 break;
     258             :         }
     259           5 : }
     260             : 
     261             : void
     262           5 : spdk_nvme_print_command(uint16_t qid, struct spdk_nvme_cmd *cmd)
     263             : {
     264           5 :         assert(cmd != NULL);
     265             : 
     266           5 :         if (qid == 0 || cmd->opc == SPDK_NVME_OPC_FABRIC) {
     267           0 :                 nvme_admin_qpair_print_command(qid, cmd);
     268             :         } else {
     269           5 :                 nvme_io_qpair_print_command(qid, cmd);
     270             :         }
     271           5 : }
     272             : 
     273             : void
     274           5 : spdk_nvme_qpair_print_command(struct spdk_nvme_qpair *qpair, struct spdk_nvme_cmd *cmd)
     275             : {
     276           5 :         assert(qpair != NULL);
     277           5 :         assert(cmd != NULL);
     278             : 
     279           5 :         spdk_nvme_print_command(qpair->id, cmd);
     280           5 : }
     281             : 
     282             : static const struct nvme_string status_type[] = {
     283             :         { SPDK_NVME_SCT_GENERIC, "GENERIC" },
     284             :         { SPDK_NVME_SCT_COMMAND_SPECIFIC, "COMMAND SPECIFIC" },
     285             :         { SPDK_NVME_SCT_MEDIA_ERROR, "MEDIA ERROR" },
     286             :         { SPDK_NVME_SCT_PATH, "PATH" },
     287             :         { SPDK_NVME_SCT_VENDOR_SPECIFIC, "VENDOR SPECIFIC" },
     288             :         { 0xFFFF, "RESERVED" },
     289             : };
     290             : 
     291             : static const struct nvme_string generic_status[] = {
     292             :         { SPDK_NVME_SC_SUCCESS, "SUCCESS" },
     293             :         { SPDK_NVME_SC_INVALID_OPCODE, "INVALID OPCODE" },
     294             :         { SPDK_NVME_SC_INVALID_FIELD, "INVALID FIELD" },
     295             :         { SPDK_NVME_SC_COMMAND_ID_CONFLICT, "COMMAND ID CONFLICT" },
     296             :         { SPDK_NVME_SC_DATA_TRANSFER_ERROR, "DATA TRANSFER ERROR" },
     297             :         { SPDK_NVME_SC_ABORTED_POWER_LOSS, "ABORTED - POWER LOSS" },
     298             :         { SPDK_NVME_SC_INTERNAL_DEVICE_ERROR, "INTERNAL DEVICE ERROR" },
     299             :         { SPDK_NVME_SC_ABORTED_BY_REQUEST, "ABORTED - BY REQUEST" },
     300             :         { SPDK_NVME_SC_ABORTED_SQ_DELETION, "ABORTED - SQ DELETION" },
     301             :         { SPDK_NVME_SC_ABORTED_FAILED_FUSED, "ABORTED - FAILED FUSED" },
     302             :         { SPDK_NVME_SC_ABORTED_MISSING_FUSED, "ABORTED - MISSING FUSED" },
     303             :         { SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT, "INVALID NAMESPACE OR FORMAT" },
     304             :         { SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR, "COMMAND SEQUENCE ERROR" },
     305             :         { SPDK_NVME_SC_INVALID_SGL_SEG_DESCRIPTOR, "INVALID SGL SEGMENT DESCRIPTOR" },
     306             :         { SPDK_NVME_SC_INVALID_NUM_SGL_DESCIRPTORS, "INVALID NUMBER OF SGL DESCRIPTORS" },
     307             :         { SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID, "DATA SGL LENGTH INVALID" },
     308             :         { SPDK_NVME_SC_METADATA_SGL_LENGTH_INVALID, "METADATA SGL LENGTH INVALID" },
     309             :         { SPDK_NVME_SC_SGL_DESCRIPTOR_TYPE_INVALID, "SGL DESCRIPTOR TYPE INVALID" },
     310             :         { SPDK_NVME_SC_INVALID_CONTROLLER_MEM_BUF, "INVALID CONTROLLER MEMORY BUFFER" },
     311             :         { SPDK_NVME_SC_INVALID_PRP_OFFSET, "INVALID PRP OFFSET" },
     312             :         { SPDK_NVME_SC_ATOMIC_WRITE_UNIT_EXCEEDED, "ATOMIC WRITE UNIT EXCEEDED" },
     313             :         { SPDK_NVME_SC_OPERATION_DENIED, "OPERATION DENIED" },
     314             :         { SPDK_NVME_SC_INVALID_SGL_OFFSET, "INVALID SGL OFFSET" },
     315             :         { SPDK_NVME_SC_HOSTID_INCONSISTENT_FORMAT, "HOSTID INCONSISTENT FORMAT" },
     316             :         { SPDK_NVME_SC_KEEP_ALIVE_EXPIRED, "KEEP ALIVE EXPIRED" },
     317             :         { SPDK_NVME_SC_KEEP_ALIVE_INVALID, "KEEP ALIVE INVALID" },
     318             :         { SPDK_NVME_SC_ABORTED_PREEMPT, "ABORTED - PREEMPT AND ABORT" },
     319             :         { SPDK_NVME_SC_SANITIZE_FAILED, "SANITIZE FAILED" },
     320             :         { SPDK_NVME_SC_SANITIZE_IN_PROGRESS, "SANITIZE IN PROGRESS" },
     321             :         { SPDK_NVME_SC_SGL_DATA_BLOCK_GRANULARITY_INVALID, "DATA BLOCK GRANULARITY INVALID" },
     322             :         { SPDK_NVME_SC_COMMAND_INVALID_IN_CMB, "COMMAND NOT SUPPORTED FOR QUEUE IN CMB" },
     323             :         { SPDK_NVME_SC_COMMAND_NAMESPACE_IS_PROTECTED, "COMMAND NAMESPACE IS PROTECTED" },
     324             :         { SPDK_NVME_SC_COMMAND_INTERRUPTED, "COMMAND INTERRUPTED" },
     325             :         { SPDK_NVME_SC_COMMAND_TRANSIENT_TRANSPORT_ERROR, "COMMAND TRANSIENT TRANSPORT ERROR" },
     326             :         { SPDK_NVME_SC_LBA_OUT_OF_RANGE, "LBA OUT OF RANGE" },
     327             :         { SPDK_NVME_SC_CAPACITY_EXCEEDED, "CAPACITY EXCEEDED" },
     328             :         { SPDK_NVME_SC_NAMESPACE_NOT_READY, "NAMESPACE NOT READY" },
     329             :         { SPDK_NVME_SC_RESERVATION_CONFLICT, "RESERVATION CONFLICT" },
     330             :         { SPDK_NVME_SC_FORMAT_IN_PROGRESS, "FORMAT IN PROGRESS" },
     331             :         { SPDK_NVME_SC_INVALID_VALUE_SIZE, "INVALID VALUE SIZE" },
     332             :         { SPDK_NVME_SC_INVALID_KEY_SIZE, "INVALID KEY SIZE" },
     333             :         { SPDK_NVME_SC_KV_KEY_DOES_NOT_EXIST, "KV KEY DOES NOT EXIST" },
     334             :         { SPDK_NVME_SC_UNRECOVERED_ERROR, "UNRECOVERED ERROR" },
     335             :         { SPDK_NVME_SC_KEY_EXISTS, "KEY EXISTS" },
     336             :         { 0xFFFF, "GENERIC" }
     337             : };
     338             : 
     339             : static const struct nvme_string command_specific_status[] = {
     340             :         { SPDK_NVME_SC_COMPLETION_QUEUE_INVALID, "INVALID COMPLETION QUEUE" },
     341             :         { SPDK_NVME_SC_INVALID_QUEUE_IDENTIFIER, "INVALID QUEUE IDENTIFIER" },
     342             :         { SPDK_NVME_SC_INVALID_QUEUE_SIZE, "INVALID QUEUE SIZE" },
     343             :         { SPDK_NVME_SC_ABORT_COMMAND_LIMIT_EXCEEDED, "ABORT CMD LIMIT EXCEEDED" },
     344             :         { SPDK_NVME_SC_ASYNC_EVENT_REQUEST_LIMIT_EXCEEDED, "ASYNC LIMIT EXCEEDED" },
     345             :         { SPDK_NVME_SC_INVALID_FIRMWARE_SLOT, "INVALID FIRMWARE SLOT" },
     346             :         { SPDK_NVME_SC_INVALID_FIRMWARE_IMAGE, "INVALID FIRMWARE IMAGE" },
     347             :         { SPDK_NVME_SC_INVALID_INTERRUPT_VECTOR, "INVALID INTERRUPT VECTOR" },
     348             :         { SPDK_NVME_SC_INVALID_LOG_PAGE, "INVALID LOG PAGE" },
     349             :         { SPDK_NVME_SC_INVALID_FORMAT, "INVALID FORMAT" },
     350             :         { SPDK_NVME_SC_FIRMWARE_REQ_CONVENTIONAL_RESET, "FIRMWARE REQUIRES CONVENTIONAL RESET" },
     351             :         { SPDK_NVME_SC_INVALID_QUEUE_DELETION, "INVALID QUEUE DELETION" },
     352             :         { SPDK_NVME_SC_FEATURE_ID_NOT_SAVEABLE, "FEATURE ID NOT SAVEABLE" },
     353             :         { SPDK_NVME_SC_FEATURE_NOT_CHANGEABLE, "FEATURE NOT CHANGEABLE" },
     354             :         { SPDK_NVME_SC_FEATURE_NOT_NAMESPACE_SPECIFIC, "FEATURE NOT NAMESPACE SPECIFIC" },
     355             :         { SPDK_NVME_SC_FIRMWARE_REQ_NVM_RESET, "FIRMWARE REQUIRES NVM RESET" },
     356             :         { SPDK_NVME_SC_FIRMWARE_REQ_RESET, "FIRMWARE REQUIRES RESET" },
     357             :         { SPDK_NVME_SC_FIRMWARE_REQ_MAX_TIME_VIOLATION, "FIRMWARE REQUIRES MAX TIME VIOLATION" },
     358             :         { SPDK_NVME_SC_FIRMWARE_ACTIVATION_PROHIBITED, "FIRMWARE ACTIVATION PROHIBITED" },
     359             :         { SPDK_NVME_SC_OVERLAPPING_RANGE, "OVERLAPPING RANGE" },
     360             :         { SPDK_NVME_SC_NAMESPACE_INSUFFICIENT_CAPACITY, "NAMESPACE INSUFFICIENT CAPACITY" },
     361             :         { SPDK_NVME_SC_NAMESPACE_ID_UNAVAILABLE, "NAMESPACE ID UNAVAILABLE" },
     362             :         { SPDK_NVME_SC_NAMESPACE_ALREADY_ATTACHED, "NAMESPACE ALREADY ATTACHED" },
     363             :         { SPDK_NVME_SC_NAMESPACE_IS_PRIVATE, "NAMESPACE IS PRIVATE" },
     364             :         { SPDK_NVME_SC_NAMESPACE_NOT_ATTACHED, "NAMESPACE NOT ATTACHED" },
     365             :         { SPDK_NVME_SC_THINPROVISIONING_NOT_SUPPORTED, "THINPROVISIONING NOT SUPPORTED" },
     366             :         { SPDK_NVME_SC_CONTROLLER_LIST_INVALID, "CONTROLLER LIST INVALID" },
     367             :         { SPDK_NVME_SC_DEVICE_SELF_TEST_IN_PROGRESS, "DEVICE SELF-TEST IN PROGRESS" },
     368             :         { SPDK_NVME_SC_BOOT_PARTITION_WRITE_PROHIBITED, "BOOT PARTITION WRITE PROHIBITED" },
     369             :         { SPDK_NVME_SC_INVALID_CTRLR_ID, "INVALID CONTROLLER ID" },
     370             :         { SPDK_NVME_SC_INVALID_SECONDARY_CTRLR_STATE, "INVALID SECONDARY CONTROLLER STATE" },
     371             :         { SPDK_NVME_SC_INVALID_NUM_CTRLR_RESOURCES, "INVALID NUMBER OF CONTROLLER RESOURCES" },
     372             :         { SPDK_NVME_SC_INVALID_RESOURCE_ID, "INVALID RESOURCE IDENTIFIER" },
     373             :         { SPDK_NVME_SC_SANITIZE_PROHIBITED, "SANITIZE PROHIBITED" },
     374             :         { SPDK_NVME_SC_ANA_GROUP_IDENTIFIER_INVALID, "ANA GROUP IDENTIFIER INVALID" },
     375             :         { SPDK_NVME_SC_ANA_ATTACH_FAILED, "ANA ATTACH FAILED" },
     376             :         { SPDK_NVME_SC_INSUFFICIENT_CAPACITY, "INSUFFICIENT CAPACITY" },
     377             :         { SPDK_NVME_SC_NAMESPACE_ATTACH_LIMIT_EXCEEDED, "NAMESPACE ATTACH LIMIT EXCEEDED" },
     378             :         { SPDK_NVME_SC_PROHIBIT_CMD_EXEC_NOT_SUPPORTED, "PROHIBIT COMMAND EXEC NOT SUPPORTED" },
     379             :         { SPDK_NVME_SC_IOCS_NOT_SUPPORTED, "IOCS NOT SUPPORTED" },
     380             :         { SPDK_NVME_SC_IOCS_NOT_ENABLED, "IOCS NOT ENABLED" },
     381             :         { SPDK_NVME_SC_IOCS_COMBINATION_REJECTED, "IOCS COMBINATION REJECTED" },
     382             :         { SPDK_NVME_SC_INVALID_IOCS, "INVALID IOCS" },
     383             :         { SPDK_NVME_SC_IDENTIFIER_UNAVAILABLE, "IDENTIFIER UNAVAILABLE" },
     384             :         { SPDK_NVME_SC_STREAM_RESOURCE_ALLOCATION_FAILED, "STREAM RESOURCE ALLOCATION FAILED"},
     385             :         { SPDK_NVME_SC_CONFLICTING_ATTRIBUTES, "CONFLICTING ATTRIBUTES" },
     386             :         { SPDK_NVME_SC_INVALID_PROTECTION_INFO, "INVALID PROTECTION INFO" },
     387             :         { SPDK_NVME_SC_ATTEMPTED_WRITE_TO_RO_RANGE, "WRITE TO RO RANGE" },
     388             :         { SPDK_NVME_SC_CMD_SIZE_LIMIT_SIZE_EXCEEDED, "CMD SIZE LIMIT SIZE EXCEEDED" },
     389             :         { SPDK_NVME_SC_ZONED_BOUNDARY_ERROR, "ZONED BOUNDARY ERROR" },
     390             :         { SPDK_NVME_SC_ZONE_IS_FULL, "ZONE IS FULL" },
     391             :         { SPDK_NVME_SC_ZONE_IS_READ_ONLY, "ZONE IS READ ONLY" },
     392             :         { SPDK_NVME_SC_ZONE_IS_OFFLINE, "ZONE IS OFFLINE" },
     393             :         { SPDK_NVME_SC_ZONE_INVALID_WRITE, "ZONE INVALID WRITE" },
     394             :         { SPDK_NVME_SC_TOO_MANY_ACTIVE_ZONES, "TOO MANY ACTIVE ZONES" },
     395             :         { SPDK_NVME_SC_TOO_MANY_OPEN_ZONES, "TOO MANY OPEN ZONES" },
     396             :         { SPDK_NVME_SC_INVALID_ZONE_STATE_TRANSITION, "INVALID ZONE STATE TRANSITION" },
     397             :         { 0xFFFF, "COMMAND SPECIFIC" }
     398             : };
     399             : 
     400             : static const struct nvme_string media_error_status[] = {
     401             :         { SPDK_NVME_SC_WRITE_FAULTS, "WRITE FAULTS" },
     402             :         { SPDK_NVME_SC_UNRECOVERED_READ_ERROR, "UNRECOVERED READ ERROR" },
     403             :         { SPDK_NVME_SC_GUARD_CHECK_ERROR, "GUARD CHECK ERROR" },
     404             :         { SPDK_NVME_SC_APPLICATION_TAG_CHECK_ERROR, "APPLICATION TAG CHECK ERROR" },
     405             :         { SPDK_NVME_SC_REFERENCE_TAG_CHECK_ERROR, "REFERENCE TAG CHECK ERROR" },
     406             :         { SPDK_NVME_SC_COMPARE_FAILURE, "COMPARE FAILURE" },
     407             :         { SPDK_NVME_SC_ACCESS_DENIED, "ACCESS DENIED" },
     408             :         { SPDK_NVME_SC_DEALLOCATED_OR_UNWRITTEN_BLOCK, "DEALLOCATED OR UNWRITTEN BLOCK" },
     409             :         { SPDK_NVME_SC_END_TO_END_STORAGE_TAG_CHECK_ERROR, "END TO END STORAGE TAG CHECK ERROR" },
     410             :         { SPDK_OCSSD_SC_OFFLINE_CHUNK, "RESET OFFLINE CHUNK" },
     411             :         { SPDK_OCSSD_SC_INVALID_RESET, "INVALID RESET" },
     412             :         { SPDK_OCSSD_SC_WRITE_FAIL_WRITE_NEXT_UNIT, "WRITE FAIL WRITE NEXT UNIT" },
     413             :         { SPDK_OCSSD_SC_WRITE_FAIL_CHUNK_EARLY_CLOSE, "WRITE FAIL CHUNK EARLY CLOSE" },
     414             :         { SPDK_OCSSD_SC_OUT_OF_ORDER_WRITE, "OUT OF ORDER WRITE" },
     415             :         { SPDK_OCSSD_SC_READ_HIGH_ECC, "READ HIGH ECC" },
     416             :         { 0xFFFF, "MEDIA ERROR" }
     417             : };
     418             : 
     419             : static const struct nvme_string path_status[] = {
     420             :         { SPDK_NVME_SC_INTERNAL_PATH_ERROR, "INTERNAL PATH ERROR" },
     421             :         { SPDK_NVME_SC_ASYMMETRIC_ACCESS_PERSISTENT_LOSS, "ASYMMETRIC ACCESS PERSISTENT LOSS" },
     422             :         { SPDK_NVME_SC_ASYMMETRIC_ACCESS_INACCESSIBLE, "ASYMMETRIC ACCESS INACCESSIBLE" },
     423             :         { SPDK_NVME_SC_ASYMMETRIC_ACCESS_TRANSITION, "ASYMMETRIC ACCESS TRANSITION" },
     424             :         { SPDK_NVME_SC_CONTROLLER_PATH_ERROR, "CONTROLLER PATH ERROR" },
     425             :         { SPDK_NVME_SC_HOST_PATH_ERROR, "HOST PATH ERROR" },
     426             :         { SPDK_NVME_SC_ABORTED_BY_HOST, "ABORTED BY HOST" },
     427             :         { 0xFFFF, "PATH ERROR" }
     428             : };
     429             : 
     430             : const char *
     431          10 : spdk_nvme_cpl_get_status_string(const struct spdk_nvme_status *status)
     432             : {
     433             :         const struct nvme_string *entry;
     434             : 
     435          10 :         switch (status->sct) {
     436           6 :         case SPDK_NVME_SCT_GENERIC:
     437           6 :                 entry = generic_status;
     438           6 :                 break;
     439           1 :         case SPDK_NVME_SCT_COMMAND_SPECIFIC:
     440           1 :                 entry = command_specific_status;
     441           1 :                 break;
     442           1 :         case SPDK_NVME_SCT_MEDIA_ERROR:
     443           1 :                 entry = media_error_status;
     444           1 :                 break;
     445           0 :         case SPDK_NVME_SCT_PATH:
     446           0 :                 entry = path_status;
     447           0 :                 break;
     448           1 :         case SPDK_NVME_SCT_VENDOR_SPECIFIC:
     449           1 :                 return "VENDOR SPECIFIC";
     450           1 :         default:
     451           1 :                 return "RESERVED";
     452             :         }
     453             : 
     454           8 :         return nvme_get_string(entry, status->sc);
     455             : }
     456             : 
     457             : const char *
     458           0 : spdk_nvme_cpl_get_status_type_string(const struct spdk_nvme_status *status)
     459             : {
     460           0 :         return nvme_get_string(status_type, status->sct);
     461             : }
     462             : 
     463             : void
     464           5 : spdk_nvme_print_completion(uint16_t qid, struct spdk_nvme_cpl *cpl)
     465             : {
     466           5 :         assert(cpl != NULL);
     467             : 
     468             :         /* Check that sqid matches qid. Note that sqid is reserved
     469             :          * for fabrics so don't print an error when sqid is 0. */
     470           5 :         if (cpl->sqid != qid && cpl->sqid != 0) {
     471           0 :                 SPDK_ERRLOG("sqid %u doesn't match qid\n", cpl->sqid);
     472             :         }
     473             : 
     474           5 :         SPDK_NOTICELOG("%s (%02x/%02x) qid:%d cid:%d cdw0:%x sqhd:%04x p:%x m:%x dnr:%x\n",
     475             :                        spdk_nvme_cpl_get_status_string(&cpl->status),
     476             :                        cpl->status.sct, cpl->status.sc, qid, cpl->cid, cpl->cdw0,
     477             :                        cpl->sqhd, cpl->status.p, cpl->status.m, cpl->status.dnr);
     478           5 : }
     479             : 
     480             : void
     481           5 : spdk_nvme_qpair_print_completion(struct spdk_nvme_qpair *qpair, struct spdk_nvme_cpl *cpl)
     482             : {
     483           5 :         spdk_nvme_print_completion(qpair->id, cpl);
     484           5 : }
     485             : 
     486             : bool
     487          32 : nvme_completion_is_retry(const struct spdk_nvme_cpl *cpl)
     488             : {
     489             :         /*
     490             :          * TODO: spec is not clear how commands that are aborted due
     491             :          *  to TLER will be marked.  So for now, it seems
     492             :          *  NAMESPACE_NOT_READY is the only case where we should
     493             :          *  look at the DNR bit.
     494             :          */
     495          32 :         switch ((int)cpl->status.sct) {
     496          26 :         case SPDK_NVME_SCT_GENERIC:
     497          26 :                 switch ((int)cpl->status.sc) {
     498           3 :                 case SPDK_NVME_SC_NAMESPACE_NOT_READY:
     499             :                 case SPDK_NVME_SC_FORMAT_IN_PROGRESS:
     500           3 :                         if (cpl->status.dnr) {
     501           1 :                                 return false;
     502             :                         } else {
     503           2 :                                 return true;
     504             :                         }
     505          23 :                 case SPDK_NVME_SC_INVALID_OPCODE:
     506             :                 case SPDK_NVME_SC_INVALID_FIELD:
     507             :                 case SPDK_NVME_SC_COMMAND_ID_CONFLICT:
     508             :                 case SPDK_NVME_SC_DATA_TRANSFER_ERROR:
     509             :                 case SPDK_NVME_SC_ABORTED_POWER_LOSS:
     510             :                 case SPDK_NVME_SC_INTERNAL_DEVICE_ERROR:
     511             :                 case SPDK_NVME_SC_ABORTED_BY_REQUEST:
     512             :                 case SPDK_NVME_SC_ABORTED_SQ_DELETION:
     513             :                 case SPDK_NVME_SC_ABORTED_FAILED_FUSED:
     514             :                 case SPDK_NVME_SC_ABORTED_MISSING_FUSED:
     515             :                 case SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT:
     516             :                 case SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR:
     517             :                 case SPDK_NVME_SC_LBA_OUT_OF_RANGE:
     518             :                 case SPDK_NVME_SC_CAPACITY_EXCEEDED:
     519             :                 default:
     520          23 :                         return false;
     521             :                 }
     522           2 :         case SPDK_NVME_SCT_PATH:
     523             :                 /*
     524             :                  * Per NVMe TP 4028 (Path and Transport Error Enhancements), retries should be
     525             :                  * based on the setting of the DNR bit for Internal Path Error
     526             :                  */
     527           2 :                 switch ((int)cpl->status.sc) {
     528           2 :                 case SPDK_NVME_SC_INTERNAL_PATH_ERROR:
     529           2 :                         return !cpl->status.dnr;
     530           0 :                 default:
     531           0 :                         return false;
     532             :                 }
     533           4 :         case SPDK_NVME_SCT_COMMAND_SPECIFIC:
     534             :         case SPDK_NVME_SCT_MEDIA_ERROR:
     535             :         case SPDK_NVME_SCT_VENDOR_SPECIFIC:
     536             :         default:
     537           4 :                 return false;
     538             :         }
     539             : }
     540             : 
     541             : static void
     542           7 : nvme_qpair_manual_complete_request(struct spdk_nvme_qpair *qpair,
     543             :                                    struct nvme_request *req, uint32_t sct, uint32_t sc,
     544             :                                    uint32_t dnr, bool print_on_error)
     545             : {
     546           7 :         struct spdk_nvme_cpl    cpl;
     547             :         bool                    error;
     548             : 
     549           7 :         memset(&cpl, 0, sizeof(cpl));
     550           7 :         cpl.sqid = qpair->id;
     551           7 :         cpl.status.sct = sct;
     552           7 :         cpl.status.sc = sc;
     553           7 :         cpl.status.dnr = dnr;
     554             : 
     555           7 :         error = spdk_nvme_cpl_is_error(&cpl);
     556             : 
     557           7 :         if (error && print_on_error && !qpair->ctrlr->opts.disable_error_logging) {
     558           5 :                 SPDK_NOTICELOG("Command completed manually:\n");
     559           5 :                 spdk_nvme_qpair_print_command(qpair, &req->cmd);
     560           5 :                 spdk_nvme_qpair_print_completion(qpair, &cpl);
     561             :         }
     562             : 
     563           7 :         nvme_complete_request(req->cb_fn, req->cb_arg, qpair, req, &cpl);
     564           7 : }
     565             : 
     566             : void
     567           2 : nvme_qpair_abort_queued_reqs(struct spdk_nvme_qpair *qpair)
     568             : {
     569             :         struct nvme_request             *req;
     570           2 :         STAILQ_HEAD(, nvme_request)     tmp;
     571             : 
     572           2 :         STAILQ_INIT(&tmp);
     573           2 :         STAILQ_SWAP(&tmp, &qpair->queued_req, nvme_request);
     574             : 
     575           5 :         while (!STAILQ_EMPTY(&tmp)) {
     576           3 :                 req = STAILQ_FIRST(&tmp);
     577           3 :                 STAILQ_REMOVE_HEAD(&tmp, stailq);
     578           3 :                 if (!qpair->ctrlr->opts.disable_error_logging) {
     579           3 :                         SPDK_ERRLOG("aborting queued i/o\n");
     580             :                 }
     581           3 :                 nvme_qpair_manual_complete_request(qpair, req, SPDK_NVME_SCT_GENERIC,
     582           3 :                                                    SPDK_NVME_SC_ABORTED_SQ_DELETION, qpair->abort_dnr, true);
     583             :         }
     584           2 : }
     585             : 
     586             : /* The callback to a request may submit the next request which is queued and
     587             :  * then the same callback may abort it immediately. This repetition may cause
     588             :  * infinite recursive calls. Hence move aborting requests to another list here
     589             :  * and abort them later at resubmission.
     590             :  */
     591             : static void
     592           7 : _nvme_qpair_complete_abort_queued_reqs(struct spdk_nvme_qpair *qpair)
     593             : {
     594             :         struct nvme_request             *req;
     595           7 :         STAILQ_HEAD(, nvme_request)     tmp;
     596             : 
     597           7 :         if (spdk_likely(STAILQ_EMPTY(&qpair->aborting_queued_req))) {
     598           6 :                 return;
     599             :         }
     600             : 
     601           1 :         STAILQ_INIT(&tmp);
     602           1 :         STAILQ_SWAP(&tmp, &qpair->aborting_queued_req, nvme_request);
     603             : 
     604           2 :         while (!STAILQ_EMPTY(&tmp)) {
     605           1 :                 req = STAILQ_FIRST(&tmp);
     606           1 :                 STAILQ_REMOVE_HEAD(&tmp, stailq);
     607           1 :                 nvme_qpair_manual_complete_request(qpair, req, SPDK_NVME_SCT_GENERIC,
     608             :                                                    SPDK_NVME_SC_ABORTED_BY_REQUEST, 1, true);
     609             :         }
     610             : }
     611             : 
     612             : uint32_t
     613           0 : nvme_qpair_abort_queued_reqs_with_cbarg(struct spdk_nvme_qpair *qpair, void *cmd_cb_arg)
     614             : {
     615             :         struct nvme_request     *req, *tmp;
     616           0 :         uint32_t                aborting = 0;
     617             : 
     618           0 :         STAILQ_FOREACH_SAFE(req, &qpair->queued_req, stailq, tmp) {
     619           0 :                 if ((req->cb_arg != cmd_cb_arg) &&
     620           0 :                     (req->parent == NULL || req->parent->cb_arg != cmd_cb_arg)) {
     621           0 :                         continue;
     622             :                 }
     623             : 
     624           0 :                 STAILQ_REMOVE(&qpair->queued_req, req, nvme_request, stailq);
     625           0 :                 STAILQ_INSERT_TAIL(&qpair->aborting_queued_req, req, stailq);
     626           0 :                 if (!qpair->ctrlr->opts.disable_error_logging) {
     627           0 :                         SPDK_ERRLOG("aborting queued i/o\n");
     628             :                 }
     629           0 :                 aborting++;
     630             :         }
     631             : 
     632           0 :         return aborting;
     633             : }
     634             : 
     635             : static inline bool
     636          15 : nvme_qpair_check_enabled(struct spdk_nvme_qpair *qpair)
     637             : {
     638             :         struct nvme_request *req;
     639             : 
     640             :         /*
     641             :          * Either during initial connect or reset, the qpair should follow the given state machine.
     642             :          * QPAIR_DISABLED->QPAIR_CONNECTING->QPAIR_CONNECTED->QPAIR_ENABLING->QPAIR_ENABLED. In the
     643             :          * reset case, once the qpair is properly connected, we need to abort any outstanding requests
     644             :          * from the old transport connection and encourage the application to retry them. We also need
     645             :          * to submit any queued requests that built up while we were in the connected or enabling state.
     646             :          */
     647          15 :         if (spdk_unlikely(nvme_qpair_get_state(qpair) == NVME_QPAIR_CONNECTED &&
     648             :                           !qpair->ctrlr->is_resetting)) {
     649           1 :                 nvme_qpair_set_state(qpair, NVME_QPAIR_ENABLING);
     650             :                 /*
     651             :                  * PCIe is special, for fabrics transports, we can abort requests before disconnect during reset
     652             :                  * but we have historically not disconnected pcie qpairs during reset so we have to abort requests
     653             :                  * here.
     654             :                  */
     655           1 :                 if (qpair->ctrlr->trid.trtype == SPDK_NVME_TRANSPORT_PCIE &&
     656           0 :                     !qpair->is_new_qpair) {
     657           0 :                         nvme_qpair_abort_all_queued_reqs(qpair);
     658           0 :                         nvme_transport_qpair_abort_reqs(qpair);
     659             :                 }
     660             : 
     661           1 :                 nvme_qpair_set_state(qpair, NVME_QPAIR_ENABLED);
     662           3 :                 while (!STAILQ_EMPTY(&qpair->queued_req)) {
     663           2 :                         req = STAILQ_FIRST(&qpair->queued_req);
     664           2 :                         STAILQ_REMOVE_HEAD(&qpair->queued_req, stailq);
     665           2 :                         if (nvme_qpair_resubmit_request(qpair, req)) {
     666           0 :                                 break;
     667             :                         }
     668             :                 }
     669             :         }
     670             : 
     671             :         /*
     672             :          * When doing a reset, we must disconnect the qpair on the proper core.
     673             :          * Note, reset is the only case where we set the failure reason without
     674             :          * setting the qpair state since reset is done at the generic layer on the
     675             :          * controller thread and we can't disconnect I/O qpairs from the controller
     676             :          * thread.
     677             :          */
     678          15 :         if (spdk_unlikely(qpair->transport_failure_reason != SPDK_NVME_QPAIR_FAILURE_NONE &&
     679             :                           nvme_qpair_get_state(qpair) == NVME_QPAIR_ENABLED)) {
     680             :                 /* Don't disconnect PCIe qpairs. They are a special case for reset. */
     681           0 :                 if (qpair->ctrlr->trid.trtype != SPDK_NVME_TRANSPORT_PCIE) {
     682           0 :                         nvme_ctrlr_disconnect_qpair(qpair);
     683             :                 }
     684           0 :                 if (qpair->transport_failure_reason == SPDK_NVME_QPAIR_FAILURE_RESET) {
     685             :                         /*
     686             :                          * For multi-process, a synchronous reset may not reconnect
     687             :                          * foreign IO qpairs. So we will reconnect them here instead.
     688             :                          */
     689           0 :                         nvme_ctrlr_reinitialize_io_qpair(qpair->ctrlr, qpair);
     690             :                 }
     691           0 :                 return false;
     692             :         }
     693             : 
     694          15 :         return nvme_qpair_get_state(qpair) == NVME_QPAIR_ENABLED;
     695             : }
     696             : 
     697             : void
     698           2 : nvme_qpair_resubmit_requests(struct spdk_nvme_qpair *qpair, uint32_t num_requests)
     699             : {
     700             :         uint32_t i;
     701             :         int resubmit_rc;
     702             :         struct nvme_request *req;
     703             : 
     704           2 :         assert(num_requests > 0);
     705             : 
     706           2 :         for (i = 0; i < num_requests; i++) {
     707           2 :                 if (qpair->ctrlr->is_resetting) {
     708           0 :                         break;
     709             :                 }
     710           2 :                 if ((req = STAILQ_FIRST(&qpair->queued_req)) == NULL) {
     711           1 :                         break;
     712             :                 }
     713           1 :                 STAILQ_REMOVE_HEAD(&qpair->queued_req, stailq);
     714           1 :                 resubmit_rc = nvme_qpair_resubmit_request(qpair, req);
     715           1 :                 if (spdk_unlikely(resubmit_rc != 0)) {
     716           1 :                         SPDK_DEBUGLOG(nvme, "Unable to resubmit as many requests as we completed.\n");
     717           1 :                         break;
     718             :                 }
     719             :         }
     720             : 
     721           2 :         _nvme_qpair_complete_abort_queued_reqs(qpair);
     722           2 : }
     723             : 
     724             : static void
     725           1 : nvme_complete_register_operations(struct spdk_nvme_qpair *qpair)
     726             : {
     727             :         struct nvme_register_completion *ctx, *tmp;
     728           1 :         struct spdk_nvme_ctrlr *ctrlr = qpair->ctrlr;
     729           1 :         STAILQ_HEAD(, nvme_register_completion) operations;
     730             : 
     731           1 :         STAILQ_INIT(&operations);
     732           1 :         nvme_ctrlr_lock(ctrlr);
     733           1 :         STAILQ_FOREACH_SAFE(ctx, &ctrlr->register_operations, stailq, tmp) {
     734             :                 /* We need to make sure we complete the register operation in
     735             :                  * the correct process.
     736             :                  */
     737           0 :                 if (ctx->pid != getpid()) {
     738           0 :                         continue;
     739             :                 }
     740           0 :                 STAILQ_REMOVE(&ctrlr->register_operations, ctx, nvme_register_completion, stailq);
     741           0 :                 STAILQ_INSERT_TAIL(&operations, ctx, stailq);
     742             :         }
     743           1 :         nvme_ctrlr_unlock(ctrlr);
     744             : 
     745           1 :         while (!STAILQ_EMPTY(&operations)) {
     746           0 :                 ctx = STAILQ_FIRST(&operations);
     747           0 :                 STAILQ_REMOVE_HEAD(&operations, stailq);
     748           0 :                 if (ctx->cb_fn != NULL) {
     749           0 :                         ctx->cb_fn(ctx->cb_ctx, ctx->value, &ctx->cpl);
     750             :                 }
     751           0 :                 spdk_free(ctx);
     752             :         }
     753           1 : }
     754             : 
     755             : int32_t
     756          10 : spdk_nvme_qpair_process_completions(struct spdk_nvme_qpair *qpair, uint32_t max_completions)
     757             : {
     758             :         int32_t ret;
     759             :         struct nvme_request *req, *tmp;
     760             : 
     761             :         /* Complete any pending register operations */
     762          10 :         if (nvme_qpair_is_admin_queue(qpair)) {
     763           1 :                 nvme_complete_register_operations(qpair);
     764             :         }
     765             : 
     766          10 :         if (spdk_unlikely(qpair->ctrlr->is_failed &&
     767             :                           nvme_qpair_get_state(qpair) != NVME_QPAIR_DISCONNECTING)) {
     768           2 :                 if (qpair->ctrlr->is_removed) {
     769           1 :                         nvme_qpair_set_state(qpair, NVME_QPAIR_DESTROYING);
     770           1 :                         nvme_qpair_abort_all_queued_reqs(qpair);
     771           1 :                         nvme_transport_qpair_abort_reqs(qpair);
     772             :                 }
     773           2 :                 return -ENXIO;
     774             :         }
     775             : 
     776           8 :         if (spdk_unlikely(!nvme_qpair_check_enabled(qpair) &&
     777             :                           !(nvme_qpair_get_state(qpair) == NVME_QPAIR_CONNECTING ||
     778             :                             nvme_qpair_get_state(qpair) == NVME_QPAIR_DISCONNECTING))) {
     779             :                 /*
     780             :                  * qpair is not enabled, likely because a controller reset is
     781             :                  *  in progress.
     782             :                  */
     783           3 :                 return -ENXIO;
     784             :         }
     785             : 
     786             :         /* error injection for those queued error requests */
     787           5 :         if (spdk_unlikely(!STAILQ_EMPTY(&qpair->err_req_head))) {
     788           0 :                 STAILQ_FOREACH_SAFE(req, &qpair->err_req_head, stailq, tmp) {
     789           0 :                         if (spdk_get_ticks() - req->submit_tick > req->timeout_tsc) {
     790           0 :                                 STAILQ_REMOVE(&qpair->err_req_head, req, nvme_request, stailq);
     791           0 :                                 nvme_qpair_manual_complete_request(qpair, req,
     792           0 :                                                                    req->cpl.status.sct,
     793           0 :                                                                    req->cpl.status.sc, qpair->abort_dnr, true);
     794             :                         }
     795             :                 }
     796             :         }
     797             : 
     798           5 :         qpair->in_completion_context = 1;
     799           5 :         ret = nvme_transport_qpair_process_completions(qpair, max_completions);
     800           5 :         if (ret < 0) {
     801           2 :                 if (ret == -ENXIO && nvme_qpair_get_state(qpair) == NVME_QPAIR_DISCONNECTING) {
     802           0 :                         ret = 0;
     803             :                 } else {
     804           2 :                         SPDK_ERRLOG("CQ transport error %d (%s) on qpair id %hu\n",
     805             :                                     ret, spdk_strerror(-ret), qpair->id);
     806           2 :                         if (nvme_qpair_is_admin_queue(qpair)) {
     807           1 :                                 nvme_ctrlr_fail(qpair->ctrlr, false);
     808             :                         }
     809             :                 }
     810             :         }
     811           5 :         qpair->in_completion_context = 0;
     812           5 :         if (qpair->delete_after_completion_context) {
     813             :                 /*
     814             :                  * A request to delete this qpair was made in the context of this completion
     815             :                  *  routine - so it is safe to delete it now.
     816             :                  */
     817           0 :                 spdk_nvme_ctrlr_free_io_qpair(qpair);
     818           0 :                 return ret;
     819             :         }
     820             : 
     821             :         /*
     822             :          * At this point, ret must represent the number of completions we reaped.
     823             :          * submit as many queued requests as we completed.
     824             :          */
     825           5 :         if (ret > 0) {
     826           2 :                 nvme_qpair_resubmit_requests(qpair, ret);
     827             :         } else {
     828           3 :                 _nvme_qpair_complete_abort_queued_reqs(qpair);
     829             :         }
     830             : 
     831           5 :         return ret;
     832             : }
     833             : 
     834             : spdk_nvme_qp_failure_reason
     835           0 : spdk_nvme_qpair_get_failure_reason(struct spdk_nvme_qpair *qpair)
     836             : {
     837           0 :         return qpair->transport_failure_reason;
     838             : }
     839             : 
     840             : void
     841           0 : spdk_nvme_qpair_set_abort_dnr(struct spdk_nvme_qpair *qpair, bool dnr)
     842             : {
     843           0 :         qpair->abort_dnr = dnr ? 1 : 0;
     844           0 : }
     845             : 
     846             : bool
     847           0 : spdk_nvme_qpair_is_connected(struct spdk_nvme_qpair *qpair)
     848             : {
     849           0 :         return nvme_qpair_get_state(qpair) >= NVME_QPAIR_CONNECTED &&
     850           0 :                nvme_qpair_get_state(qpair) <= NVME_QPAIR_ENABLED;
     851             : }
     852             : 
     853             : int
     854           8 : nvme_qpair_init(struct spdk_nvme_qpair *qpair, uint16_t id,
     855             :                 struct spdk_nvme_ctrlr *ctrlr,
     856             :                 enum spdk_nvme_qprio qprio,
     857             :                 uint32_t num_requests, bool async)
     858             : {
     859             :         struct nvme_request *req;
     860             :         size_t req_size_padded;
     861             :         uint32_t i;
     862             : 
     863           8 :         qpair->id = id;
     864           8 :         qpair->qprio = qprio;
     865             : 
     866           8 :         qpair->in_completion_context = 0;
     867           8 :         qpair->delete_after_completion_context = 0;
     868           8 :         qpair->no_deletion_notification_needed = 0;
     869             : 
     870           8 :         qpair->ctrlr = ctrlr;
     871           8 :         qpair->trtype = ctrlr->trid.trtype;
     872           8 :         qpair->is_new_qpair = true;
     873           8 :         qpair->async = async;
     874           8 :         qpair->poll_status = NULL;
     875           8 :         qpair->num_outstanding_reqs = 0;
     876             : 
     877           8 :         STAILQ_INIT(&qpair->free_req);
     878           8 :         STAILQ_INIT(&qpair->queued_req);
     879           8 :         STAILQ_INIT(&qpair->aborting_queued_req);
     880           8 :         TAILQ_INIT(&qpair->err_cmd_head);
     881           8 :         STAILQ_INIT(&qpair->err_req_head);
     882             : 
     883           8 :         req_size_padded = (sizeof(struct nvme_request) + 63) & ~(size_t)63;
     884             : 
     885             :         /* Add one for the reserved_req */
     886           8 :         num_requests++;
     887             : 
     888           8 :         qpair->req_buf = spdk_zmalloc(req_size_padded * num_requests, 64, NULL,
     889             :                                       SPDK_ENV_SOCKET_ID_ANY, SPDK_MALLOC_SHARE);
     890           8 :         if (qpair->req_buf == NULL) {
     891           0 :                 SPDK_ERRLOG("no memory to allocate qpair(cntlid:0x%x sqid:%d) req_buf with %d request\n",
     892             :                             ctrlr->cntlid, qpair->id, num_requests);
     893           0 :                 return -ENOMEM;
     894             :         }
     895             : 
     896         243 :         for (i = 0; i < num_requests; i++) {
     897         235 :                 req = (void *)((uintptr_t)qpair->req_buf + i * req_size_padded);
     898             : 
     899         235 :                 req->qpair = qpair;
     900         235 :                 if (i == 0) {
     901           8 :                         qpair->reserved_req = req;
     902             :                 } else {
     903         227 :                         STAILQ_INSERT_HEAD(&qpair->free_req, req, stailq);
     904             :                 }
     905             :         }
     906             : 
     907           8 :         return 0;
     908             : }
     909             : 
     910             : void
     911           2 : nvme_qpair_complete_error_reqs(struct spdk_nvme_qpair *qpair)
     912             : {
     913             :         struct nvme_request             *req;
     914             : 
     915           3 :         while (!STAILQ_EMPTY(&qpair->err_req_head)) {
     916           1 :                 req = STAILQ_FIRST(&qpair->err_req_head);
     917           1 :                 STAILQ_REMOVE_HEAD(&qpair->err_req_head, stailq);
     918           1 :                 nvme_qpair_manual_complete_request(qpair, req,
     919           1 :                                                    req->cpl.status.sct,
     920           1 :                                                    req->cpl.status.sc, qpair->abort_dnr, true);
     921             :         }
     922           2 : }
     923             : 
     924             : void
     925           1 : nvme_qpair_deinit(struct spdk_nvme_qpair *qpair)
     926             : {
     927             :         struct nvme_error_cmd *cmd, *entry;
     928             : 
     929           1 :         nvme_qpair_abort_queued_reqs(qpair);
     930           1 :         _nvme_qpair_complete_abort_queued_reqs(qpair);
     931           1 :         nvme_qpair_complete_error_reqs(qpair);
     932             : 
     933           2 :         TAILQ_FOREACH_SAFE(cmd, &qpair->err_cmd_head, link, entry) {
     934           1 :                 TAILQ_REMOVE(&qpair->err_cmd_head, cmd, link);
     935           1 :                 spdk_free(cmd);
     936             :         }
     937             : 
     938           1 :         spdk_free(qpair->req_buf);
     939           1 : }
     940             : 
     941             : static inline int
     942           7 : _nvme_qpair_submit_request(struct spdk_nvme_qpair *qpair, struct nvme_request *req)
     943             : {
     944           7 :         int                     rc = 0;
     945             :         struct nvme_request     *child_req, *tmp;
     946             :         struct nvme_error_cmd   *cmd;
     947           7 :         struct spdk_nvme_ctrlr  *ctrlr = qpair->ctrlr;
     948           7 :         bool                    child_req_failed = false;
     949             : 
     950           7 :         nvme_qpair_check_enabled(qpair);
     951             : 
     952           7 :         if (spdk_unlikely(nvme_qpair_get_state(qpair) == NVME_QPAIR_DISCONNECTED ||
     953             :                           nvme_qpair_get_state(qpair) == NVME_QPAIR_DISCONNECTING ||
     954             :                           nvme_qpair_get_state(qpair) == NVME_QPAIR_DESTROYING)) {
     955           9 :                 TAILQ_FOREACH_SAFE(child_req, &req->children, child_tailq, tmp) {
     956           6 :                         nvme_request_remove_child(req, child_req);
     957           6 :                         nvme_request_free_children(child_req);
     958           6 :                         nvme_free_request(child_req);
     959             :                 }
     960             : 
     961           3 :                 rc = -ENXIO;
     962           3 :                 goto error;
     963             :         }
     964             : 
     965           4 :         if (req->num_children) {
     966             :                 /*
     967             :                  * This is a split (parent) request. Submit all of the children but not the parent
     968             :                  * request itself, since the parent is the original unsplit request.
     969             :                  */
     970           0 :                 TAILQ_FOREACH_SAFE(child_req, &req->children, child_tailq, tmp) {
     971           0 :                         if (spdk_likely(!child_req_failed)) {
     972           0 :                                 rc = nvme_qpair_submit_request(qpair, child_req);
     973           0 :                                 if (spdk_unlikely(rc != 0)) {
     974           0 :                                         child_req_failed = true;
     975             :                                 }
     976             :                         } else { /* free remaining child_reqs since one child_req fails */
     977           0 :                                 nvme_request_remove_child(req, child_req);
     978           0 :                                 nvme_request_free_children(child_req);
     979           0 :                                 nvme_free_request(child_req);
     980             :                         }
     981             :                 }
     982             : 
     983           0 :                 if (spdk_unlikely(child_req_failed)) {
     984             :                         /* part of children requests have been submitted,
     985             :                          * return success since we must wait for those children to complete,
     986             :                          * but set the parent request to failure.
     987             :                          */
     988           0 :                         if (req->num_children) {
     989           0 :                                 req->cpl.status.sct = SPDK_NVME_SCT_GENERIC;
     990           0 :                                 req->cpl.status.sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR;
     991           0 :                                 return 0;
     992             :                         }
     993           0 :                         goto error;
     994             :                 }
     995             : 
     996           0 :                 return rc;
     997             :         }
     998             : 
     999             :         /* queue those requests which matches with opcode in err_cmd list */
    1000           4 :         if (spdk_unlikely(!TAILQ_EMPTY(&qpair->err_cmd_head))) {
    1001           0 :                 TAILQ_FOREACH(cmd, &qpair->err_cmd_head, link) {
    1002           0 :                         if (!cmd->do_not_submit) {
    1003           0 :                                 continue;
    1004             :                         }
    1005             : 
    1006           0 :                         if ((cmd->opc == req->cmd.opc) && cmd->err_count) {
    1007             :                                 /* add to error request list and set cpl */
    1008           0 :                                 req->timeout_tsc = cmd->timeout_tsc;
    1009           0 :                                 req->submit_tick = spdk_get_ticks();
    1010           0 :                                 req->cpl.status.sct = cmd->status.sct;
    1011           0 :                                 req->cpl.status.sc = cmd->status.sc;
    1012           0 :                                 STAILQ_INSERT_TAIL(&qpair->err_req_head, req, stailq);
    1013           0 :                                 cmd->err_count--;
    1014           0 :                                 return 0;
    1015             :                         }
    1016             :                 }
    1017             :         }
    1018             : 
    1019           4 :         if (spdk_unlikely(ctrlr->is_failed)) {
    1020           0 :                 rc = -ENXIO;
    1021           0 :                 goto error;
    1022             :         }
    1023             : 
    1024             :         /* assign submit_tick before submitting req to specific transport */
    1025           4 :         if (spdk_unlikely(ctrlr->timeout_enabled)) {
    1026           0 :                 if (req->submit_tick == 0) { /* req submitted for the first time */
    1027           0 :                         req->submit_tick = spdk_get_ticks();
    1028           0 :                         req->timed_out = false;
    1029             :                 }
    1030             :         } else {
    1031           4 :                 req->submit_tick = 0;
    1032             :         }
    1033             : 
    1034             :         /* Allow two cases:
    1035             :          * 1. NVMe qpair is enabled.
    1036             :          * 2. Always allow fabrics commands through - these get
    1037             :          * the controller out of reset state.
    1038             :          */
    1039           4 :         if (spdk_likely(nvme_qpair_get_state(qpair) == NVME_QPAIR_ENABLED) ||
    1040           0 :             (req->cmd.opc == SPDK_NVME_OPC_FABRIC &&
    1041           0 :              nvme_qpair_get_state(qpair) == NVME_QPAIR_CONNECTING)) {
    1042           4 :                 rc = nvme_transport_qpair_submit_request(qpair, req);
    1043             :         } else {
    1044             :                 /* The controller is being reset - queue this request and
    1045             :                  *  submit it later when the reset is completed.
    1046             :                  */
    1047           0 :                 return -EAGAIN;
    1048             :         }
    1049             : 
    1050           4 :         if (spdk_likely(rc == 0)) {
    1051           3 :                 if (SPDK_DEBUGLOG_FLAG_ENABLED("nvme")) {
    1052           0 :                         spdk_nvme_print_command(qpair->id, &req->cmd);
    1053             :                 }
    1054           3 :                 req->queued = false;
    1055           3 :                 return 0;
    1056             :         }
    1057             : 
    1058           1 :         if (rc == -EAGAIN) {
    1059           0 :                 return -EAGAIN;
    1060             :         }
    1061             : 
    1062           1 : error:
    1063           4 :         if (req->parent != NULL) {
    1064           0 :                 nvme_request_remove_child(req->parent, req);
    1065             :         }
    1066             : 
    1067             :         /* The request is from queued_req list we should trigger the callback from caller */
    1068           4 :         if (spdk_unlikely(req->queued)) {
    1069           1 :                 if (rc == -ENXIO) {
    1070           0 :                         nvme_qpair_manual_complete_request(qpair, req, SPDK_NVME_SCT_GENERIC,
    1071             :                                                            SPDK_NVME_SC_ABORTED_SQ_DELETION,
    1072           0 :                                                            qpair->abort_dnr, true);
    1073             :                 } else {
    1074           1 :                         nvme_qpair_manual_complete_request(qpair, req, SPDK_NVME_SCT_GENERIC,
    1075             :                                                            SPDK_NVME_SC_INTERNAL_DEVICE_ERROR,
    1076             :                                                            true, true);
    1077             :                 }
    1078           1 :                 return rc;
    1079             :         }
    1080             : 
    1081           3 :         nvme_cleanup_user_req(req);
    1082           3 :         nvme_free_request(req);
    1083             : 
    1084           3 :         return rc;
    1085             : }
    1086             : 
    1087             : int
    1088           4 : nvme_qpair_submit_request(struct spdk_nvme_qpair *qpair, struct nvme_request *req)
    1089             : {
    1090             :         int rc;
    1091             : 
    1092           4 :         if (spdk_unlikely(!STAILQ_EMPTY(&qpair->queued_req) && req->num_children == 0)) {
    1093             :                 /*
    1094             :                  * Requests that have no children should be sent to the transport after all
    1095             :                  * currently queued requests. Requests with children will be split and go back
    1096             :                  * through this path.  We need to make an exception for the fabrics commands
    1097             :                  * while the qpair is connecting to be able to send the connect command
    1098             :                  * asynchronously.
    1099             :                  */
    1100           0 :                 if (req->cmd.opc != SPDK_NVME_OPC_FABRIC ||
    1101           0 :                     nvme_qpair_get_state(qpair) != NVME_QPAIR_CONNECTING) {
    1102           0 :                         STAILQ_INSERT_TAIL(&qpair->queued_req, req, stailq);
    1103           0 :                         req->queued = true;
    1104           0 :                         return 0;
    1105             :                 }
    1106             :         }
    1107             : 
    1108           4 :         rc = _nvme_qpair_submit_request(qpair, req);
    1109           4 :         if (rc == -EAGAIN) {
    1110           0 :                 STAILQ_INSERT_TAIL(&qpair->queued_req, req, stailq);
    1111           0 :                 req->queued = true;
    1112           0 :                 rc = 0;
    1113             :         }
    1114             : 
    1115           4 :         return rc;
    1116             : }
    1117             : 
    1118             : static int
    1119           3 : nvme_qpair_resubmit_request(struct spdk_nvme_qpair *qpair, struct nvme_request *req)
    1120             : {
    1121             :         int rc;
    1122             : 
    1123             :         /*
    1124             :          * We should never have a request with children on the queue.
    1125             :          * This is necessary to preserve the 1:1 relationship between
    1126             :          * completions and resubmissions.
    1127             :          */
    1128           3 :         assert(req->num_children == 0);
    1129           3 :         assert(req->queued);
    1130           3 :         rc = _nvme_qpair_submit_request(qpair, req);
    1131           3 :         if (spdk_unlikely(rc == -EAGAIN)) {
    1132           0 :                 STAILQ_INSERT_HEAD(&qpair->queued_req, req, stailq);
    1133             :         }
    1134             : 
    1135           3 :         return rc;
    1136             : }
    1137             : 
    1138             : void
    1139           1 : nvme_qpair_abort_all_queued_reqs(struct spdk_nvme_qpair *qpair)
    1140             : {
    1141           1 :         nvme_qpair_complete_error_reqs(qpair);
    1142           1 :         nvme_qpair_abort_queued_reqs(qpair);
    1143           1 :         _nvme_qpair_complete_abort_queued_reqs(qpair);
    1144           1 :         if (nvme_qpair_is_admin_queue(qpair)) {
    1145           0 :                 nvme_ctrlr_abort_queued_aborts(qpair->ctrlr);
    1146             :         }
    1147           1 : }
    1148             : 
    1149             : int
    1150           4 : spdk_nvme_qpair_add_cmd_error_injection(struct spdk_nvme_ctrlr *ctrlr,
    1151             :                                         struct spdk_nvme_qpair *qpair,
    1152             :                                         uint8_t opc, bool do_not_submit,
    1153             :                                         uint64_t timeout_in_us,
    1154             :                                         uint32_t err_count,
    1155             :                                         uint8_t sct, uint8_t sc)
    1156             : {
    1157           4 :         struct nvme_error_cmd *entry, *cmd = NULL;
    1158           4 :         int rc = 0;
    1159             : 
    1160           4 :         if (qpair == NULL) {
    1161           1 :                 qpair = ctrlr->adminq;
    1162           1 :                 nvme_ctrlr_lock(ctrlr);
    1163             :         }
    1164             : 
    1165           4 :         TAILQ_FOREACH(entry, &qpair->err_cmd_head, link) {
    1166           1 :                 if (entry->opc == opc) {
    1167           1 :                         cmd = entry;
    1168           1 :                         break;
    1169             :                 }
    1170             :         }
    1171             : 
    1172           4 :         if (cmd == NULL) {
    1173           3 :                 cmd = spdk_zmalloc(sizeof(*cmd), 64, NULL, SPDK_ENV_LCORE_ID_ANY, SPDK_MALLOC_DMA);
    1174           3 :                 if (!cmd) {
    1175           0 :                         rc = -ENOMEM;
    1176           0 :                         goto out;
    1177             :                 }
    1178           3 :                 TAILQ_INSERT_TAIL(&qpair->err_cmd_head, cmd, link);
    1179             :         }
    1180             : 
    1181           4 :         cmd->do_not_submit = do_not_submit;
    1182           4 :         cmd->err_count = err_count;
    1183           4 :         cmd->timeout_tsc = timeout_in_us * spdk_get_ticks_hz() / 1000000ULL;
    1184           4 :         cmd->opc = opc;
    1185           4 :         cmd->status.sct = sct;
    1186           4 :         cmd->status.sc = sc;
    1187           4 : out:
    1188           4 :         if (nvme_qpair_is_admin_queue(qpair)) {
    1189           0 :                 nvme_ctrlr_unlock(ctrlr);
    1190             :         }
    1191             : 
    1192           4 :         return rc;
    1193             : }
    1194             : 
    1195             : void
    1196           3 : spdk_nvme_qpair_remove_cmd_error_injection(struct spdk_nvme_ctrlr *ctrlr,
    1197             :                 struct spdk_nvme_qpair *qpair,
    1198             :                 uint8_t opc)
    1199             : {
    1200             :         struct nvme_error_cmd *cmd, *entry;
    1201             : 
    1202           3 :         if (qpair == NULL) {
    1203           1 :                 qpair = ctrlr->adminq;
    1204           1 :                 nvme_ctrlr_lock(ctrlr);
    1205             :         }
    1206             : 
    1207           3 :         TAILQ_FOREACH_SAFE(cmd, &qpair->err_cmd_head, link, entry) {
    1208           3 :                 if (cmd->opc == opc) {
    1209           3 :                         TAILQ_REMOVE(&qpair->err_cmd_head, cmd, link);
    1210           3 :                         spdk_free(cmd);
    1211           3 :                         break;
    1212             :                 }
    1213             :         }
    1214             : 
    1215           3 :         if (nvme_qpair_is_admin_queue(qpair)) {
    1216           0 :                 nvme_ctrlr_unlock(ctrlr);
    1217             :         }
    1218           3 : }
    1219             : 
    1220             : uint16_t
    1221           0 : spdk_nvme_qpair_get_id(struct spdk_nvme_qpair *qpair)
    1222             : {
    1223           0 :         return qpair->id;
    1224             : }
    1225             : 
    1226             : uint32_t
    1227           0 : spdk_nvme_qpair_get_num_outstanding_reqs(struct spdk_nvme_qpair *qpair)
    1228             : {
    1229           0 :         return qpair->num_outstanding_reqs;
    1230             : }

Generated by: LCOV version 1.15